Skip to main content

Full text of "SECOND EDITION STATISTICS IN RESEARCH"

See other formats


$t0.50 


STATISTICS 
IN  RESEARCH 


Which  statistical  tool  shoaLl  I  use? 

How  should  I  use  if? 

How  can  I  interpret  my  results? 

SUCH  QUESTIONS  plague  every  engineer, 
scientist,  research  worker,  student,  and 
teacher,  and  they  will  tmd  this  book  an 
indispensable  reference  which  contains 
the  answers  to  these  and  many  other 
questions. 

Modern  techniques  arc  presented  in 
ways  conducive  to  easy  leaning  and 
application.  The  most  difficult  part  of 
statistical  method  is  learning  when  and 
where  to  apply  a  particular  technique. 
Statistics  in  Research  carefully  indicates 
the  assumptions  underlying  ruii  tech 
nique  so  that  it  will  be  applied  properly 
within  its  limitations, 

Especially    noteworthy    features    in 
clude; 

*  emphasis  on  theory  for  a  thorough 
undorstaiuling  of  fundamentals 


*  a  comprehensive  coveuv^  of  re 
gression  analysis 

*  useful  check-lists  on  the  various 
aspects    of   experimental    design 

*  presentation    of    rigorously    de 
veloped  procedures  in  the  form 
best  suited  for  computation 

*  inclusion  of  a  number  of  non-para 
metric  techniques  in  recognition 
of  the  growing  importance  of  this 
area  in  research 

*  a  separate  chapter  on  quality  con.- 
trol 

*  an  excellent  collection  of  impor 
tant  and  useful  tables  with  stand 
ardized  formats 

*  statistical  inference  (both  estima 
tion  and  testing  hypotheses)  well 
organised,  thoroughly  discussed 

(Continued  on  beck  flap) 


OCT  8     1975 
NOY     4 


jMAj;  MAR  2  9  1976 

JUN  1     1976 


2     1976 

7  -  T976 

SEP  1 1 1978 


!MAI  JUL  14  1992" 


If  III      f      j 

°°°28  0651 


519-9  085s2 
Ostle 

Statistics  in  research 


63-09055 


519.9  085s2 
Ostle     $10.50 
Statistics  in  research 


63-09055 


•M|I»J.«.I..    ,  », 


STATISTICS  IN  RESEARCH 


i  . 


-1" 


r 


SECOND  EDITION 

STATISTICS   IN   RESEARCH 

BASIC  CONCEPTS  AND  TECHNIQUES  FOR  RESEARCH  WORKERS 


BERNARD  OSTLE 

PROFESSOR    OF    ENGINEERING,    ARIZONA    STATE   UNIVERSITY 


THE   IOWA  STATE   UNIVERSITY  PRESS 

_xVm#4,    IOWA,    U.S.A. 


About  the  author  .  .  . 

BERNARD  OSTLES,  Professor  of  ^Engineering  at  Arizona  State  University,  includes 
operations  research,  quality  control,  reliability,  the  statistical  design  of  experi 
ments,  and  the  statistical  analysis  of  data  in  his  principal  areas  of  teaching 
and  research.  He  also  serves  as  special  consultant  for  several  industrial  firms- 
EJarlier,  Dr.  Gstle  was  Supervisor  of  the  Statistics  Section  in  the  Reliability 
Department  of  Sandia  Corporation,  Albuquerque,  New  Mexico.  From  1952  to 
J957  he  was  Profossor  of  Mathematics,  Agricultural  Experiment  Station  Statisti 
cian,  and  Director  of  the  Statistical  Laboratory  at  Montana  State  College. 
Prior  to  that  he  taught  statistics  at  Iowa  State  University  and  the  University  of 
Minnesota.  In  addition,  he  served  as  Special  Lecturer  in  Statistics  at  the  Uni 
versity  of  New  Mexico,  1958-60,  and  as  Special  Lecturer  in  Statistics  for  the 
National  Science  Foundation  Summer  Institute  at  Oklahoma  State  University 
in  1959. 

He  received  his  B.A.  degree  from  the  University  of  British  Columbia  in  1945 
with  honors  in  mathematics,  and  the  M.A.  degree  In  economics  from  the  same 
institution  in  194(5.  He  wan  awarded  the  Ph.D.  degree  in  statistics  at  Iowa  State 
University  iix  1949.  He  IB  a  member  of  the  American  Society  for  Quality  Control, 
the  American  Statistical  Association,  the  Operations  Tlosearch  Society  of 
America,  Alpha  Pi  Mu,  Phi  Kappa  Phi,  Pi  Mu  Kpsilon,  and  Sigma  Xi.  Dr.  Ostle 
alno  is  the  author  of  numerous  artiolos  in  various  scientific  journals. 


©  1963  The  Iowa  State  University  Press. 
All  rights  rofiorvod. 
Manufactured  in  U.S.A. 


Library  of  Congress  catalog  card  number:  03-7548. 


To 
RUTH  JEAN 


KANSAS  cny  cm.)  puer  re 

I   O  ,  £,    O  630905 


THE  PAST  FEW  YEARS  have  seen  many  changes  in  the  science  of  statis 
tics.  New  techniques  of  analysis  and  inference  have  been  developed 
by  mathematical  statisticians  while  applied  statisticians  have  been 
busily  engaged  in  applying  both  old  and  new  techniques  in  novel 
situations.  These  facts,  plus  the  tendency  toward  an  increased  use  of 
mathematics  in  research,  indicated  that  a  revision  of  this  text  was 
needed. 

This  revised  edition  has  been  prepared  with  the  following  goals  in 
view:  (1)  to  provide  a  book  giving  those  statistical  methods  that  have 
been  found  useful  by  workers  in  many  areas  of  scientific  research,  (2) 
to  present  these  methods  as  integral  parts  of  a  complete  discipline,  and 
(3)  to  provide  a  textbook  that  will  facilitate  teaching  the  science  of 
statistics. 

In  attempting  to  achieve  the  above  goals,  I  have  taken  the  position 
that  it  is  possible  to  write  a  book  which  will  prove  acceptable  to  stu 
dents,  teachers,  and  research  workers  in  many  fields  of  specialization. 
Consequently,  this  book  presents  the  techniques  of  modern  statistics  as 
statistical  methods  per  se*  To  demonstrate  the  universality  of  statistical 
methods,  many  examples  from  varied  fields  of  application  ai*e  included. 

In  this  book  considerable  attention  has  been  given  to  the  assumptions 
underlying  the  techniques  presented.  Withoxit  a  thorough  understand 
ing  of  the  limitations  of  various  techniques,  one  might  apply  them  in 
situations  where  they  should  not  be  used.  The  learning  of  methods  is 
easy;  learning  when  and  where  to  use  them  is  not  so  easy.  I  have  at 
tempted  to  achieve  a  reasonable  balance  between  these  two  ends. 

This  edition  has  been  designed  in  such  a  manner  that  it  should  prove 
xiseful  for  several  purposes,  namely:  (1)  as  a  text  for  a  standard  course 
in  statistical  methods,  (2)  as  a  text  for  an  integrated  course  in  theory 
and  methods  for  students  in  engineering  and  the  physical  sciences,  and 
(3)  as  a  reference  book  for  research  workers  and  other  users  of  statistical 
methods,  whether  they  be  affiliated  with  government,  industry,  re 
search  institutes,  or  universities. 

Because  of  the  multiple-purpose  design  of  the  book,  some  topics  will 
be  of  interest  only  to  special  groups.  In  addition,  changes  in  order  of 
presentation  by  individual  teachers  will  also  occur.  However,  it  is  my 
belief  that  a  reasonable  compromise  among  descriptive  statistics, 
mathematical  statistics,  statistical  methods,  and  the  design  and  analy 
sis  of  experiments  has  been  achieved,  and  that  the  book  will  prove 
suitable  for  all  the  purposes  for  which  it  was  planned. 

I  am  indebted  to  Sir  Ronald  A.  Fisher,  to  Dr.  Frank  Yates,  and 
to  Oliver  and  Boyd  Ltd.,  Edinburgh,  for  permission,  to  reprint  Table  III 
from  Statistical  Tables  for  Biological,  Agricultural  and  Medical  Research. 
I  am  also  indebted  to  Dr.  O.  L,  Davies  and  to  Oliver  and  Boyd  Ltd., 

CvIII 


viii  PREFACE 

Edinburgh,  for  permission  to  reprint  Table  6.G1  from,  the  second 
edition  of  Statistical  Methods  in  Research  and  Production,  and  Tables 
7.7,  7.72,  E,  E.I,  G,  and  H  from  the  second  edition  of  The  Design  and 
Analysis  of  Industrial  Experiments.  Many  other  persons  have  also 
graciously  given  permission  for  the  reproduction  of  published  material, 
and  acknowledgment  has  been  made  at  the  appropriate  places  in  the 
text.  The  author  is  deeply  appreciative  and  wishes  to  express  his 
thanks  for  their  cooperation. 

Acknowledgment  is  due  to  Paul  G.  Homeyer,  David  V.  Huntsberger, 
Bmil  H.  Jebe,  Oscar  Kempthorne,  and  George  W.  Snedecor  for  their 
encouragement  during  the  preparation  of  the  first  edition.  To  my 
former  co-workers  at  Sandia  Corporation,  Albuquerque,  New  Mexico, 
and  in  particular  to  John  M.  Wiesen,  I  also  wish  to  express  my  appreci 
ation.  The  suggestions  which  resulted  from  our  many  stimulating  con 
versations  contributed  greatly  to  the  improvement  of  this  new  edition. 

My  greatest  personal  indebtedness  is  to  my  wife,  Ruth  Jean  Ostle, 
without  whose  help  this  revision  would  still  be  far  from  complete.  In 
particular,  I  wish  to  thank  her  publicly  for  her  diligence  in  typing  the 
manuscript,  for  her  editorial  assistance,  and  for  her  tmfailing  patience 
and  understanding  during  the  entire  project. 


BERNABB  OST&E 
Tempe,  Arizona 


TABLE  OF  CONTENTS 

1.  THE  ROLE   OF   STATISTICS   IN  RESEARCH 

1.1  The  Nature  and  Purpose  of  Research 1 

1.2  Research  and  Scientific  Method 2 

1.3  What  Is  Statistics? 2 

1.4  Statistics  and  Research 3 

1.5  Further  Remarks  on  Science,  Scientific  Method,  and  Statistics.  .  3 

1.6  Applications  of  Statistics  in  Research 5 

1.7  Summary 12 

Problems 14 

References  and  Further  Reading 14 

2.  MATHEMATICAL   CONCEPTS 

Set  Theory 17 

Notation 18 

Permutations  and  Combinations 20 

Some  Useful  Identities  and  Series 20 

Some  Important  Functions 21 

Matrices , 22 

Linear  Equations 24 

Problems 25 

References  and  Further  Reading 28 

3.  A  SUMMARY  OF  BASIC  THEORY  IN  PROBABILITY  AND  STATISTICS 

Probability 29 

Mathematical  Expectation 33 

Probability  Distributions 33 

Expected  Values 35 

Other  Descriptive  Measures 36 

Special  Probability  Distributions 37 

Problems 37 

References  and  Further  Reading 43 

4.  ELEMENTS   OF  SAMPLING  AND   DESCRIPTIVE   STATISTICS 

4.1  The  Population  and  the  Sample 44 

4.2  Types  of  Samples 45 

4.3  Sampling  From  a  Specified  Population 46 

4.4  Presentation  of  Data , 47 

4.5  Calculation  of  Sample  Statistics 52 

4.6  The  Arithmetic  Mean 53 

4.7  The  Midrange 55 

4.8  The  Median 55 

4.9  Percentile,  Decile,  and  Quartile  Limits 56 

4.10  The  Mode 58 

4.11  The  Hange 60 

4.12  The  Standard  Deviation  and  Variance 60 

4.13  The  Coefficient  of  Variation 64 

4.14  Summary 65 

Problems 66 

Uxl 


x  CONTENTS 

5.  SAMPLING  DISTRIBUTIONS 

5.1  Sample  Moments 70 

5.2  Variance  of  the  Sample  Moan 70 

5.3  TchebyohefTPs  Inequality 71 

5.4  Law  of  Large  Numbers 72 

5.5  Central  Limit  Theorem 72 

5.6  Random  Sampling  From  a  Specified  Population 73 

5.7  The  Hypergeomotrie  Distribution , 73 

5.8  The  Binomial  Distribution 74 

5.9  Binomial  Approximation  to  the  Hypergeo metric 74 

5.10  Poisson  Approximation  to  the  Binomial 75 

5.11  Normal  Approximation  to  the  Binomial 76 

5.12  The  Multinomial  Distribution 78 

5.13  The    Negative    Binomial    Distribution    and    the    Geometric    Dis 
tribution 79 

5.14  Distribution,  of  a.  Linear  Combination  of  Normally   Distributed 
Variables 80 

5.15  Distribution  of  the  Sample  Moan  for  Normal  Populations 80 

5,1(5    Distribution  of  the  Difference  of  Two  Sample  Moan** 81 

5.17  Chi-Square  Distribution 81 

5.18  Distribution  of  the  Sum  of  Squares  of  Independent  Standard  Nor 
mal  Variates . 82 

5.19  Distributions  of  the  Sample  Variance  ami  Standard  Deviation  for 
Normal  Population** , 82 

5.20  Distribution  of  "StxulontV  t 83 

5.21  Distribution  of  F .  .  ,  . 83 

5.22  Order  Statistics 84 

Problems 85 

References  and  Further  Rending 86 

6.  STATISTICAL  INFERENCE;   ESTIMATION 

6,1      Some  Preliminary  Idean 87 

l       0.2      Methods  of  Obtaining  Point  Kntimatorw 88 

6.3  Maximum  Likelihood  Katimutorw 89 

6.4  Confidence  Intervals:  Crtmenil  DIHCUHHIOII , 89 

(5.5      Confidence  Interval  for  the  Mean  of  a  Normal  Population  ,  ,  .  .  .  00 

6.6  Confidence  Interval  for  the  Moan  of  a  Nonnormal  Population  .  .  92 

6.7  Confidence*  Interval  for  the  Variance  of  a  Normal  Population.  ,  .  03 

6.8  Confidence)  Interval  for  p,  the  Parameter  of  a  Binomial  Population  94 

6.0  Confidence  Interval  for  the  Difference  Between  the  Means*  of  Two 
Normal  Populations , , ,  .  ,  95 

6.10  Cofiderico  Interval  for  the  Ratio  of  th&  Variances  of  Two  Normal 
Populations , 97 

6.11  Tolerance  Limitn:  General  DuicuaHum , ...,,.,  98 

6.12  Tolerance   Limits   (Two-Bided;   One-Sided)    for   Normal    Popula 
tions  ..,.,.....,. ..,,>,...  98 

6.13  Distribution- Free  Tolerance*  LimitH .  .  100 

Prohlemn .  . , 101 

Reference**  and  Further  Heading ,  .  , 105 

7.  STATISTICAL  INFERENCE:   TESTING   HYPOTHESES 

7.1  <  Jeneral  OonHidorntumH ,,,,.,,,...  107 

7.2  KHtnhliHhment  of  Tewt  Procedure**.  , , *  111 


CONTENTS  xi 

7.3  Normal  Population;  H'.jm  —  juiQ  Versus  A:/A^pt0  .................  113 

7.4  Normal  Population;  ^IM^/XO  Versus  A:A*>MOJ  or  H:JJL>JULO  Versus 

A  :M<MO  .................................................  113 

7.5  Normal  Population;  H:<r2=<rl  Versus  A:a-2^o-Q  ................  114 


7.6  Normal    Population;     H:<72<o-o    Versus    A:<T*>CT*,     or    Hi<rz><rl 
Versus  A  :o-2<o-Q  ..........................................  115 

7.7  Binomial  Population;  H:p~pQ  Versus  A:p^pQ  ...............  115 

7.8  Binomial    Population;    H:p<.p0    Versus    A:p>pQ,     or    ff-p>po 
Versus  A  :p  <p0  ...........................................  118 

7.9  Two  Normal  Populations;  j£T  :  /-n  =  >U2  Versus  A'.J^I^JJ.^  ..........  119 

7.10  Two  Normal  Populations;  £T:/zi<M2  Versus  A:/xi>M2,  or  H  :jun  >^t2 
Versus  A  :/xi  <At2  ..........................................  122 

7.11  Two  Normal  Populations,  H  i&l  =<T%  Versus  A  '.a\^<r\  ..........  123 

7.12  Two  Normal  Populations;  H'.a\  <cr~  Versus  A  iv\  ><r%,  or  H:crl>cr^ 
Versus    Ai<r\<<ri  ..........................................  123 

7.13  Multinomial  Data  .........................................  124 

7.14  Poisson  Data  .............................................  124 

7.15  Chi-Square  Test  of  Goodness  of  Fit  .........................  126 

7.16  Binomial  Population;  More  Than  One  Sample  ................  128 

7.17  Contingency  Tables  .......................................  129 

7.18  Special  Approximate  Methods  for  2X2  Tables  ................  131 

7.19  The  Exact  Method  for  2  X  2  Tables  ..........................  132 

7.20  Several  Normal  Populations;  H  :  MI  =  ju2  =    •  •  -    =Mfc  ............  133 

7.21  Several  Normal  Populations;  ff:crl~<r%~   •  •  -    =cr£  ............  136 

7.22  Sample  Size  ..............................................  136 

7.23  Sequential  Tests  ..........................................  140 

Problems  ................................................  143 

References  and  Further  Reading  ............................  157 

8.   REGRESSION   ANALYSIS 

8.1  Functional  Relations  Among  Variables  .......................  159 

8.2  A  Word  of  Caution  About  Fimctional  Relations  ..............  160 

8.3  The  Choice  of  a  Functional  Relation  ........................  160 

8.4  Curve  Fitting  ............................................  160 

8.5  The  Method  of  Least  Squares  ...............................  161 

8.6  Graphical  Interpretation  of  the  Method  of  Least  Squares  ......  162 

8.7  Simple,  Linear  Regression  ............  .  .....................  164 

8.8  Partitioning  the,  Sum  of  Squares  of  tho  Dependent  Variable,  ....  164 

8.9  A  Practical  .Example  ......................................  167 

8.10  AwHumptionH  Necessary  for  JBJstimation  and  Testing  Hypotheses  in 
Simple  Linear  Regression  ..................................  168 

8.11  K&timates   of   Krror   Associated   With   Simple   Linear   degression 
Analyses  .....  .  ...........................................  170 

8.12  Confidence  and  Prediction  Intervals  In  Simple  Linear  Regression  170 

8.13  Tests  of  Hypotheses  in  Simple  Linear  Regression  ..............  174 

8.14  Inverse  Prediction  in  Simple  Linear  Regression.  ,  .  .  .  .  .........  176 

8.35   The  Abbreviated  Doolittle  Method.  ...  ......................  177 

8.16  Some  Additional  Remarks  With  Regard  to  Generalized  Regression 
Analyses  .................................................  186 

8.17  Teats  for  Lack  of  Fit  ......................................  188 

8.18  Nonlinear  Models  .........................................  190 

8.19  Second  Order  Models  ......................................  191 

8-20   Orthogonal  Polynomials  ...................................  192 


xii  CONTENTS 

8.21  Simple  Exponential  Regression 194 

8.22  The  Special  Case:  77  =£.X" 196 

8.23  Weighted  Regressions 197 

8.24  Sampling  From  a  Bivariate  Normal  Popxilation 198 

8.25  Adjusted  F  Values 199 

8.26  The  Problem  of  Several  Samples  or  Groups 201 

8.27  Some  Uses  of  Regression  Analysis 205 

Problems 206 

References  and  Further  Reading 221 

9.    CORRELATION   ANALYSIS 

9.1  Measures  of  Association  , 222 

9.2  An  Intuitive  Approach  to  Correlation 222 

9.3  The  Correlation  Index 223 

9.4  Correlation  in  Simple  Linear  Regression 223 

9.5  Sampling  From  a  Bivariate  Normal  Population 225 

9.6  Correlation  in  Multiple  Linear  Regression 227 

9.7  The  Correlation  Ratio 229 

9.8  Biaerial  Correlation 231 

9.9  Tetrachoric  Correlation 232 

9.10  Coefficient  of  Contingency ,  .  .* , 232 

9.11  Rank  Correlation 233 

9.12  IntraehusH  Correlation 235 

9.13  Correlations  of  Sums  and  Differences ,....,...  238 

Problems , 239 

References  and  Further  Reading , 243 

10.  DESIGN   OF  BXPERIMBNTAL  INVESTIGATIONS 

10»  1     Some  General  Remarks , 244 

10.2  What  IB  Moant  by  "The  Design  of  an  Experiment"? ,  .  244 

10.3  The  Need  for  an  Experimental  Dcmgn, 244 

10.4  The  Purpose  of  an  Kxporimental  Design 245 

10*5     Battle  Principles  of  Experimental  Demgn ,  .  .  ,  .  246 

10.6  Replication 246 

10.7  Experimental  Krror  and  Experimental  Units 247 

10.8  Confounding , , 248 

10*9     Randomisation *.*«,*. ....»..,,,.,,.  249 

10.10   Local  Control .  .  .  „ „ « 250 

10*11    Balancing,  Blocking,  and  Grouping.  I.,......,............*..  251 

10,12  Treatments  and  Treatment  Combinations *  .  , ,  4  .  252 

10. IS  Factors,  Factor  levels,  and  Factorials. t  ,  .  .  253 

10,14   Effects  and  Interactions ..«,.... 256 

10*35  Treatment  Comparisons.  k  * .  ,  * 261 

10.10  Htepa  in  Designing  an  Kxperimcmt.  .  * . .  - *  . 264 

10.17  IlhitftrutioziH  of  the  Statistician's  Approach  to  !>emgn  Problem**.  ,  266 

10.18  Advantagen  and  Disadvantages  of  Statistically  Designed  Kxperi- 
mont.H ,  .  , «  *  .  ,  27 1 

10.19  Hummary ,  , 273 

Problomn ,.,,., 273 

eeH  and  Further  Reading.  ,  ,  . ,  .  .  ,  275 


CONTENTS  xiii 

11.  COMPLETELY   RANDOMIZED    DESIGN 

11.1  Definition  of  a  Completely  Randomized  Design 278 

11.2  Completely    Randomized    Design    With    One    Observation    per 
Experimental  Unit 279 

11.3  The  Relation    Between   a    Completely   Randomized    Design   and 
"Student's"  Z-Test  of  JET:  AXIOMS  Versus  A: v\^m 288 

11.4  Subsampling  in  a  Completely  Randomized  Design 288 

11.5  Expected  Mean  Squares,  Components  of  Variance,  Variances  of 
Treatment  Means,  and  Relative  Efficiency 298 

11.6  Some  Remarks  Concerning  F-Ratios  That  Are  Less  Than  Unity  .  .  301 

11.7  Satterthwaite's  Approximate  Test  Procedure 302 

11.8  Selected  Treatment  Comparisons:  General  Discussion 303 

11.9  Selected  Treatment  Comparisons:  Orthogonal  and  TSTonorthogonal 
Contrasts 306 

11.10  All  Possible  Comparisons  Among  Treatment  Means 310 

11.11  Response   Curves:  A  Regression  Analysis  of  Treatment   Means 
When    the    Various    Treatments    Are    Different    Levels    of    One 
Quantitative  Factor 312 

11.12  Analysis  of  a  Completely  Randomized  Design  Involving  Factorial 
Treatment  Combinations 316 

11.13  Nonconformity  to  Assumed  Statistical  Models 338 

11.14  The    Relation    Between    Analysis    of    Variance    and    Regression 
Analysis 340 

11.15  Presentation  of  Results 341 

Problems 344 

References  and  Further  Reading 360 

12.  RANDOMIZED    COMPLETE   BLOCK   DESIGN 

12.1  Definition  of  a  Randomized  Complete  Block  Design 363 

12.2  Randomized  Complete  Block  Design  With  One  Observation  per 
Experimental  Unit 364 

12.3  The    Relation   Between  a  Randomized    Complete  Block  Design 
and  "Student's"  if-Test  of  H :  &&  **  0  When  Paired  Observations  Are 
Available 368 

12.4  Subaampling  in  a  Randomized  Complete  Block  Design 368 

12.5  Preliminary  Testa  of  Significance 371 

12.6  legitimation,  of  Components  of  Variance  and  Relative  Efficiency  .  373 

12.7  Efficiency  of  a  Randomized  Complete  Block  Design  Relative  to  a 
Completely  Randomized  Design 374 

12.8  Selected  Treatment  Comparisons 376 

12.9  Subdivision  of  the   Experimental   Error  Sxim.  of  Squares   When 
Considering  Selected  Treatment  Comparisons 376 

12.10  All  Possible  Comparisons  Among  Treatment  Means 380 

12.11  Response  Curves  in  a  Randomized  Complete  Block  Design.  ....  380 

12.12  Factorial  Treatment  Combinations  in  a  Randomized  Complete 
Block  Design 380 

12.13  Missing  Data  in  a  Randomized  Complete  Block  Design. . 390 

Problems 394 

References  and  Further  Reading 408 

13.  OTHER  DESIGNS 

13,1      Latin  arid  Graeco-Latirj.  Squares 410 


xiv  CONTENTS 

13.2  Split  Plots 415 

13.3  Complete  Factorials  Without  Replication,  Fractional  Factorials, 

and  Incomplete  Blocks 417 

13.4  Unequal  but  Proportionate  Subclass  Nximbers 421 

13.5  Unequal  arid  Disproportionate  Subclass  Numbers 423 

13.6  Response  Surface  Techniques 424 

13.7  Random  Balance 425 

13.8  Other  Designs  and  Techniques 425 

Problems 426 

References  and  Further  Reading 434 

14,  ANALYSIS    OF   COVARIANCE 

14.1  Uses  of  Oovnrianec  Analysis. 437 

14.2  Assumptions   Underlying  Analyses  of  Oovarianee 43S 

14.3  Completely  Randomized  Design 439 

14.4  Randomized  Complete  Block  Design 444 

14.5  Latin  Square  Design 449 

14.6  Two-Factor  Factorial  in  a  Randomized  Complete  Block  Design .  .  452 

14.7  Covariaixce  When  the  X  Variable  Is  AfTeeted  by  the  Treatments  .  .  456 

14.8  Multiple  Co  variance , , 457 

Problems , 460 

References  and  Further  Rending , 465 

15.  DISTRIBtJTI ON-FREE   METHODS 

35.1  DiHtributitm-Froe  Methodw  Included  in  Previous  Chapters 466 

15.2  The  Sign  Tout - 466 

15.3  The  Signed  Rank  Tost 468 

15.4  The  Rxm  Tewt , 470 

15.5  The  Kolmogorov-Smirnov  Tent  of  Goodno.wfl  of  Fit 471 

15.6  Median  Testa 473 

Problems , , 473 

References  and  Further  Reading 474 

16.  STATISTICAL    QUALITY  CONTROL 

16.1  Control  Charts 477 

16.2  Acceptance  Sampling  Plans ,  ,  ,  .  ,  4<S5 

Problems .  . 49 1 

Roferouees  und  Further  Reading. ,...,,...,...  408 

17,  SOME   OTHE&  TECHNIQUES   AND   APPLICATIONS 

17.1  Some  Pseudo  t  Statistics ,.,..,.,,,..,,, 500 

17.2  A  PHoudo  F  Statistic .  .  ,  , , 501 

J 7.3      Kvolutionary  Operation. * .,.,»..  501 

1 7.4  TolerancoH * 502 

17.5  The  Entimatum  of  8y«tem  Reliability, ...,.,..., 506 

Problems . , 508 

Reference  and  Further  Heading*  .  ,  , .....,, 510 

APPENDIX 

I .    ( Jreek  Alphabet .  ,  ,  ,  , ,  .  .  .  511 

iJ,   Cumulative  PoiHwon  Dmtributicm 512 

3.   CUimulativ<»  Standard  Normal  Difttrihution.  ,  .  ...,..,, ,  .  ,  ,  ,  517 


CONTENTS  xv 

4.  Cumulative  Chi-Square  Distribution 523 

5.  Cumulative  ^-Distribution 528 

6.  Cumulative  F- Distribution 529 

7.  Random  Numbers 544 

S.    Control  Chart  Constants 548 

9.    Number  of  Observations  for  £-Test  of  Mean 550 

10.  Number    of    Observations    for    £-Test    of    Difference    Between    Two 
Means 552 

11.  Number  of  Observations  Required  for  the  Comparison  of  a  Popula 
tion  Variance  With  a  Standard  Value  Using  the  %2-Test 554 

12.  Number    of    Observations    Required    for    the    Comparison    of    Two 
Population  Variances  Using  the  F-Test, 555 

13.  Critical  Values  of  r  for  the  Sign  Test 556 

14.  Table  of  Critical  Values  of  T  in  the  Wilcoxon  Signed  Rank  Test ....  557 

15.  Table  of  Critical  Values  of  r  in  the  Run  Test 558 

10.    Table  of  Critical  Values  of  D  in  the  Kolmogorov-Smirnov  Goodness 

of  Fit  Test 560 

17.    Percentage  Points  of  Pseudo  t  and  F  Statistics 561 

INDEX 565 


CH APTE  R  1 

THE  ROLE  OF  STATISTICS 
IN  RESEARCH 

EVERY  DAY  each  of  us  engages  in  some  observation  in  which  statistics 
is  used.  Such  common,  events  as  noting  the  weather  forecast,  weighing 
oneself,  checking  the  position  of  a  favorite  ball  team  in  its  league,  or 
testing  a  new  food  product  are  typical.  The  element  of  statistics  creeps 
in  when  you  mentally  evaluate  your  research.  In  weighing  yourself,  you 
automatically  compare  your  observation  with  your  average  weight 
(deviation  from  the  mean)  and  conclude  the  present  weight  is  usual 
(no  significance  to  the  difference)  or  unusual  (a  significant  difference) , 
basing  your  judgment  upon  previous  measurements  of  your  weight  and 
your  knowledge  of  the  variation  generally  observed.  These  common 
results  are  easily  obtained,  are  of  only  local  importance,  and  are  soon 
forgotten.  However,  the  formal  research  which  means  so  much  to 
improving  man's  lot  is  of  infinitely  greater  importance  and  must  be 
conducted  with  much  greater  care.  It  is  with  the  latter  type  of  research 
that  this  book  is  concerned. 

1.1      THE  NATURE  AND   PURPOSE  OF  RESEARCH 

Research,  according  to  Webster,  is  studious  inquiry  or  examination — 
critical  and  exhaustive  investigation  or  experimentation  having  for  its 
aim  the  discovery  of  new  facts  and  their  correct  interpretation.  It  also 
aims  at  revising  accepted  conclusions,  theories,  or  laws  in  the  light  of 
newly  discovered  facts  or  the  practical  applications  of  such  new  or 
revised  conclusions.  Research,  therefore,  means  continued  search  for 
knowledge  and  understanding;  scientific  research  is  continued  research 
using  scientific  methods.  Scientific  research  is  essentially  compounded 
of  two  elements:  observation,  by  which  knowledge  of  certain  facts  is 
obtained  through  sense-perception;  and  reasoning,  by  which  the  mean 
ing  of  these  facts,  their  interrelation,  and  their  relation  to  the  existing 
body  of  knowledge  are  ascertained  insofar  as  the  present  state  of 
knowledge  and  the  investigator's  ability  permit. 

In  any  discussion  of  research,  two  important  facts  should  be  noted. 
They  are :  (1)  there  is  an  ever  increasing  trend  towards  extreme  speciali 
zation  on  the  part  of  individual  scientists,  and  (2)  most  research  prob 
lems  are  such  that  many  disciplines  and  fields  of  specialization  can 
contribute  in  a  significant  manner  to  their  solutions.  Thus,  it  is  evident 
that  more  and  more  research  will  be  handled  on  an  interdisciplinary 
team  basis  rather  than  by  individual  scientists  working  in  "solitary 
confinement/'  (NOTE:  This  is  not  to  say  that  very  little  individual 

til 


2  CHAPTER    1,    THE    ROLE    OF    STATISTICS    IN    RESEARCH 

research  will  continue  to  be  performed.  Sucli  research  always  should 
and  always  will  be  performed.  The  statement  was  meant  to  imply 
only  that  research  is  becoming  predominantly  a  team  or  cooperative 
effort.) 

In  summary,  research  is  an  inquiry  into  the  nature  of,  the  reasons  for, 
and  the  consequences  of  any  particular  set  of  circumstances- — whether 
these  circumstances  are  experimentally  cozitrolled  or  recorded  just  as 
they  occur.  Further,  research  implies  the  researcher  is  interested  in 
more  than  particular  results — he  is  interested  in  the  repeatability  of 
the  results  and  in  their  extension  to  more  complicated  and  general 
situations. 

1.2  RESEARCH   AND  SCIENTIFIC   METHOD 

Although  the  techniques  of  investigation  may  vary  considerably 
from  one  science  to  another,  the  philosophy  common  to  all  is  generally 
referred  to  as  scientific  method.  There  are,  perhaps,  as  many  definitions 
of  scientific  method  as  there  are  workers  in  research.  For  our  purposes, 
the  following  will  be  used:  Since  the  ideal  of  science  is  to  achieve  a  systcm- 
atic  interrelation  of  facts }  scientific  method  must  be  a  pursuit  of  tkzs  ideal 
by  experimentation,  observation,  logical  arguments  from  accepted  postu 
lates  r  and  a  combination  of  these  three  in  varying  proportions.  Therefore, 
research  and  scientific  method  are  closely  related,  if  not  one  and  the 
same  thing. 

1.3  WHAT   IS  STATISTICS? 

Statistics  has  often  been  classified  as  a  method  of  research  along 
with,  or  in  opposition  to,  .such  methods  as  case  studios,  the  historical 
approach,  and  the  experimental  method.  Since  this  classification  fre 
quently  leads  to  confused  and  incorrect  thinking,  it  is  not  wise.  It  is 
better  to  regard  statistics  as  supplying  a  kit  of  tools  which  can  be 
extremely  valuable  in  research.  This  book  will  stress  gaining  an  tinder- 
standing  of  these  tools  and  learning  which  tool  should  be  xised  in  vari 
ous  situations  arising  in  scientific  research.  Only  when  you  know  which 
tool  to  xiso,  how  to  use  it,  and  how  to  interpret  your  results  can  you 
hope  to  do  productive  research.  To  summarise:  the  science  of  statistics 
has  mxieh  to  offer  the  research  worker  in  planning,  analyzing,  and  inter 
preting  the  results  of  his  investigations,  and  tliis  book  is  devoted  to 
an  exposition  of  those  methods  and  techniques  that  have  proved  useful 
in  many  fields  of  inquiry, 

As  is  the  case  with  many  words  in  the  English  language,  the  word 
statistics  is  used  in  a  variety  of  ways,  each  correct  in  its  own  sphere. 
In  the  plural  sense,  it  is  usually  taken  to  be  synonymous  with  data. 
However,  to  the  statistician,  there  is  another  meaning  of  the*  word. 
This  moaning  is  the  plural  of  the  word  statistic,  which  refers  to  a  quan 
tity  calculated  from  sample  observations*  (These  terms  will  be  defined 
in  considerable  detail  in  later  chapters.)  In  the  singular  sense,  statistics 
is  a  science,  and  it  is  in  this  sense  the  word  will  be  employed  most  fro- 


quently  in  this  book.  The  science  of  statistics  deals  with: 

(1)  Collecting  and  summarizing  data. 

(2)  Designing  experiments  and  surveys. 

(3)  Measuring  the  magnitude  of  variation  in  both  experimental 
and  survey  data. 

(4)  Estimating    population    parameters    and    providing    various 
measures  of  the  accuracy  and  precision  of  these  estimates. 

(5)  Testing  hypotheses  about  populations. 

(6)  Studying  relationships  among  two  or  more  variables. 

1.4  STATISTICS  AND   RESEARCH 

As  indicated  in  the  preceding  section,  statistics  enters  into  research 
and/or  scientific  method  through  experimentation  and  observation. 
That  is,  experimental  and  survey  investigations  are  integral  parts  of 
scientific  method,  and  these  procedures  invariably  lead  to  the  use  of 
statistical  techniques.  Since  statistics,  when  properly  used,  makes  for 
more  efficient  research,  it  is  recommended  that  all  researchers  become 
familiar  with  the  basic  concepts  and  techniques  of  this  useful  science. 

Because  statistics  is  such  a  valuable  tool  for  the  researcher,  it  some 
times  gets  overworked.  That  is,  there  are  many  cases  where  statistics 
is  used  as  a  crutch  for  poorly  conceived  and/or  executed  research.  In 
addition,  there  are  cases  in  which  statistics  is  employed  in  good  faith 
but,  unfortunately,  insufficient  attention  is  paid  to  the  assumptions 
required  for  a  valid  use  of  the  methods  employed.  For  these  and  other 
reasons,  it  is  essential  that  the  user  of  statistics  clearly  understands  the 
techniques  he  employs.  Consequently,  in  this  book  careful  attention 
will  be  given  to  both  the  methods  and  the  underlying  assumptions  in 
the  hope  that  such  an  approach  will  lead  to  the  proper  application  and 
use  of  statistics  in  scientific  research. 

1.5  FURTHER      REMARKS     ON      SCIENCE,      SCIENTIFIC 
METHOD,  AND  STATISTICS 

In  the  preceding  sections,  your  attention  has  been  called  to  the  close 
connection  that  statistics  has  with  experimentation,  scientific  method, 
and  research.  However,  in  each  case,  the  discussion  was  quite  brief. 
Because  these  various  topics  and  their  interrelationships  are  so  impor 
tant  to  the  remainder  of  this  book,  a  few  additional  remarks  are  justi 
fied.  To  expedite  the  discussion,  the  following  questions  and  answers 
have  deon  devised: 

What  Is  Logic? 

Logic  deals  with  the  relation  of  implication  among  propositions,  that 
is,  the  relation  betweeii  premises  and  conclusions.  In  scientific  method, 
logic  aids  in  formulating  our  propositions  explicitly  and  accurately  so 
that  their  possible  alternatives  become  clear.  When  faced  with  alterna 
tive  hypotheses,  logic  develops  their  consequences  so  that  when  these 


4  CHAPTER    1,    THE    ROLE    OF    STATISTICS    IN    RESEARCH 

consequences  are  compared  with  observable  phenomena  we  have  a 
means  of  testing  which  hypotheses  are  to  be  eliminated  and  which 
one  is  most  in  harmony  with  the  observed  facts. 

What  Is  Science? 

Science  is  knowledge  which  is  general  and  systematic — knowledge 
from  which  specific  propositions  are  deduced  in  accordance  with  a  few 
general  principles.  Although  all  the  sciences  differ,  a  universal  feature 
is  "scientific  method/'  which  consists  of  searching  for  general  laws 
which  govern  behavior  and  of  asking  such  questions  as:  Is  it  so?  To 
what  extent  is  it  so?  Why  is  it  so?  What  general  conditions  or  consider 
ations  determine  it  to  be  so? 

What  Is  Scientific  Method? 

Scientific  method  is  the  pursuit  of  truth  as  determined  by  logical 
considerations.  The  ideal  of  science  is  to  achieve  a  systematic  interrela 
tion  of  facts;  scientific  method,  using  the  approach  of  "systematic 
doubt/'  attempts  to  discover  what  the  facts  really  are. 

What  Is  Experimentation? 

The  function  of  experimentation  is  the  elimination  of  untenable 
theories*  Experimentation  is  used  to  test  hypotheses  and  to  discover 
now  relationships  among  variables.  It  must  be  remembered,  however, 
that  no  hypothesis  which  states  a  general  proposition  can  be  demon 
strated  to  be  absolutely  true;  only  probable  inferences  are  possible. 

What  Part  Does  Experimentation  Play  in  Scientific  Method? 

Experimentation  is  only  a  means  toward  an  end.  It  is  a  tool  of  scien 
tific  method.  Conclusions  drawn  from  experimental  data  are  frequently 
criticized.  Such  criticisms  are  usually  based  on  one  or  more  of  the 
following  arguments:  (1)  the  interpretation  ia  faulty,  (2)  the  original 
aasumptiorm  are  faxilty,  or  (3)  the  experiment  waa  poorly  designed  or 
badly  executed.  Obviously,  careful  attention  should  be  given  to  the 
design  of  the  experiment  HO  that  the  procedures  used  are  both  valid  and 
efficient. 

What  Is  Experimental  Design? 

Experimental  design  is  the  plan  u^ed  in  experimentation.  It  involves 
the  assignment  of  treatments  to  the  experimental  units  and  a  thorough 
underst a nding  of  the  analysis  to  be  performed  when  the  data  become 
available,, 

What  Is  the  Relationship  Between  Statistics  and  Experi 
mental  Design? 

vStatintics  enters  into  experimental  design  because,  even  in  the  best 
planned  experiments,  one  cannot  control  all  the  factors  and  because 


1.6       APPLICATIONS    OF    STATISTICS    IN    RESEARCH  5 

one  wishes  to  make  inferences  based  on  the  observed  sample  data.  To 
be  of  any  practical  use,  these  uncertain  inferences  must  be  accompanied 
by  probability  statements  expressing  the  degree  of  confidence  which 
the  researcher  has  in  such  inferences.  To  make  certain  that  such  prob 
ability  statements  will  be  possible,  the  experiments  should  be  designed 
in  accordance  with  the  principles  of  the  science  of  statistics. 

1.6      APPLICATIONS  OF  STATISTICS    IN    RESEARCH 

Early  applications  of  statistics  were  mainly  concerned  with  reduc 
tion  of  large  amounts  of  observed  data  to  the  point  where  general 
trends  (if  they  existed)  became  apparent.  At  the  same  time,  emphasis 
in  many  sciences  turned  from  the  study  of  individuals  to  the  study  of 
the  behavior  of  aggregates  of  individuals.  Statistical  methods  were  ad 
mirably  suited  to  such  studies,  aggregate  data  fitting  consistently  with 
the  concept  of  a  population. 

The  next  major  development  in  statistics  arose  to  meet  the  need  for 
improved  analytical  tools  in  the  agricultural  and  biological  sciences. 
Better  analytical  tools  were  needed  to  improve  the  process  of  interpre 
tation  of,  and  generalization  from,  sample  data.  For  example,  the 
farmer  is  faced  with  the  task  of  maintaining  a  high  level  of  produc 
tivity  of  field  crops.  To  aid  him,  the  agronomist  conducts  an  endless 
number  of  experiments  to  determine  differences  among  yields  of  various 
crop  varieties,  effects  of  various  fertilizers,  and  the  best  methods  of 
cultivation.  On  the  basis  of  the  results  of  his  experiments,  he  is  ex 
pected  to  make  accurate  and  useful  recommendations  to  the  farm 
operator.  Clearly  then,  statistics,  being  a  science  of  inductive  inference 
using  probabilistic  methods,  should  be  of  great  value  to  the  researcher 
in  agronomy. 

In  early  agronomic  experimentation,  in  order  to  compare  a  number 
of  fertilizers,  it  was  thought  necessary  to  devote  only  a  single  plot  to 
each  treatment  and  determine  yields  in  order  to  arrive  at  valid  con 
clusions  concerning  relative  values  of  the  treatments.  However,  the 
agronomists  soon  found  that  the  yields  of  a  series  of  plots  treated  alike 
differed  greatly  among  themselves,  even  when  soil  conditions  appeared 
uniform  and  experimental  conditions  were  carefully  designed  to  reduce 
errors  in  harvesting.  For  this  reason,  it  became  necessary  to  find  some 
mearivS  for  determining  whether  differences  in  yields  were  due  to  dif 
ferences  in  treatments  or  to  uncontrollable  factors  which  also  con- 
tribxite  to  the  variability  of  plot  yields.  Statistical  methods  were  ap 
plied,  and  their  value  in  scientific  investigation  of  agronomic  practices 
was  soon  proved. 

Closely  related  to  agronomy  is  the  science  of  plant  breeding.  The 
ultimate  objective  of  any  plant-breeding  research  program  is  the  de 
velopment  of  improved  varieties  or  hybrids.  A  variety  may  be  im 
proved  in  many  possible  ways,  e.g.,  in  ability  to  use  plant  nutrients, 
in  disease  or  insect  resistance,  in  cold  tolerance,  or  in  its  suitability  to 
the  needs  or  fancies  of  the  grower  and/or  consumer.  Plants  are  organ- 


6  CHAPTER    1,    THE    ROLE    OF    STATISTICS    IN    RESEARCH 

isms  conditioned  by  genetic  factors  and  by  the  environment  in  which 
they  grow.  The  plant  breeder,  therefore,  utilizes  the  principles  of  genet 
ics  in  attempting  to  improve  inheritable  characteristics  of  plant  va 
rieties,  just  as  the  producer  attempts  to  obtain  high  production  by 
maintaining  a  favorable  environment.  However,  results  of  past  genetic 
studies  do  not  provide  all  the  answers  relative  to  the  inheritance  of 
plant  characteristics.  Thus,  plant  breeders  continually  carry  out  basic 
genetic  research  in  each  crop— along  with  practical  plant-breeding  pro 
cedures — in  order  to  ensure  future  progress. 

Development  of  a  superior  new  variety  by  hybridization  is  seldom 
a  haphazard  occurrence.  Usually  the  breeder  has  in  mind  the  charac 
teristics  desired  for  his  particular  purpose  or  area.  Growing  many  plant 
selections  to  decide  which  excel  in  a  quantitatively  inherited  character 
requires  growing  thorn  in  a  randomised,  replicated  field  design.  Choice 
of  design  depends  on  the  numbers  involved;  the  uniformity  of  the  soil; 
the  accxiracy  and  precision  of  the  particular  estimates  deemed  neces 
sary  to  get  the  desired  results;  the  time,  effort,  and  money  available; 
and  perhaps  other  factors.  The  data  collected  are  then  analyzed  in 
accordance  with  the  plan  of  the  experiment,  which  was  designed  to 
make  possible  proper  comparisons  among  the  strains  being  tested.  The 
statistical  methods  employed  must,  of  course,  have  a  logical  relation 
ship  to  the  biological  processes  under  consideration,  as  well  as  to  the 
way  in  which  the  experiment  was  conducted,  if  they  are  to  be  useful. 
After  the  data  have  been  analyzed  statistically,  the  results  must  be 
interpreted  in  view  of  the  assumptions  made  and  of  the  existing 
knowledge  so  that  some  conclusion  may  be  reached  with  regard  to 
accepting  or  rejecting  the  hypotheses  being  tested.  Selection  of  the 
strain  to  be  released  as  a  variety,  or  of  those  to  be  tested  further, 
may  then  bo  made  with  assxirance  that  the  decision  will,  in  all  likeli 
hood,  bo  a  reasonable  one. 

Other  research  areas  in  which  good  use  of  statistical  theories  and 
methods  is  made  are  poultry  breeding,  animal  breeding,  and  animal 
nutrition*  Poultry  brooding,  for  example,  is  concerned  with  the  raising 
of  more  efficient  and  more  productive  fowl.  Increased  egg  production, 
egg  sixe,  egg  color,  interior  egg  quality,  more  efficient  meat  production, 
long  life,  disease  resistance,  and  high  fertility  are  some  of  the  factors 
with  which  the  poultry  breeder  IH  actively  concerned.  If  a  statistically 
sound  research  program  is  adopted,  the  researcher  will  be  able  to  reach 
defensible  conclusions  and  bring  about  more  efficient  use  of  resources. 

One  of  the  more  important  uses  of  statistics  in  breeding  work  is  the 
separation  of  environmental  and  hereditary  effects*  The  literature  of 
the  field  i#  full  of  reports  dealing  with  this  type  of  research,  both  with 
poultry  and  domestic  animals.  For  the  reader  interested  in  this  par 
ticular  urea  of  research,  we  refer  to  such  writers  as  Ilutt  (25)  and 
Lush  (29),1 

1  Numborn  in  pnr<*nfrhonoH  designate  rofarfmoow  linted  at  mid  of  chapter. 


1.6       APPLICATIONS    OF    STATISTICS    IN    RESEARCH  7 

In  the  field  of  animal  nutrition,  many  experiments  have  been  devised 
to  discover  the  significance  of  various  vitamins  in  the  different  phases 
of  animal  production.  In  such  investigations,  several  groups  of  animals, 
as  homogeneous  as  possible,  are  selected  for  experimentation.  These 
homogeneous  groups  are  usually  formed  by  considering  such  criteria  as 
age,  weight,  sex?  heredity,  vigor,  and  previous  nutrition.  A  check  group 
is  chosen  and  fed  a  standard  ration.  The  other  groups  are  fed  different 
levels  of  the  vitamin  in  question,  one  of  them  on  a  ration  a  great  deal 
higher  than  the  standard  ration  for  the  vitamin  and  another  on  a  ration 
containing  little,  or  none,  of  the  vitamin.  The  remainder  of  the  groups 
are  fed  rations  somewhere  between  the  extremes.  The  animals  are  on 
the  randomly  assigned  rations  for  a  given  period  of  time,  and  the  re 
searcher  records  such  data  as  daily  gain  in  weight,  economy  of  gain, 
livability,  etc.  If  the  experiment  has  been  properly  designed  in  accord 
ance  with  established  statistical  principles,  conclusions  of  great  value 
to  the  farmer  may  then  be  drawn.  Of  course.,  much  work  of  a  more 
complex  nature  than  this  simple  example  has  also  been  done  in  animal 
nutrition  research.  Consultation  of  technical  journals  in  this  field  will 
reveal  many  instances  where  statistics  has  been  of  great  help. 

In  the  past,  many  persons  thought  statistics  had  no  place  in  the 
so-called  "exact  sciences'7  such  as  chemistry,  physics,  and  the  various 
branches  of  engineering.  These  fields  are  concerned  with  exact  measure 
ment,  with  quantities  that  can.  be  measured  with  a  ruler,  thermometer, 
flow  meter,  thickness  gauge,  telescope,  or  pressure  gauge.  Therefore, 
the  doubters  asked,  why  use  a  "pseudo-science'' — statistics — that  at 
best  merely  estimates  quantities?  As  the  true  meaning  of  statistics  and 
its  application  has  come  to  wider  attention,  these  persons  have  readily 
admitted  there  is  indeed  a  place  for  this  important  tool  in  the  exact 
sciences.  In  fact,  it  has  become  apparent  that  ail  of  these  sciences 
themselves  are  based  on  statistical  concepts.  For  example,  it  is  evident 
that  the  pressure  exerted  by  a  gas  is  actually  an  average  pressure — an 
average  effect  of  forces  exerted  by  individual  molecules  as  they  strike 
the  wall  of  a  container.  A  similar  situation  is  true  in  regard  to  tem 
perature. 

Since  the  popularly  accepted  theory  is  that  all  matter  is  made  up  of 
small  particles,  it  does  not  require  much  imagination  to  see  that  a 
statistical  approach  is  the  logical  one  to  adopt  in  investigations  of  the 
ultimate  nature  of  matter.  Such  particles  are  actually  part  of  an  almost 
inconceivably  large  population — one  that  is,  for  all  practical  pur 
poses,  our  closest  approach  to  the  infinite  population.  All  of  these 
particles  exhibit  individual  behavior  characteristics.  With  the  com 
paratively  crude  devices  of  the  exact  sciences  we  can  generally  only 
note  the  results  of  group  behavior — an  average  effect — and  until  re 
cently  science  has  been  limited  to  this.  But  even  in  these  crude  applica 
tions  statistics  plays  its  role.  For  instance,  examine  the  chart  of  the 
elements  in  any  chemistry  classroom.  The  atomic  weights  shown  on 
this  chart  are  actually  "weighted  averages'7  of  the  atomic  weights  of 


8  CHAPTER    1,    THE    ROLE    OF    STATISTICS    IN    RESEARCH 

individual  isotopes  of  the  given  element,  the  "weights"  being  the  fre 
quency  of  occurrence  of  the  element  in  a  normal  or  naturally  occurring 
mixture. 

Statistics  has  also  invaded  the  fields  of  meteorology  and  astronomy. 
The  modern  science  of  meteorology  is  to  a  great  degree  dependent  upon 
statistical  methods  for  its  existence.  The  methods  which  give  weather 
forecasting  the  accuracy  it  has  today  have  been  developed  using  mod 
ern  sample  survey  techniques.  Thus,  weather  stations  throughout  the 
United  States  are  able  to  give  us  highly  accurate  predictions  for  their 
individual  areas.  In  addition,  by  suitable  selection  of  gathering  points 
and  proper  treatment  of  the  data,  an  over-all  picture  of  the  weather 
for  larger  areas  is  pieced  together.  Again  we  may  see  statistical  sampling 
in  action  when  we  turn  our  attention  to  snow  survey  teams  which  de 
termine  the  amount  of  snow  present  in  a  given  area  and  thus  the 
quantity  of  water  to  be  drained  from  that  area  following  a  thaw.  In  the 
more  theoretical  aspects  of  meteorology,  statistical  inference  and 
analysis  are  being  xised  to  develop  new  techniques  for  advancing  the 
field.  In  astronomy,  statistics  havS  long  played  a  major  role.  One  hundred 
years  ago  the  uncertainty  in  the  measurement  of  the  semimajor  axis 
of  the  earth's  elliptical  orbit  was  1  part  in  20.  Today  statistical  methods 
have  reduced  this  uncertainty  to  1  part  in  10,000. 

Statistics  is  now  playing  an  important  role  in  engineering.  For 
example,  such  topic**  as  the  study  of  heat  transfer  through  insulating 
materials  per  unit  time,  performance  guarantee  testing  programs,  pro 
duction  control,  inventory  control,  standardisation  of  fits  and  toler 
ances  of  machine  parts,  job  analyses  of  technical  personnel,  studies  in 
volving  the  fatigue  of  metals  (endurance  properties),  corrosion  studies, 
time  and  motion  studies,  operations  research  and  analysis,  quality 
control,  reliability  analyses,  and  many  other  specialized  problem?*  in  re 
search  and  development  make  great  use  of  probabilistic*,  and  Btatintical 
methods* 

Because  the  above  problems  are  but  a  small  portion  of  those  to 
which  the  science  of  statistics  IB  being  applied  in  industry,  the  reader 
can  readily  appreciate  that  the  application  of  statistical  methods  to  the 
field  of  engineering  is  riot  limited  to  a  few  areas  but  is  general  in  nature. 
As  an  indication  of  the  wide  scope  of  industrial  statistics,  P.  L.  Algor 
of  the  General  Electric  Corporation  has  listed  the  following  ten  major 
areas  of  application: 

1*  Defining  the  value  of  observations 
2*  Design  of  experiments 

3.  Detection  of  causes 

4.  Production  quality  control 

5*  (letting  more  out  of  the  inspection  dollar 

6.  Design  specifications 

7.  Measurement  of  human  attributes 
<S,  Operational  research 

9.   Market  research,  including  opinion  polling 
10.    Determining  trends*3 


1.6       APPLICATIONS    OF    STATISTICS    IN    RESEARCH  9 

If  applied  statistics  is  to  play  a  primary  role  in  the  future  of  engi 
neering,  or,  to  be  more  general,  in  that  of  industry,  it  is  quite  evident 
that  there  is  a  great  need  for  specific  training  of  personnel  entering  the 
field.  This  training  is  needed  for  the  young  engineer  as  well  as  for  the 
young  businessman,  since  each  must  be  capable  of  dealing  with  combi 
nations  of  men  and  machines.  Professor  S.  S.  Wilks  of  Princeton  Uni 
versity  has  made  this  statement  of  the  problem : 

The  statistical  problems  which  the  future  scientist  or  engineer  will  en 
counter  will  cut  across  traditional  lines.  Therefore,  in  order  that  he  may  be 
properly  equipped  to  deal  with  these  problems,  he  should  have  a  fairly 
broad  statistical  training.  The  training  should  cover  not  only  statistical 
quality  control  methods  as  the  term  is  now  understood,  but  the  design  of 
experiments,  analysis  of  variance,  and  many  other  topics.  It  should  be  built 
into  the  training  of  scientists  and  engineers,  as  calculus  is  now  made  part 
of  their  basic  education.3 

Agricultural  engineering,  which  combines  the  practices  of  engineer 
ing  and  agriculture,  has  also  benefited  greatly  from  the  use  of  statisti 
cal  methods.  In  this  field,  statistics  has  helped  the  researcher  with  such 
varied  projects  as  the  testing  of  weed-control  machinery,  certain  eco 
nomic  aspects  of  farm  electrification,  comparison  of  various  drying 
methods  for  grain,  determination  of  the  effects  of  drying  rate  on  pop 
corn,  irrigation  research,  roofing  studies  for  farm  buildings,  and  meth 
ods  of  cultivation. 

Statistics  is  also  proving  an  important  tool  in  food  technology  re 
search.  Foods  exhibit  to  a  marked  degree  what  is  widely  called  "bio 
logical  variation."  Their  constitution  is  heterogeneous,  and  their  com 
plexity  is  such  that  duplication  is  highly  improbable.  Food  properties 
are  affected  not  only  by  the  multiplicity  of  factors  influencing  their 
growth  but  also  by  the  infinite  variety  of  processing  and  storage  con 
ditions  to  which  they  may  be  subjected.  Thus,  it  is  impossible  to  give  a 
general  answer  to  a  question  such  as  "What  is  the  moisture  content  of 
corn?"  Before  attempting  to  answer,  one  would  first  have  to  ask  "What 
variety  ...  at  what  stage  of  its  growth  or  processing  cycle  .  .  ,  where 
was  it  grown?"  and  such  questions.  Having  obtained  the  necessary 
specifications,  the  food  technologist  might  be  able  to  quote  an  average 
value*  In  short,  he  might  specify  a  frequency  distribution  of  moisture 
content  of  sweet  corn  under  the  stated  conditions. 

This  type  of  problem  was  encountered  by  Bard  (6)  in  his  investiga 
tion  of  certain  palatability  factors — tenderness,  juiciness,  and  fiber 
cohesivenevss— -of  canned  beef  as  conditions  of  time  and  temperature  of 
processing  were  varied.  In  his  work,  a  statistical  approach  dictated  the 
design  of  the  experiment,  and  analysis  of  variance  was  freely  employed 
to  delineate  between  variation  due  to  raw  material  and  that  caused  by 

2  P.  L.  Alger,  "The  growing  importance  of  statistical  methods  in  industry/' 
General  Electric  Review,  Dec.,  1048,  p.  12. 

3  S.  S,  Wilks,   "Statistical  training  for  industry,"  Analytical  Chemistry,  Vol. 
19,  Dec.,  1947,  p.  955. 


10  CHAPTER    1,    THE    ROLE    OF    STATISTICS    IN    RESEARCH 

processing  treatments.  Another  example  in  food  technology  is  provided 
by  Bernhard  (7)  in  his  comparison  of  several  techniques  of  estimation 
of  the  frequency  of  occurrence  of  insect  fragments  in  cream-style 
corn.  The  conventional  methods  utilize  castor  oil  to  separate  the 
insect  fragments  by  flotation.  Bernhard  wished  to  compare  the  effi 
ciency  of  castor  oil  and  lard  oil,  each  at  three  different  temperatures 
for  each  of  four  different  times  of  mixing  oil  and  food  samples.  Of  course, 
repeated  samples  of  any  one  set  of  determination  conditions  could  be 
expected  to  yield  variable  results  for  the  number  of  insect  fragments 
present.  Thus,  statistical  methods  were  required  to  enable  the  variation 
among  mixing  times,  temperatures,  and  oils  to  be  analyzed. 

One  of  the  most  difficult  areas  of  food  research  is  that  of  evaluating 
a  food  product  in  terms  of  consumer  reaction.  It  is  well  known  that 
most  objective  tests  of  food  acceptability  (such  as  laboratory  measure 
ments  of  shear  strength,  etc.)  must  bo  correlated  with  consumer  pref 
erence  by  means  of  taste-panel  observations  in  order  to  achieve  firm 
standing.  The  problems  of  the  "taste  panel"  are  many.  To  what  extent 
is  the  taste  panel  representative  of  the  entire  population  of  tasters? 
How  is  variation  from  sample  to  sample  of  a  food  product  distinguished 
from  variation  from  tnstor  to  taster?  How  can  subjective  evaluation  of 
a  particular  property  of  food,  for  example,  odor,  be  separated  from 
evaluation  of  another  property,  such  as  flavor?  To  what  extent  can  or 
should  restrictions  such  as  instruction*)  to  evaluate  a  narrow  area  be 
placed  upon  the  taster,  in  view  of  the  fact  that  the  heart  of  the  taste- 
panel  system  is  the  use  of  the  integrated  pattern  of  individual  reaction 
to  a  complex  event? 

These  and  many  other  sxich  problems  of  food  evaluation  are  not  en 
tirely  solved.  Kven  the  basic  justification  for  the  introduction  of  sta 
tistical  analysis  is  not  always  clear.  For  example,  a  group  of  tasters 
may  be  asked  to  rank  in  order  of  merit  five  varieties  of  corn.  In  search 
ing  for  a  method  of  evaluating  results  of  this  type  of  problem,  many 
workers  have  followed  the  procedure  of  allotting  a  number  to  each  rank, 
e.g.,  5  for  first,  4  for  second,  etc.  These  figures  are  then  treated  an  num 
bers  ami  analyxed  by  analysis  of  variance  to  check  for  significant  varia 
tion  among;  the  five  varieties*  Huch  a  procedure  is  not  entirely  valid 
because  analysis  of  variance  can  only  be  used  with  numbers,  and  the 
ranking  figures  nre  not  originally  wet  down  an  qxmntitativo  relative  esti 
mates  of  tanto  reaction.  However,  in  view  of  the  lack  of  exact  methods 
of  analysis,  the  technique  mentioned  can  provide  valuable  assistance 
to  the  research  worker  who  deals  with  food  products, 

In  the  social  sciences,  statistical  methods  also  find  wide  application* 
Because  of  their  vital  interest  in  public  opinion,  the  major  political 
parties  have  become  acquainted  with  the  statistician*  In  economic  re 
search,  stat.lstieal  methods  are  almost  indispensable.  Economic  laws 
refer  to  muss  or  group  phenomena,  and  the  determination  of  these  laws 
often  depends  upon  the  judicious  use  of  statistical  techniques. 

In  marketing  research,  an  objective  may  be  increasing  eonsmmption 


1.6      APPLICATIONS    OF    STATISTICS    IN    RESEARCH  11 

of  those  foods  shown  by  nutritional  studies  to  be  inadequately  supplied 
in  the  average  diet.  The  initial  role  of  statistics  here  is  merely  one  of 
finding  consumption  per  capita  and  comparing  it  with  some  goal.  Of 
course,  the  nature  of  the  distribution  of  consumption  per  capita  is  as 
important  as  the  average.  Another  objective  is  the  analyzing  of  mar- 
keting  methods  in  order  to  find  the  least  costly  way  of  doing  the  job. 
As  a  result,  a  smaller  portion  of  society's  efforts  need  be  expended  on 
product  handling. 

Measuring  demand  is  another  of  the  many  difficult  tasks  in  eco 
nomics.  The  research  worker  must  have  a  knowledge  of  consumer  pref 
erences,  supply  of  money,  its  distribution,  etc.  In  measuring  supply, 
he  must  have  an  intimate  acquaintance  with  marketing  functions, 
services,  and  costs  and  be  familiar  with  trends  in  operational  efficiency, 
both  physical  and  managerial.  Data  on  these  particulars  can  only  be 
digested  and  made  available  through  statistical  procedures. 

In  production  economics  probably  the  most  important  comparisons 
are  made  when  two  or  more  characteristics  are  simultaneously  studied 
or  measured.  This  involves  statistical  techniques  known  as  regression 
and  correlation.  These  tools  are  invaluable  to  the  economist.  By  using 
them,  one  factor  can  be  shown  in  its  relationship  with  other  factors.  For 
instance,  if  we  made  the  hypothesis  that  net  income  per  acre  becomes 
higher  as  farm  size  increases,  we  would  want  to  find  the  influence  of 
farm  size  on  net  income  per  acre.  We  might  then  collect  data  for  ten 
units  of  each  farm  size,  ranging  from  40  acres  to  perhaps  480  acres  with 
40-acre  increments.  These  data,  if  properly  obtained  in  accordance 
with  the  rules  of  statistical  procedure,  could  then  be  analyzed  to  aid 
the  researcher  in  making  a  contribution  to  the  theory  of  production 
economics. 

It  is  possible  to  go  on  almost  indefinitely  enumerating  the  fields 
wherein  statistics  is  being,  or  could  be,  applied.  Statistics  is  utilized 
for  a  systematic  approach  to  problems  in  public  health  studies,  epide 
miology,  demography,  biological  assay,  psychology,  education,  sociol 
ogy,  and  in  various  areas  of  home  economics.  Oddly  enough,  statistics 
is  not  confined  to  the  so-called  scientific  world,  for  it  also  is  applied  in 
the  arts.  It  has  been  used  to  aid  in  determining  the  authorship  of  cer 
tain  manuscripts  by  analyzing  the  length  of  sentences.  Authenticity  of 
paintings  has  al§o  been  established  by  analyzing  the  frequency  of  brush 
strokes. 

Although  statistics  is  very  much  in  the  realm  of  an  applied  science, 
it  has  its  theoretical  basis  in  mathematics.  Development  of  the  theo 
retical  branch  of  the  science  is  as  important  as  that  of  the  applied 
branch  if  progress  in  the  field  is  to  continue.  Unfortunately,  there  is  a 
gap  between  statistical  theory  and  application  much  the  same  as  exists 
in  other  sciences.  This  gap  is  steadily  being  closed,  but  the  job  is  far 
from  completion.  Thus  it  is  not  surprising  to  find  statistics  in  use  as 
both  a  science  and  an  art.  It  is  a  science  because  its  methods  are  basi 
cally  systematic  and  of  wide  application.  And  it  is  an  art  because  sue- 


12  CHAPTER    1,    THE    ROLE    OF    STATISTICS    IN    RESEARCH 

cess  in  its  application  depends  on  the  skill,  special  experience,  and 
knowledge  of  the  person  using  it.  The  research  worker  will  become  more 
appreciative  of  this  fact  as  he  gains  a  greater  understanding  of  statisti 
cal  methods  and  their  uses. 

1.7     SUMMARY 

The  scope  of  statistics  might  be  summarized  as  concerned  with:  the 
presentation  and  summarization  of  data,  the  estimation  of  population 
quantities  and  the  testing  of  hypotheses,  the  determination  of  the  ac 
curacy  of  estimates,  the  measurement  and  study  of  variation,  and  the 
design  of  experiments  and  surveys.  Inherently  and  inextricably  in 
volved  in  all  of  the  above-mentioned  areas  is  the  process  known  as 
methods  of  reduction  of  data,  or  the  computational  aspects  of  statistics, 

The  statistical  method  is  one  of  the  devices  by  which  men  try  to 
understand  the  generality  of  life.  Out  of  a  welter  of  single  events,  hu 
man  beings  seek  endlessly  for  general  trends.  Controlled,  objective 
methods  by  which  group  trends  'are  abstracted  from  observations  on 
many  separate  individuals  are  called  statistical  methods.  These  meth 
ods  are  especially  adapted  to  the  elucidation  of  quantitative  data  which 
have  been  affected  by  many  factors.  Statistical  methods  are  fxmda- 
mentally  the  same  whether  employed  in  the  analysis  of  physical  phe 
nomena,  the  study  of  educational  measurements,  the  study  of  data 
resulting  from  biological  experiments,  or  the  analysis  of  quantitative 
material  in  economics.  Agriculturists,  biologists,  chemists,  physicists, 
and  other  researchers  all  attempt  to  eliminate  the  many  miusanee  fac 
tors  which  influence  the  variables  under  investigation  and  to  concen 
trate  their  attention  upon  one  or  two  of  the  most  powerful  factors 
affecting  the  phenomena  being  studied.  Yet,  many  disturbances  are 
always  present  and  thus  statistical  methods  of  analysis  nre  vitally 
necessary*  Wherever  there  is  a  mass  of  numerical  data  that  admits  of 
explanation,  the  statistician  shoxild  consider  itB  analysis  his  field  of 
endeavor. 

To  utilise  statistical  methods  to  advantage,  a  person  should: 

(1)  Be  well  versed  in  the  subject  matter  of  the  field  in  which  the 
research  is  to  be  conducted* 

(2)  Know  how  to  organize  masses  of  data  for  efficient  tabulation 
and  how  to  lay  out  economical  routines  for  handling  data  and 
computation. 

(3)  Know  effective  means  of   presenting   data   in   tabular   and 
graphic  form. 

(4)  Have  some  knowledge  of  the  mathematical  theory  of  statistics 
in  order  to  have  assurance  there  is  a  fair  correspondence  be 
tween  his  data  and  the  assumptions  underlying  the  formulas 
ho  XISOH. 

(5)  Bo  acquainted  with  a  variety  of  ntatistieal  techniques,  the 
limitations  and  advantages  of  each,  the  assximptions  upon 
which  they  are  based,  the  place  each  occxipicn  hi  a  logical 


1  .7       SUMMARY  1 3 

analysis  of  the  data,  and  the  interpretations  which  can  be 
made  from  them. 

Statistics,  then,  boils  down  to  numerical  results,  the  methods  and 
processes  used  in  obtaining  them,  the  methods  and  means  for  estimate 
ing  their  reliability,  and  the  drawing  of  inferences  from  these  resultslv 

During  the  past  half-century,  the  thinking  world  appears  to  have 
awakened  to  an  unusually  deep  appreciation  and  respect  for  numerical 
facts.  There  has  been  a  growing  tendency  to  reduce  observations  and 
accumulated  data  to  an  orderly  arrangement,  making  possible  the 
evaluation  of  results  by  means  of  a  systematic  method  of  analysis. 

Formerly,  many  persons  believed  statistical  analysis  could  be  used 
only  in  certain  highly  specialized  fields.  However,  more  and  more 
methods  of  statistical  analysis  are  finding  their  way  into  scientific 
workshops  in  all  fields.  This  is  due  largely  to  the  fact  that  some  of  the 
enthusiastic  supporters  of  statistical  methods  have  worked  faithfully 
to  develop  and  explain  methods  useful  to  and  usable  by  those  persons 
not  specifically  trained  in  higher  mathematics. 

In  the  field  of  statistical  analysis  advancement  has  been  rapid  in 
recent  years.  Many  useful  methods  are  now  available  for  the  analysis 
of  data  arising  from  different  sources.  A  clear  grasp  of  simple  and 
standardized  statistical  procedures  will  go  far  to  elucidate  principles  of 
experimentation.  However,  one  must  remember  that  these  procedures 
are  in  themselves  only  a  means  to  a  more  important  end.  As  fundamen 
tal  and  pervasive  as  statistical  thinking  is  in  the  modern  world,  it  must 
not  be  considered  an  end  in  itself.  The  statistical  method  is  a  tool  for 
organizing  facts  so  they  are  rendered  more  available  for  study.  A  sta 
tistical  study  can  only  describe  what  is;  it  cannot  determine  what 
ought  to  be,  except  insofar  as  it  may  throw  light  upon  probable  con 
comitants  and  consequences  of  certain  situations.  It  is  fatuous  to  sup 
pose  the  statistical  method  can  provide  mechanical  substitutes  for 
thinking,  although  it  is  often  an  indispensable  aid  to  thinking.  Men 
see  increased  prevalence  of  the  statistical  method  in  scientific  studies; 
and,  sometimes,  failing  to  grasp  underlying  reasons  for  this  develop 
ment,  they  assume  the  use  of  tables,  formulas,  and  numerical  sum 
maries  is  a  badge  of  respectability.  As  a  result,  some  studies,  truly 
subjective  in  nature,  are  invested  with  a  false  show  of  objectivity. 
Thus,  a  vast  superstructure  of  computation  is  raised  upon  a  foundation 
inappropriate  to  such  treatment.  When  such  a  picture  is  painted,  it  is 
neither  good  statistics  nor  good  philosophy. 

Most  statistical  studies  will  not  answer  all  the  questions  we  would 
like  to  have  answered  regarding  a  given  problem.  From  the  very  nature 
of  statistical  work,  results  are  apt  to  be  partial  and  fragmentary,  rather 
than  complete  and  final.  Therefore,  the  researcher  must  make  up  his 
mind  that  questions  must  sometimes  be  left  unanswered.  He  must  also 
on  occasion  freely  admit  his  study  has  limitations.  Any  shortcomings  in 
his  work  and  the  danger  of  attributing  more  than  claimed  for  his  in 
vestigation  should  be  pointed  out  to  his  readers  by  the  researcher. 


14  CHAPTER    1,    THE    ROLE    OF    STATISTICS    IN    RESEARCH 

It  is  also  imperative  that  conclusions  drawn  from  observational  re 
sults  be  based  on  a  detailed  knowledge  of  procedures  employed  in  the 
investigation.  The  interpretive  function  in  statistical  analysis  is  one  of 
the  most  important  contributions  of  statistics,  and  the  statistician 
should  plan  experiments  and  investigations  which  will  yield  maximum 
information  and  valid  conclusions  from  scientific  research  data.  In 
ference  from  the  particular  to  the  general  must  be  attended  with  some 
degree  of  uncertainty,  and  research  workers  in  all  fields  of  science  must 
recognize  the  role  statistics  plays  in  this,  the  most  important  aspect  of 
research. 

The  role  of  statistics  in  research  is,  then,  to  function  as  a  tool  in  de 
signing  research,  in  analyzing  its  data,  and  in  drawing  conclusions 
therefrom.  A  greater  and  more  important  role  can  scarcely  be  en 
visioned.  In  utility  to  research,  statistics  is  second  only  to  the  mathe 
matics  and  common  sense  from  which  it  is  derived.  Clearly  the  science 
of  statistics  cannot  be  ignored  by  any  research  worker  even  though  he 
may  not  have  occasion  to  use  applied  statistics  in  all  of  its  detail  and 
ramifications. 

Problems 

1.1  Discuss  the  following  terms  or  phrases:  (a)  observation  mid  descrip 
tion;  (b)  cause  and  effect;  (c)  analysis  and  synthesis;  (d)  assumption, 
postulate,   and  hypothesis;    (e)    testing   of  hypotheses;   (f)    deduction 
and  induction. 

1.2  What   do    you   believe   operations   researchers    mean    by   the    phrase 
"measure  of  effectiveness"? 

1.3  Saaty  (39),  in  a  chapter  entitled  "8ome  Remarks  on  Scientific  Method 
in  Operations  Research/'  refers  to:  (a)  the  jxulgment  phase,   (b)  the 
research  phase,  and  (c)  the  action  phase.  Give  your  interpretations 
of  these  three  phases.  Then  compare  your  views  with  those  of  Saaty. 

1.4  Read  Chapter   12,   "Some  Thoughts  'on  Creativity,"  in  8aaty   (H<>). 
Then  prepare  a  brief  report  on  your  reactions  to  his  ideas* 

1.5  Prepare  a  report  on  the  pros  and  eons  of:  (a)  individual  research  and 
(b)  interdisciplinary  team  research. 

. ,6  DmcuHS  the  similarities  and  dissimilarities  of  pure  and  applied  research. 
K7        Prepare  a  report  on  the  subject  of  "scientific  method/1 

1.8  Prepare  a  report  on  your  interpretation  of  "the  role  of  statistics  in 
research," 

1.9  By  consulting  the  technical  journal**  in  your  area  of  specialization, 
prepare  and  submit  a  list  of  references  (properly  documented)  which 
illustrate  the  use  of  statistical  methods. 

1.10  Submit  a  list  of  publications  (bookn,  monographo,  papors,  etc.)  which 
you  believe  would  be  worthwhile  additions  to  the  references*  prenerited 
with  this  chapter. 

References  and  Further  Reading 

1,  Algcr>  P.  L*   The  growing  importance   of   Htatintical    method**   in  industry. 
General  Ktertric  Review,  51  (No.  12): II,  1948, 

2.  American  Management  Association.  Getting  the  Moat  from  Product 
and  Development.  Now  York>  1055. 


REFERENCES    AND    FURTHER    READING  15 

3     ^  Making  Effective  Use  of  Research  and  Development.  New  York,  1956. 

4.   .  Engineering   and  Research  in   Small  and   Medium  Size    Companies. 

New  York,  1957. 

5.  Anderson,  J.  A.  The  role  of  statistics  in  technical  papers.  Trans.  Amer.  Assn. 
Cereal  Chemists ,  3:69,  1945. 

6.  Bard,  J.  C.  Changes  in  tenderness,  fiber  cohesiveness  and  moisture  content 
of  canned  beef  due  to  thermal  processing.   Master  of  Science  thesis,  Iowa 
State  University,  Ames,  1950. 

7.  Bernhard,  F.  L.  Recovery  and  identification  of  insect  fragments  from  cream 
style  corn.  Master  of  Science  thesis,  Iowa  State  University,  Ames,  1951. 

8.  Beveridge,  W.  I.  B.   The  Art  of  Scientific  Investigation.  W.  W.  Norton  and 
Co.,  New  York,  1951. 

9.  Buros,  O,  K.  (editor)  Research  and  Statistical  Methodology  Books  and  Reviews 
1933-38.  Rutgers  University  Press,  New  Brunswick,  N.J.,  1938. 

10. .  (editor)  The  Second  Yearbook  of  Research  and  Statistical  Methodology 

Books  and  Reviews.  The  Gryphon  Press,  Highland  Park,  N.J.,  1941. 

11.   .  (editor)    Statistical  Methodology  Reviews  194-1-50.  John   Wiley  and 

Sons,  Inc.,  New  York,  1951. 

12.  Bush,  G.  P.,  and  Hattery,  L.  H.  (editors)  Teamwork  in  Research.  American 
University  Press,  Washington,  D.C.,   1953. 

13.  Chapanis,  A,  R,  JE.  Research  Techniques  in  Human  Engineering.  The  Johns 
Hopkins  Press,  Baltimore,  1959. 

14.  Churchman,  C.  W.  Theory  of  Experimental  Inference.  The  Macmillan  Com 
pany,  New  York,  1948. 

15    ^  Ackoff,  R.  L.,  and  ArnofT,  B.  L.  Introduction  to  Operations  Research. 

John  Wiley  and  Sons,  Inc.,  New  York,  1957. 

16.  Cohen,  M.  R.,  and  Nagel,  E.  An  Introduction  to  Logic  and  Scientific  Method. 
Harcourt,  Brace  and  Company,  New  York,  1934. 

17.  Cox,  G.  M.  The  value  and  usefulness  of  statistics  in  research.  Lecture  given 
for   the    USDA    Committee    on    Experimental    Design.    Washington,    D.C., 
Jan.  11,  1951. 

18.  Flagle,    C.    D.,    Huggins,   W.   H.,   and  Roy,   R.   H.   Operations  Research  and 
Systems  Engineering.  The  Johns  Hopkins  Press,  Baltimore,  1960. 

19.  Frecnlmati,    P.    The    Principles   of  Scientific   Research.   Public   Affairs   Press, 
Washington,  D.C.,  1950. 

20.  Good,  C.  V.,  and  Scatea,  D.  EL  Methods  of  Research:  Educational,  Psycho 
logical,  Sociological.  Appleton-Century-Crofts,  New  York,  1954. 

21.  Goode,  H.  H.,  and  Machol,  R.  E.  System  Engineering,  An  Introduction  to  the 
Design  of  Large  Scale  Systems.   McGraw-Hill  Book  Co.,   Inc.,    New  York, 
1957. 

22.  Gryna,  F.  M.,  Jr.,  McAfee,  N.  J.,  Ryerson,  C.  M.,  and  Zwerling,  S.  (editors) 
Reliability   Training   Text.  Second  Ed.   Institute  of  Radio   Engineers,   Inc., 
New  York,  1959. 

23.  Hawley,   G.  C).,  and  Ostle,  B.  Training  for  reliability  and  quality  control. 
Proc.  of  the  Seventh  National  Symposium  on  Reliability  and  Quality  Control  in 
Electronics,  pp.  91-96,  Jan.  9-11,  1961. 

24.  Hill  way,  T.  Introduction  to  Research.  Houghton  MifTiin  Co.,  Boston,  1956. 

25.  Hutt,  F.  B.,  Genetics  of  the  Fowl.  McGraw-Hill  Book  Company,  Inc.,  New 
York,  1949. 

26.  Jeffreys,  It.  Scientific  Inference.  Cambridge  University  Press,  London,  1937. 

27.  JOVQIIS,   W.  S,   The  Principles  of  Science.  Second  Ed.   Macmillan  and  Co., 
New  York,  1877. 

28.  Johnnon,   P.  O.   Modern  statistical  science  and  its  function  in  educational 
and  psychological  research,  Sci.  Monthly ,  72:385,  1951. 

29.  Lush,  J.  L.  Animal  Breeding  Plans.  The  Iowa  State  University  Press,  Ames, 
1945. 

30.  Luszki,   M-  TO.  B*  Interdisciplinary  Team  Research:  Methods  and  Problems. 
Now  York  University  Press,  New  York,  1958. 


16  CHAPTER    1,    THE    ROLE    OF    STATISTICS    IN    RESEARCH 

31.  Miller,  D.  W.,  and  Starr,  M.  K.  Executive  Decisions  and  Operations  Research. 
Prentice-Hall,  Inc.,  Englewood  Cliffs,  N.J.,  1960. 

32.  Morse,   P.   M.,   and   Kimball,    G,    K.   Methods  of  Operations   Research.  John 
Wiley  and  Sons,  Inc.,  New  York,  1051. 

33.  Ostle,   B.  Planning  Experimental  Programs.  Humble  Oil  and    Refining   Co., 
Refining    Department    Technical    and    Research     Divisions,    Research    and 
Development  Division,  Baytown,  Tex.  RL.  44M.52,  6-20-1,  Aug.  2O,   1952. 

34. .  Statistical  problems  in  designing  range  management  investigations. 

Report  on   Meeting  of  Kconomics  of  Range   Resource    Development   Com 
mittee   of   the    Western    Agricultural    Economics    Research    Council,    Reno, 
Nev.,  Oct.,  1954. 

35_ ._  Statistical  problems  in  designing  experiments  to  study  the  economics 

of  fertilizer  application.  Conf.  Proc.,  Farm  Management  Research  Com 
mittee  of  the  Western  Agricultural  lilconomics  Research  Council,  Corvallis, 
Ore.,  Jan.,  1956. 

3(} — .  Statistics  in  engineering.  Jour.  Engin,  Educ.,  47:(No.  5):410— 14, 

Jan.,  1957. 

37^  _™_>  and  Tischer,  R.  G.  Statistical  methods  in  food  research.  Advances  in 
Food  Research,  5:161-259,  1954. 

38.  Popper,  K.  R.  The  Logic  of  Scientific  Discovery.  Basic  Books,  New  York,  1959. 

39.  Saaty,    T.   L.    Mathematical  Methods  of  Operations   Research.    McGraw-Hill 
Book  Co.,  Inc.,  New  York,  1959. 

40.  Rasicmi,    M.,    Yaapaii,    A.,   and  Friedman,    I...    Operations  Research:    Methods 
and  Problems.  John  Wiley  and  Sons,  Inc.,  Now  York,   1959. 

41.  Sehenek,    H.,    Jr.    Theories    of  Engineering    Experimentation.    McGraw-Hill 
Book  Company,  Inc.,  New  York,  1961, 

42.  Snedoeor,  G.  W.  On  a  uniqxio  feature  of  statistics.  Jour.  Amer.  Stat*  Assn., 
44:1,  1949. 

43.  ,____%  The  statistical  part  of  the  scientific  method,  Ann.  N.Y+  Acad.  Sci., 
52:792,  1950. 

44.  Taton,  R.  (trans,  by  A.  J.  Pome.rans)   Reason  and  Chance  in  Scientific  Dis 
covery.  Philosophical  Library,  New  York,  1957. 

45.  Walker,    H.    3VL    Statistical    literacy    in    the    social    sciences.    Amer.    Stat.y 
5  (No.  1):6,  1951. 

46.  Whitney,   F.   Li.    The   Elements  of  Research.   Third   Ed.   Prentice-Hall,   Inc., 
New  York,  1950. 

47.  Wiesen,  J.   M.,  and  Oatle,   B.  Some  problems  in  the  preparation  of  quality 
specifications.    Proc*   of  the   Fifth   National   Symposium   on   Reliability   and 
Quality  Control  in  Electronics,  pp.  375-80,  Jan/  12-14,  1959. 

48.  Wilks,  H.  8,  Statistical  training  for  induntry.  AnaL  Wusm.,  19:953,  1947, 

49.  WilHon,    K»    B.   An   Introduction  to   $rie,ntijtc   /^search.    McCrraw-Hill   liook 
(Company,  Inc.,  New  York,  1952, 

50.  Worthing,  A.  G.,  and  (tofFner,  J.  Treatment  of  Experimental  Data.  John  Wiley 
and  Sons,  Ino.,  New  York,  1943* 


CH  APTE  R    2 

MATHEMATICAL  CONCEPTS 

IT  is  DIFFICULT  to  achieve  a  clear  understanding  of  statistical  methods 
without  discussing,  to  some  extent,  the  underlying  theory.  Since  the 
theory  of  statistics  is  intimately  associated  with  the  theory  of  prob 
ability  and,  further,  since  probability  is  an  important  branch  of  mathe 
matics,  this  implies  that  every  student  of  statistical  methods  should 
be  willing  to  "use  a  little  mathematics  once  in  a  while!"  Consequently, 
it  seems  desirable  to  present  here  a  few  basic  mathematical  concepts, 
formulas,  and  techniques  which  may  prove  helpful  to  the  reader.  These 
ideas  will  be  presented  as  definitions  and/or  theorems1  (without  proofs). 

SET  THEORY 

The  subject  of  the  theory  of  sets  is  fundamental  in  mathematics.  In 
this  text,  however,  we  shall  be  concerned  only  with  a  few  basic  concepts 
which  are  useful  in  the  theory  of  probability. 

Definition  2.1     A  set  is  a  collection  of  elements.2 

Definition  2.2  The  universal  set  is  the  set  consisting  of  all  elements 
under  discussion.  (NOTE:  The  universal  set  is  some 
times  referred  to  as  a  space.) 

Definition  2.3     The  null  set  is  the  set  containing  no  elements  at  all. 

Definition  2.4  Associated  with  each  set,  A,  is  another  set,  A',  called 
the  complement  of  A  and  defined  to  be  the  set  consist 
ing  of  all  the  elements  of  the  universal  set  which  are  not 
elements  of  A. 

Definition  2.5  For  any  two  sets,  A  and  JB,  the  union  of  A  and  B  is  the 
set  consisting  of  all  elements  which  are  either  in  A  or 
in  B  or  in  both  A  and  B.  The  union  of  A  and  B  is  com 
monly  denoted  by  A^JB. 

Definition  2.6  For  any  two  sets,  A  and  J5,  the  intersection  of  A  and  B 
is  the  set  consisting  of  all  elements  which  are  both  in 
A  and  R.  The  intersection  of  A  and  B  is  commonly 
denoted  by  AC\B  or  by  AB. 

Theorem  2.1       If  A  and  B  arc  two  sots  which  have  no  common  ele 
ments,  then  the  sot  AJS  is  the  null  set. 
A  useful  device  for  illustrating  the  properties  of  the  algebra  of  sets 

is  the  Venn  diagram.  In  such  a  diagram,  the  points  interior  to  a  rec- 

1  The  expression  "theorems"  will  bo  xisod  in  a  very  broad  sense  to  describe 
various  proportion,  propositions,  theorems,  ote.,  which  result  from  the  definitions. 
While  not  strictly  correct,  this  procedure  will  materially  reduce  the  number  of 
terms  to  bo  absorbed  by  the  reader. 

a  The  term  element  will  bo  left  undefined. 

C17J 


18 


CHAPTER    2,    MATHEMATICAL    CONCEPTS 


FIG.  2.1— A  Simple  Venn  diagram. 


tangle  constitute  the  xmiversal  set.  Arbitrary  sets  within  the  universal 
set  (that  is?  subsets  of  the  universal  set)  will  be  represented,  for  con 
venience,  by  the  points  interior  to  circles  within  the  rectangle.  In 
Figure  2.1  the  set  A  is  shaded  by  vertical  lines,  the  set  B  is  shaded  by 
horizontal  lines.  Since  A  I)  9*0,  that  is,  does  not  equal  the  null  set,  AB 
appears  as  the  erosshatehed  area. 

Probability  theory,  which  we  shall  sximmari^e  in  the  next  chapter, 
depends  on  the  number  of  elements  in  a  set.  We  will  denote  the  number 
of  elements  in  any  arbitrary  set  A  by  n(A). 
Theorem  2.2     If  A  and  B  have  no  elements  in  common, 

n(A^JB)  «n(-4)4-n(jR). 

Theorem  2*3      If  A  and  B  have  no  elements  in  common,  n(AB)  —  0, 
Theorem  2,4     For  arbitrary  sets  ,*1  and  B,  it  is  true  that 


NOTATION 

As  in  all  subjects,  the  system  of  notation  employed  in  a  matter  of 
concern  to  the  reader.  Since  statistics  is  HO  entwined  with  mathematics, 
it  in  no  surprise  that  problems  of  notation  arise*  In  the  remainder  of 
this  book  every  attempt  will  be  made  to  define  and  explain  special 
symbols  and  notation.  However,  at  this  point  it  seems  appropriate  to 
mention  some  of  the  more  frequently  occurring  signs  and  symbols. 
Definition  2,7  The  absolute  value  of  a  number,  x,  denoted  by  |xf  ?  is 

its  numerical  value  neglecting  its  algebraic  nigtx*  For 

example,  j  ~~3|  —3  and  J3|  —3, 
Definition  2,8       #  =  y  is  read  "x  is  equal  to  37." 
Definition  2.9       ;rp^//  is  read  ".#  is  not  equal  to  |/.?> 


NOTATION 


19 


Definition  2.10 
Definition  2.11 
Definition  2.12 
Definition  2.13 
Definition  2.14 


x=z/  is  read  "x  is  approximately  equal  to  y* 

x<y  is  read  "re  is  less  than  y." 

x<y  is  read  "x  is  less  than  or  equal  to  y." 

x>y  is  read  "x  is  greater  than  y." 

x>y  is  read  "x  is  greater  than  or  equal  to  y.' 


Definition  2.15 


Y,  = 


Y2  + 


-    + 


NOTE:  The  Greek  capital  letter  sigma,  )?  is  known 
as  the  summation  sign.  Further,  i  is  called  the  index 
of  summation,  while  1  and  n  are  known  as  the  limits 
of  summation. 


Theorem  2.5 


Theorem  2.6 


cYi  =  c 


t-  where  ^  is  a  constant. 


Theorem  2.7 


Theorem  2.8 


+ 


,  + 


Theorem  2.9          (  ^  Y^\  -  J£    Y*  +  2  S     S 


NOTE:  In  this  theorem  the  notation  ]>D*-<y  ^s  in~ 
terpreted  to  mean  that  we  sum  all  possible  products 
YiY,  letting  i  and  j  go  from  1  to  n,  subject  only  to  the 
restriction  that  in  any  particular  term, 


Definition  2.16 


t-  -  (FO(F2) 


(Fn). 


NOTE:  In  contrast  to  Definition  2.15,  in  which  we 
introduced   23  as  ^e  summation  sign,  we  have  here 


20 


CHAPTER    2,    MATHEMATICAL    CONCEPTS 


introduced  the  Greek  capital  letter  pi,  TI?  as  the  prod 
uct  sign. 


Theorem  2.1O 


i  =  (i)(2) 


(n)  =  nl 


NOTE:  The  symbol  nl  is  called  n  factorial,   or  fac 
torial  n. 
Definition  2. 17     0 ! «  1 . 

NOTE:  This  will  prove  useful  later. 

PERMUTATIONS   AND  COMBINATIONS 

Permutations  and  combinations  are  concerned  with  the  different 
subgroups  and  arrangements  that  can  be  formed  from  a  given  set.  A 
permutation  is  a  particular  sequence  (i.e.,  arrangement)  of  a  given  set 
or  subset  of  elements,  while  u  combination  is  the  set  or  subset  without 
reference  to  the  order  of  the  contahied  elements. 


Definition  2,18 


If  an  event,  A,  can  occur  in  n(A)  ways  and  if  a  differ 
ent  event,  B,  can  occxir  in  n(75)  ways,  then  the  event 
"either  A  or  R"  can  occur  in  n(A)+n(B)  ways  pro 
vided  A  and  B  cannot  occur  simxiltaneously. 
NOTE:  You  will  notice  the  similarity  between  this 
definition  and  Theorem  2.2, 

If  an  event,  A,  can  occur  in  n(A}  ways  and  a  sub 
sequent  event,  fij  can  occur  in  n(B)  ways,  then  the 
event  "both  A  and  B)f  can  occur  in  n(A)  -n(B)  ways. 
An  r-pormutation  of  n  things  is  an  ordered  selection  or 
arrangement  of  r  of  them. 

An  r-combination  of  n  things  is  a  selection  of  r  of  them 
without  regard  to  order. 

The  number  of  different  permutations  which  can  be 
formed  from  n  distinct  objects  taken  r  at  a  time  is 
JP(n,  r)=tt(/i— 1)  *  •  -  (n  — r+l)«n!/(n  — r)! 
The  number  of  different  permutations  which  can  be 
formed  from  n  objects  taken  n  at  a  time,  given  that 
n»-  are  of  type  it  where  i=*  1,  2,  •  -  *  ,  A:,  and  ^nt- «=n« 
is  P(n\  niy  na,  •  -  •  »  M*)  ^nl/niln^l  *  *  -  n&\ 
The  number  of  different  combinations  which  can  be 
formed  from  n  dintinct  objects  taken  r  at  a  time  ia 


Definition.  2,19 

Definition  2.20 
Definition  2.21 
Definition  2.22 

Definition  2,23 

Definition  2*24 

Definition  2.25     <7(n,  r)  «  0  for  r  <0  and  r>n. 

SOME  USEFUL   IDENTITIES  AND  SERIES 

In  statistical  work  it  in  often  necessary  to  sum  a  series  of  terms  or 
simplify  a  particular  expression.  A  few  of  the  more  useful  results  are 
given  here  for  roady  reference. 


SOME    IMPORTANT    FUNCTIONS  21 


Theorem  2.11        (a  +  by  = 

r=0 

Theorem  2.12       If,  in  Theorem  2.11,  we  let  a=l  and  &  =  Z,  we  obtain 


Theorem  2.13       If,   in   Theorem   2.11,    we   let   a  =  q  and   &  =  p=l—  -q 
where  0<p<l,  we  obtain 

n 

1  =  ^JL,  C(n,  r)qn~rpr. 

r»»0 

This  is  a  very  useful  expression  in  probability  and 
statistics. 


n—  1 


Definition  2.26     ex  =  exp  (oc)  === 
Theorem  2.14       (1  —  #n)/(l  —  c 

00 

Theorem  2.15       i/(l  —  x)n  =  1C  C(n  +  i  —  1,  z)^'- 
Theorem  2.16         ]T)  C(a,  i)  -C(J,  c  —  i)  =  C(a  +  &,<;). 

i— 0 
n 

Theorem  2.17        ^  i  =  n(w  +  l)/2. 

n 

Theorem  2.18         y^  ^2  ...  n^n  _^  i)(2>z  +  l)/6. 


Theorem  2.19        23  i^<  =  a?/(l  —  x)*     for  —  1  <  x  <  1. 

i— 1 

SOME    IMPORTANT   FUNCTIONS 

Some  mathematical  functions  not  always  presented  in  courses  in 
elementary  mathematics  are  of  great  interest  to  the  statistician.  Two  of 
these  will  be  presented  here  for  your  convenience. 

Definition  2,27     The  gamma  function,  denoted  by  r(y>),  is  defined  by 
the  integral 


CHAPTER    2,    MATHEMATICAL    CONCEPTS 

y-»   OO 

T-T-?       I  3CP~~~^'& — "K(Jf2C 

Jo 
for  p>0.  An  alternative  form  for  this  function  is 

r(p)   =  2   f     yip-ierSdy, 

J   0 

where  the  transformation  used  was  x  =  yz+ 
Theorem  2*20       If,  in  Definition  2.27,  we  let  p~n  where  n  is  a  positive 

integer,  we  obtain  F(n)  =  (n— 1) -r(ri  —  l)  =  (n— 1)  ! 
Theorem  2.21       F(i)  —  vV^  (^r)172. 
Definition  2,28  The  beta  function,  denoted  by  J3(p,q),  is  defined  by  the 

integral 

i 


=    I 
•^  o 


for  p>Q  and  #>CK  An  alternative  form  for  this  func 
tion  is 


,  <?> 


/W  /Ii 
^ 


0  cos*«-1  Q  de, 


whore  the  transformation  used  was  ^  =  vsi 
Theorem  2*22       /3(p,  q)  ==/5(r/,  p). 
Theorem  2.23       ft(p,  ff)  =  r(p) 4 

MATRICES 

Many  of  the  methods  to  be  discussed  in  this  book  depend  on  the 
theory  of  linear  .statistical  models.  This  theory  is  most  expeditaoxisly 
handled  in  terms  of  matrix  algebra.  Therefore,  it  is  appropriate  that 
the  reader  be  made  aware  of  the  basic  concepts*  As  in  the  preceding 
sections  of  this  chapter,  definitions  and  theorems  will  be  stated  without 
discussion. 

Definition  2.29     A  matrix  A  of  dimension  rXc  is  a  rectangular  array 
of  elements  a^  arranged  in  r  rows  and  c  columns: 


Definition  2*30 
Definition  2.31 


If  it  is  necessary  to  emphasise  the  dimension,  we  shall 
write  ArG  instead  of  A. 

If  ,4  is  of  dimension  nXl>  it  is  called  an  nXl  vector. 
A™B  when  and  only  when  .4  and  &  are  of  the  same 
dimension  and  a,;  —  6^-  for  all  i  and  j. 


MATRICES 


23 


Definition  2.32 


Definition  2.33 


Definition  2.34 


The  product  of  a  matrix  A  and  a  scalar  (ordinary) 
number  k  is  a  matrix  B  where  &#  =  kaiy-  for  all  i  and  j. 
That  is,  kA  =*=  Ak  =  5. 

The  sum  of  two  matrices,  -4  and  Z?,  can  be  defined  only 
when  A  and  B  are  of  the  same  dimension.  Then 
<A-hJ?  =  C  where  c1-y  =  a^+&^. 

The  product  of  two  matrices,  say  AB,  can  be  defined 
only  when  the  number  of  columns  in  A  equals  the  num 
ber  of  rows  in  B.  Then  AB  =  C  where 


Definition  2.35 


Theorem  2.24 
Theorem  2.25 
Theorem  2.26 
Definition  2.36 
Theorem  2.27 

Definition  2.37 


Definition  2.38 

Theorem  2.28 
Definition  2.39 

Definition  2.40 
Definition  2.41 
Definition  2.42 


«=1 

NOTE:  We  must  be  very  careful  of  the  order  of  the 
factors  when  multiplying  one  matrix  by  another. 
Even  if  AB  and  BA  are  both  defined,  they  are  not 
necessarily  equal. 

The  transpose  of  a  matrix  A  of  dimension  rXc  is  de 
noted  by  A',  where  Af  is  a  matrix  of  dimension.  eXr 
in  which  a'-/  =  c&yi.  That  is,  the  rows  of  A'  are  the  col 
umns  of  A  and  the  columns  of  Af  are  the  rows  of  A. 


=  AA,   A* 


If  r  =  c,  A  is  called  a  square  matrix, 
For  a  square   matrix   A,  we   can   write 
~AAA,  etc. 

In  a  square  matrix  of  dimension  nXn,  the  elements 
an,  #22,  -  •  •  ,  an«,  form  the  main  diagonal  and  are 
known  as  diagonal  elements. 

A  square  matrix  which  is  symmetric  with  respect  to 
its  main  diagonal  is  called  a  symmetric  matrix. 
For  a  symmetric  matrix,  A'  =  A. 

A  symmetric  matrix  in  which  a^'  =  0  for  all  i^j  is 
called  a  diagonal  matrix. 

A  diagonal  matrix  in  which  a«=l  for  all  i  is  called  a 
unit  (or  an  identity)  matrix,  and  will  be  denoted  by  /. 
A  matrix  having  all  its  elements  equal  to  zero  is  called 
the  null  matrix,  and  will  be  denoted  by  0. 
The  determinant  of  a  square  matrix  A  of  dimension 
denoted  by  \A\,  is  defined  by 


where  the  second  subscripts  n,  r%,  -  •  •  ,  rn  run  through 
all  the  n\  possible  permutations  of  the  numbers 
1,  2,  -  -  -  ,  n,  and  the  sign  of  each  term  (either  + 
or  — )  is  determined  according  to  a  well-defined  rule. 


24 


CHAPTER    2,    MATHEMATICAL    CONCEPTS 

NOTE:  If  A  is  of  dimension  2X2,  then 


Theorem  2.29 
Theorem  2.30 

Theorem  2.31 
Theorem  2.32 
Theorem  2.33 

Definition  2.43 
Definition  2.44 

Theorem  2.34 


A     = 


For  any  square  matrix  A,   \  A  j  =  J  A'  J , 
If  two  rows  (or  columns)  of  a  square  matrix  are  inter 
changed,  the  determinant  changes  its  sign. 
If  two  rows  (or  columns)  of  a  square  matrix  are  identi 
cal,  the  determinant  is  0. 

If  -4,  B,  and  C  are  square  matrices  such  that  AB=  C, 
then  \A\  -  \B\  -|C|. 

If  a  multiple  of  one  row  (column)  is  added  to  another 
row  (column)  of  a  square  matrix,  the  determinant  is 
unchanged. 

For  any  arbitrary  matrix  -4?  the  determinant  of  any 
square  submatrix  of  A  is  called  a  minor  of  A. 
For  a  square  matrix  A,  the  minor  obtained  by  deleting 
the  ith  row  and  yth  column,  multiplied  by  (•—  I)*4"*", 
is  known  as  the  cofactor  of  at-/.  We  shall  denote  the 
cofactor  of  a»y  by  cof  a*/. 

For  a  square  matrix  A  of  dimension  nXn?  the  de 
terminant  |  .4 1  may  be  found  by  evaluating 

**  n 

*>  -  2:  < 


Definition  2,45 
Definition  2.46 


If,  for  a  square  matrix,   \A\^09  then  A  is  of  rank  n 
and  A  is  said  to  bo  nonsingular. 

For  a  nonsingular  sciuare  matrix  -4,  the  inverse  of  A 
is  denoted  by  A™1  and  is  defined  by 


Theorem  2.35       For  a  nonmngulur  square  matrix  A,  it  is  true  that 
Theorem  2*36       For  a  nonsingular  square  matrix  ,4,  it  is  true  that 

i    A      i """""  *  w\   \    A  *~"  *  i 

V.**    J  **"**   V  **         J     * 

LINEAR  EQUATIONS 

Many  times  in  statistical  work  wo  find  it  necessary  to  discuss  sys 
tems  of  linear  equations  such  as: 

+  a22*a  +  •  •  •  4-  a*»x»  -  y*  ^^  ^ 


Hr 


PROBLEMS  25 

The  matrix  notation  introduced  in  the  preceding  section  gives  us  an 
extremely  concise  method  of  representing  such  systems.  For  example 
it  is  clear  that  * 


AX=  Y 

is  the  same  as  Equation  (2.1)  if 


(2.2) 


A  = 


<Zin~ 


X  = 


and     Y  = 


Theorem  2.37     If  Jl  in  Equation    (2.2)   is  nonsingular.   then 

X=A-iY.  That  is, 


oc<i   — 


j_          •- 

~rr  23  y*(cof  a^ 
A      iwi 


forj  =  1,  2,  -  -  -  ,  n. 
NOTE:  Another  way  of  writing  this  is 
matrix  A  in  which  a^ 
has  been  replaced  by 
i=l,  2,  •  •  •  ,  n 


X*    = 


That  is, 


W-i      y« 


J  =  l, 


Problems 


2.1 


2.2 
2.3 


Consider  a  box  of  resistors  that  are  color  coded  (red,  black,  or  yellow) 
according  to  resistance  rating.  Suppose  that  all  red  (JB)  resistors 
and  some^of  the  black  (2J)  resistors  are  manufactured  by  company  E. 
The  remainder  of  the  black  resistors  are  manufactured  by  company  Ft 
while  the  yellow  (F)  resistors  are  manufactured  by  company  G.  The 
universal  set  consists  of  all  the  resistors  in  the  box.  Letting  R  stand 
for  the  set  of  all  red  resistors,  B  stand  for  the  set  of  all  black  resistors, 
and  so  on,  write  as  many  equations  and  inequations  as  you  can  to 
describe  the  relations  existing  among  the  various  sets. 
List  all  subsets  of  the  set  {X,  Y,  Z}. 

Consider  the  space  consisting  of  the  26  lower  case  letters  of  the  alpha 
bet.  If  the  sets  A,   B,   and   C  are  defined  as    A «  { a,    &,    c,    d     e\ 
B~  {b,  d,f,  h,j}f  and  C-  {c,/,  i,  I,  m} ,  find: 


26  CHAPTER    2,    MATHEMATICAL    CONCEPTS 

(<*)  AVJB,  B\JC,  A\JB\JC 

(&)  AB,  J5C,  ABC 

(c) 

(d)  (A\JB)  (A\JBf) 

(e)  (A^JB)  (A'WB)  ( 

2.4  How  many  different  subsets  arc  there  in  a  set  containing  n  distinct 
elements? 

2.5  Draw  Venn  diagrams  for  the  following  and  shade  the  indicated  area: 

(a)  A^JA'BC 

(6) 

(c) 


2.6  Referring  to  Problem  2.3,  give  the  number  of  elements  in  each  set 
discussed. 

2.7  If  A"i=*4,  A^aa—  3,  AT'a  =  i,  X"4  =  7,  1^=8,  Fo  =  2,  F3  =  —  1,  and  r4  =  3, 
find: 


(4 
T,  -Y* 
*—L 


t— 1 

2.8        Given  the  following  observations: 

Fm  -  4  Fm  -  3  Fun  -  0 

—  3  Fm  —  —  3  Fin*  —  9 

«  —  1  Fm  «*  0  F*M  -•  4 

«  -  8  F4ftl  «  14  K4M  -  0 

«*  22  F4t«  *•  7  F4is  «-•  0 

find: 

433 


1— I 


PROBLEMS  27 


3-1   &-1 

2.9  Evaluate:  P(7,  3),  P(5,  5),  P(17,  2),  P(7,  4),  P(7,0). 

2.10  Evaluate:  P(ll;  2,  2,  5,  2),  P(8;  5,  3). 

2.11  Evaluate:  (7(7,  3),  C(5,  5),  C(17,  2),  (7(7,  4),  (7(7,0),  (7(8,  5),  C(8,  — ! 
C(7,  8). 

2.12  Using  the  binomial  expansion  (Theorem  2.11),  and  letting  a  =  6  = 
verify  that  26  =  32.  Write  out  each  term  and  show  its  value. 

2.13  Expand  (i+f)4. 

2.14  Find 

]C  C"(7>  y)'C(5,  4  —  a:  —  y)- 

v—o 

2.15  Evaluate: 

r  °° 

(a)       I      ics/3e  2ic^^ 
*/  o 


rl 

J  0 

«  oo 

I 

•^  o 


2.16  Show  that  C(n,  r)~C(n  — 1,  r)+C(»  — 1,  r-1). 

2.17  A  lot  contains  100  items.  A  single  sample  of  two  items  is  to  be  selected. 
How  many  differently  constituted  samples  are  possible? 

2.18  If 


a 


find:  .4-f-  J?,  4  —  B,  and  AB. 
2.19     If 


find  -45  and 

2.20  Find  the  transposes  of  the  matrices  given  in  Problems  2.18  and  2.19. 
Also   find  the  transposes  of  the  solution  matrices  in  each  of  those 
problems. 

2.21  Find  the  inverses  of  the  matrices  in  Problem  2.18. 

2.22  Transform  the  matrices  in  Problem  2.18  into  diagonal  form. 

2.23  Evaluate  \A\  by  expanding  by  minors  (see  Theorem  2.34)  for 


rl  -3    1-1 

2       1     2 
Ll       5     3J 


Is  A  singular? 

2.24     Solve  the  following  sot  of  equations  using  determinants. 

10 


28  CHAPTER    2,    MATHEMATICAL    CONCEPTS 


References  and  Further  Reading 

1.  Cram6r,   H.   Mathematical  Methods  of  Statistics.   Princeton  University  Press, 
Princeton,  N.J.,  1946. 

2.  Graybill,  F.  A.  An  Introduction  to  Linear  Statistical  Models.  VoL  I.  McGraw- 
Hill  Book  Company,  Inc.,  New  York,  1961. 

3.  Kemeny,  J,   G.,   Snell,   J.   L.,   and   Thompson,    G.    L,   Introduction  to  Finite 
Mathematics.  Prentice-Hall,  Inc.,  Englewoocl  Cliffs,  N.J.,  1957. 

4.  Riordan,  J.  An  Introduction  to  Combinatorial  Analysis.  John  Wiley  and  Sons, 
Inc.,  New  York,  1958. 

5.  Whitesitt,    J.     E.    Boolean    Algebra    and    its    Applications.    Addison-Wesley 
Publishing  Company,  Inc.,  Reading,  Mass.,  1961, 


CHAPTER    3 

A  SUMMARY  OF  BASIC  THEORY  IN 
PROBABILITY  AND  STATISTICS 

As  INDICATED  at  the  beginning  of  Chapter  2,  a  proper  appreciation  of 
statistical  methods  is  difficult  without  an  understanding  of  the  associated 
theory.  If  we  do  not  have  sufficient  grounding  in  the  theory  of  prob 
ability  and  statistics,  the  possibility  of  misapplication  of  methods 
based  on  this  theory  is  enhanced. 

PROBABILITY 

In  general,  statistics  enters  into  scientific  method  through  experi 
mentation  or  observation.  Any  investigation,  is  only  a  means  to  an  end. 
It  is  a  device  for  testing  a  stated  hypothesis  or  for  acquiring  an  amount 
of  knowledge — however  small — from  which  a  conclusion  may  be  drawn. 
Most  statements  resulting  from  scientific  investigations  are  only  in 
ferences.  They  are  uncertain  in  character.  The  measurement  of  this  im- 
certainty  by  use  of  the  theory  of  probability  is  one  of  the  most  important 
contributions  of  statistics. 

Probability  is  just  a  measure  of  the  likelihood  of  occurrence  of  a 
chance  event.  A  fairly  simple  definition  of  probability,  generally  re 
ferred  to  as  the  classical  definition  of  probability,  is : 

Definition  3.1  If  an  event  can  occur  in  N  mutually  exclusive  and 
equally  likely  ways,  and  if  n  of  these  possess  a  charac 
teristic  E,  then  the  probability  of  E  occurring  is  the 
fraction  n/N.  This  is  customarily  written  P(E}  ~n/N. 

There  is  a  natural  relation  between  set  theory  and  probability  theory 
which  is  easily  recognized  once  we  adjust  to  a  change  in  language.  In 
probability,  the  universal  set  is  called  the  sample  space?  each  subset  is 
called  an  event,  and  an  element  is  referred  to  as  a  sample  point.  Then, 
the  definition  of  probability  is: 

Definition  3.2  The  probability  of  occurrence  of  the  event  A  is  the 
ratio  of  the  number  of  sample  points  in  the  event  A 
to  the  number  of  sample  points  in  the  sample  space. 
Symbolically,  P(A}  ~n(A}/N  where  n(A}  is  the  num 
ber  of  sample  points  in  the  event  A,  and  N  is  the  num 
ber  of  sample  points  in  the  sample  space. 

Some  additional  expressions  encountered  in  probability  and  statistics 
are  the  words  experiment  and  outcome. 

C291 


30  CHAPTER    3,    THEORY    IN    PROBABILITY    AND    STATISTICS 

Definition  3.3      An  experiment  is  any  well-defined  action. 
Definition  3.4     Each  possible  result  of  an  experiment  is  called  an  out 
come  (of  the  experiment) . 

The  tie-in  between  the  two  definitions  just  given  and  the  ideas  ex 
pressed  earlier  is  as  follows:  An  outcome  is  a  sample  point,  the  totality 
of  outcomes  is  the  sample  space,  and  an  event  is  a  set  of  outcomes. 

Definition  3.5  A  random  {chance)  variable  is  a  numerically  valued 
function  defined  over  a  sample  space.  It  is  a  rule  which 
assigns  a  numerical  value  to  each  outcome  of  an  experi 
ment. 

Definition  3.6  A  discrete  random  variable  is  one  which  can  take  on 
only  a  finite  or  a  denumerable  number  of  values. 

Definition  3.7  A  continuous  random  variable  is  one  which  can  take  on 
a  continuum  of  values. 

NOTE :  A  one-dimensional  continuous  random  variable 
is  most  easily  thought  of  as  one  wrhich  can  take  on  any 
value  within  a  specified  interval  along  a  straight  line. 

The  definitions  of  probability  advanced  in  the  preceding  paragraphs 
are  such  that  difficulties  are  sometimes  encountered  in  their  use.  For 
example,  it  is  not  always  easy  to  tell  if  two  events  are  equally  likely. 
Then,  too,  how  do  we  handle  the  concept  of  an  experiment  that  can  be 
performed  infinitely  many  times? 

Before  formulating  a  new  definition  that  will  give  us  greater  flexi 
bility,  let  us  examine  some  preliminary  ideas.  Consider  a  random  ex 
periment  8  that  may  be  repeated  many  times  under  uniform  conditions. 
Each  time  the  experiment  is  performed,  observe  whether  an  event  E 
does  or  does  not  take  place.  In  the  first  n  performances  of  8,  E  will  occur 
a  certain  number  of  times,  say/.  We  shall  call  the  ratio  f/n  the  relative 
frequency  of  E  in  the  first  n  performances  of  the  experiment  8.  It  will 
be  observed  that  f/n  will  generally  tend  to  become  more  or  less  con 
stant  for  large  n.  This  phenomenon  is  sometimes  referred  to  as  statisti 
cal  regularity.  It  is  now  conjectured  that  for  given  8  and  E  we  should 
be  able  to  find  a  number  P  such  that  as  n,  the  number  of  performances 
of  8,  gets  very  large,  the  ratio  f/n  should  be  approximately  equal  to  P. 

Definition  3.8  "Whenever  we  say  that  the  probability  of  an  event  E 
with  respect  to  an  experiment  8  is  equal  to  P,  the 
concrete  meaning  of  this  assertion  will  thus  simply  be 
the  following:  In  a  long  series  of  repetitions  of  8,  it  is 
practically  certain  that  the  (relative)  frequency  of  E 
will  be  approximately  equal  to  P."1 

Theorem  3.1          For  any  event  E,  it  is  true  that  Q<P(E}  <1. 

Theorem  3.2          P(J5)  +P(not  E}  =  1. 

NOTE :  Using  the  set  notation  introduced  in  Chapter 
2,  this  would  appear  as  P(E)  +P(I?')  =  1. 

1  H.  Cramer,   Mathematical  Methods  of  Statistics,  Princeton  University  Press 
Princeton,  N.J.,   1946,  p.   148.  ' 


Theorem  3.3 


PROBABILITY 

For  arbitrary  events  A  and  B,  P(A  or  B}  = 


31 


Theorem  3.4 


NOTE  :  For  three  events,  this  extends  to 
(5)+P(C)  —  P(AB}  — 

The  extension  to  more  than  three  events 
can  be  made  quite  easily. 

If  A  and  B  are  mutually  exclusive  (i.e.,  have  no  ele 
ments    in    common),    then    P(A    or 


NOTE:  For  three  events  that  are  pairwise  disjoint 
(i.e.,  mutually  exclusive),  this  extends  to  P(A^JB^JC) 
=  P(A)+PCB)+P(C).  The  extension  to  more  than 
three  events  is  obvious. 

Let  A  be  an  event  in  an  arbitrary  sample  space  such 
that  P(A}  7^0.  Let  B  be  any  event  in  the  same  sample 
space.  Then,  the  conditional  probability  that  B  occurs, 
knowing  that  A  has  occurred,  is  defined  by  P(B\  A) 


Definition  3.9 


Theorem  3.5          For  arbitrary  events  A  and  B,  P(A  and  J?)  =P(AJ3) 


Definition  3.10 


Theorem  3.6 


Theorem  3.7 


NOTE:  For  three  events,   this  extends  to   P(ABC} 

=  P(A)  -P(B\A)  -P(C\AB).  It  should  be  realized  that 

other  permutations  of  the  factors  and  the  letters  are 

possible.  The  extension  to  more  than  three  events  is 

obvious. 

Two  events  A  and  B  are  said  to  be  statistically  inde 

pendent  if  P(A\B)=P(A)  and  P(B  \  A)  =PCB).  This 

is  equivalent  to  saying  that  A  and  B  are  statistically 

independent  if  P(AB)  =  P(A}  -P(B}. 

NOTE:  Three  events  (A,  .B,  and  C)  are  mutually  in 

dependent       if       P(A|J3)=PCA),    P(A|.BC)=P(A), 

P(ABlC')  ~P(AB),  and  so  on  for  all  possible  events. 

This  is  equivalent  to  saying  that  A,  jB,   and  C  are 

mutually  independent  if  A,  B,  and  C  are  pairwise  in 

dependent     [that    is,    PCAJ3)=P(A)-P(B),     P(AC) 

=  P(A)-P(C),     and     P(J5C)  =P(J5)  -P(C)  ],     and     if 

P(ABC}  -P(A)  -P(B)  -P(C). 

If   A    and   B    are   statistically   independent,    P(AJ5) 


NOTE:  For  three  events  that  are  mutually  inde 
pendent,  this  extends  to  P(ABC}  =P(A)  -P(B)  -P(C). 
The  extension  to  more  than  three  events  is  obvious. 
For  any  events  A  and  B, 


P(A  ^J  J5)  =   1  —  P(not 
-!-[{!- 


-P(not  B  \  not  A) 


•P{l  —  P(B\  not  A}}]. 


32  CHAPTER    3,    THEORY    IN    PROBABILITY    AND    STATISTICS 

Theorem  3.8         If  A  and  B  are  statistically  independent,  Theorem  3.7 
becomes 

P(A  VJ  B)  =  1  -  P(not  /I)  •  P(not  B) 


Theorem  3.9 


Theorem  3.10 


Definition  3.11 


Theorem  3,11 


NOTE :  This  can  easily  be  extended  to  k  events  that 
are  mutually  independent  by  writing 


P(El 


=  1  -  [{1 


{1  - 


Let  7/i,  7/2,  -  -  •  ,  7/7i  be  mutually  exclusive  events 
whose  union  is  the  sample  space.  Let  E  be  an  arbitrary 
event  in  the  same  sample  space  such  that 
Then 


Referring  to  Theorem  3.9  and  invoking  the  fact  that 

7^),  it  is  seen  that 


P(7/l 


n  //o 
+"p(//7J 


^|  77.) 
77n.    This 


and  similar  results  hold  for  77$,  -  *  • 
theorem  in  known  as  JB ayes'  theorem. 
ConRidor  an  experiment  with  only  two  possible  out 
comes,  that  i«t  /I  and  A'.  If  at  each  performance  of 
the  experiment  (i.e.,  each  trial),  P(/t)  remains  the 
Hume,  then  the  repeated  trials  are  known  as  Bernoulli 
trials. 

NOTE:  When  clincussing  HeruoxiIH  trials,  it  in  CUH- 
tomary  to  refer  to  one  of  the  two  possible  «ut<iome« 
an  a  success  and  to  the  other  as  a  failure. 
Let  fo(^;  n,  p)  denote  the  probability  that  n  Bernoxilli 
trialn  will  result  in  exactly  x  HUCCOSSOS  and  n  — j?  faii- 
tireH  when  the  probability  of  a  success  at  each  trial  is  p 
and  the  probability  of  a  failure  at  each  trial  in  #«  I 
—  p.  Then  6(x;  n,  p)  —  (^(n?  as)p*ffft""*. 
NOTE:  Probabilities  giveti  by  &(#:;  n,  p)  are  often  re 
ferred  to  as  binomial  probabilities.  Kvaluation  of  bi 
nomial  probabilities  can  be  a  tedious  task.  However, 
tables  (10,  II)  are  available  and  can  be  xined  to  good 
advantage, 


PROBABILITY    DISTRIBUTIONS 


33 


MATHEMATICAL   EXPECTATION 

Definition  3.12  Consider  the  function  6(05;  n,  p)  introduced  in  Theo 
rem  3.11.  The  expected  value  of  x,  which  will  be  de 
noted  by  E[x],  is  defined  by  E[x}-=np.  That  is,  the 
expected  number  of  successes  in  n  trials  is  defined  to 
be  np,  even  if  it  may  be  impossible  to  observe  such  a 
number. 

PROBABILITY   DISTRIBUTIONS 


Definition  3.13 

Definition  3.14 
Definition  3.15 
Theorem  3.12 

Theorem  3.13 


For  any  random  variable  X,  we  will  denote  the 
P(X<x}  by  F(x}.  Further,  F(x)  will  be  referred  to  as 
the  cumulative  distribution  function  (c.d.f.)  or,  simply, 
as  the  distribution  function  (d.f.)  of  the  random  vari 
able  X. 

If  X  is  a  discrete  random  variable,  we  will  define 
the  probability  function  (p.f.)  of  the  random  variable 
X  to  be /(#)==  P(.X  ===  re) . 

If  X  is  a  continuous  random  variable,  we  will  define 
the  continuous  probability  density  function  of  the  ran 
dom  variable  X  to  be  f(x)  —  dF(x)/dx. 
For  a  discrete  random  variable  X,  F(x}  =  ]Cy** /(?/)• 
NOTE:  If  F(x)  is  defined,  f(x)  may  be  obtained  by 
differencing.  However,  close  attention  must  be  given 
to  equality  and  inequality  signs. 
For  a  continuous  random  variable  X, 


F(*)  -    f  * f(y)dy. 

*/    —00 

Theorem  3.14       F(x)  has  the  following  properties: 


(1) 

(2) 
(3) 


if 


Theorem  3.15 


/(a?)  has  the  following  properties: 

(1)  /te)^0. 

(2)  2^  /(#)  =  1  if  X  is  a  discrete  random  variable, 

alia:  ? 


or 


f(x)dx  =  1 


Theorem  3.16 


if  X  is  a  continuous  random  variable. 
For  a  discrete  random  variable  X, 


P(a  <  X  <  6) 


-F(a)  = 


34 


CHAPTER    3,    THEORY    IN    PROBABILITY    AND    STATISTICS 


Theorem  3.17       For  a  continuous  random  variable  X, 

P(a  <  X  <  6)  =  F(b)  -  F(a)  -    f   f(x)dx. 

J  a 

Definition  3.16  For  two  random  variables  X  and  Y,  we  will  denote 
P(X<x,  Y<y)  by  F(x,  y).  Further,  F(x,  y)  will  be 
referred  to  as  the  joint  cumulative  distribution  function 
of  X  and  F, 

Definition.  3.17  If  X  and  Y  are  discrete  random  variables,  the  joint 
probability  function  of  X  and  Y  will  be  denoted  by 


Definition  3,18  If  -X"  and  Y  are  continuous  random,  variables,  the  joint 
probability  density  function  of  X  and  Y  will  be  de 
noted  by 


dxdy 
Theorem  3.18       For  discrete  random  variables  X  and  F, 


Theorem  3.19       For  continuous  random  variables  X  and  F, 

f**         /-i/ 

F(x,  y)  ==    I       ds   I       f(sj  t)dt. 


Theorem  3.20 


Theorem  3.21 


F(x,  T/)  has  the  following  properties: 

111      ft  \    •"""••   OO       ?/ )   EST  ft  ( *JT       — '    CO   )  ssss  /'[   •—   CO      •"-•    OO   )   ?sa  O 
v^.*/*^  ;^//          *     V**^?  /          *\  ?  /          v'* 

/O\       t/T/  -^s         *>*.  N  »—  1 
(^;      /*  (,  00  ;      OO;  as  1. 

(ti)  F(c&,  ?/)  ^^O/)?  which  in  the  marginal  cumu 
lative  distribution  function  of  F, 

(4)  F(*r,  <^o)  — /'\(a;)?  which  in  the  marginal  cutnula~ 
tivr,  distribution  function  of  A*"* 

^C^j  ?/)  h*1^  the  following  properties: 
(1) 


(2) 


-   1      or 


Kit    X      tUt    |/ 


n  whether  A"  and  F  are  discrete  or 
Theorem  3*22       If  X  and  F  arc  discrete,  then 


EXPECTED    VALUES 


35 


P(a  <  X  <  b,  c  <   F  < 


=  F(b,  d)  -  F(b,  c) 
-  F(a,  d)  +  F(a,  c) 


Theorem  3.23 


If  X  and  Y  are  continuous,  then 
P(a  <X<b,c<Y<d)  =  F(b,  d)  — 

-  F(a,  d} 


,  c) 

F(a,  c} 


/b  s*  d 

dx   I      f(oc,  y)dy. 
.j.  +s    n 


Definition  3.19  Associated  with  the  marginal  c.d.f.'s  in  Theorem 
3.20,  we  have  marginal  p.d.f/s  (or  p.f.'s)  denoted  by 
/i(V)  and/2(y),  respectively. 

Definition  3.20  Conditional  p.d.f.'s  (orp.f.'s)  and  c.d.f.'sare  defined  as 
follows  : 

(i) 

(2) 
(3) 
(4) 

Theorem  3.24       If    X    and     Y    are    statistically    independent,    f(&,y} 
=/iO)  -/2(2/)  and  F(x,  y]  =Fx(x)  -F2(?/). 
NOTE:  All  the  definitions  and  theorems  given  for 
two  random  variables  may  easily  be  extended  to  three 
or  more  random  variables. 

EXPECTED  VALUES 

To  aid  in  the  description  of  probability  distributions,  it  is  helpful  to 
know  something  about  their  properties.  Of  special  importance  are  those 
properties  associated  with  the  concept  of  mathematical  expectation. 

The  expected  value  of  any  function  of  a  random  variable  is  defined 
as  the  weighted  average  (weighted  by  the  probability  of  its  occurrence) 
of  the  function  over  all  possible  values  of  the  variable.  Since  expected 
values  are  used  so  much  in  statistics,  a  special  notation  has  been  de 
veloped.  The  symbol  E[  •  -  -  ]  will  be  used  to  denote  the  expected  value 
of  whatever  appears  within  the  brackets.  For  example,  the  expected 
value  of  a  function  B(X}  will  be  denoted  by  E[0(X)]. 


Definition  3.21      &[0(X)]  =    ^    «(«)'/(«),  *  discrete 


all  a; 


/oo 
6(x) 
—  BO 


xy  x  continuous. 


Definition  3.22     E[6(X,  F)]  =    23     X)    #<>>  3>)/O,  y)>  *  and  y  discrete 

all  x     all]/ 


36  CHAPTER    3,    THEORY    IN    PROBABILITY    AND    STATISTICS 


/OO  X»    OO 

dx   I       Q(x7  y)f(x,  y}dy^  x  and  y  continuous. 
^-00  *J  —00 


Theorem  3.25       For  expected  values,  the  following  properties  hold: 

(1)  E[c]  =c  where  c  is  a  constant. 


(2)   E[c6(X,  Y)]==cE[0(X,  F)]. 

[k  ~~|  X* 

T^  c<Q-(X     Y}       =    y^  c>F\9  (Y    V}1 

/  *  6tt7^^-V  ,      •*    /          —      X    /   C'l^iyA*^  >     •*/.!• 
i-1  J        t_l 

Definition  3.23     The  /cth  moment  with  respect  to  the  origin  of  X  Ls 
denoted  by  A^  —  J^pf*]- 

Definition  3.24     MI  is  known  as  the  mean,  and  it  is  commonly  denoted 
by/*. 

Definition  3.25     The  /cth  moment  about  the  moan,  or  the  A;th  central 
moment,  of  X  is  denoted  by  MA-  =  AT[(^Y  —  /*)*]. 

Definition  3.26     /xa  is  known  as  the  variance,  and  it  is  commonly  de 
noted  by  or2. 

Definition  3.27     The  positive  square  root  of  the  variance,  &,  is  known 
as  the  standard  deviation. 

Theorem  3.26       <r*=*iJi!i~-tA***K[X*]  —  (ft[X])*. 

Definition  3.28     When  dealing  with  two  random  variables,  the  product 
moments  of  X  and    F  are  defined  by  /*Jt,  =»  J$[XrY*]. 

Definition  3.29     The  central  product  moments  are  defined  by 

-Mv>)*i     where     a  v  = /£  LY 1      and 


Definition  3.30     MIA  is  known  as  the  covariance  of  J\T  and  F,  and  it  is 
commonly  denoted  by  <rXK. 

NOTE:  The  variance  of  X,  <r^t  is  sometimes  written 
as  &XX*  Similarly,  ^,  — <rrr*  These  alternative  nota 
tions  show  the  close  relation  between  variances  and 
eovarianeoB. 

Definition  3-31     The  product  moment  correlation  between  -Y  and  F  is 
defined  by  PXY^VXY/VX&Y*  **  should   be  rioted   that 


Theorem  3,27      J 

Theorem  3*28       If  X  and  F  are  statistically  independent, 


OTHER   DESCRIPTIVE   MEASURES 

Definition  3,32  The  value  of  a?  such  that  F(JT)  «p  m  called  the  100^ 
fracttte  of  the  distribution  of  the  random  variable  A". 

Definition  3*33  When  p=»tK5  in  Definition  3.32,  the  corresponding 
vahie  of  x  is  known  as  the  median  of  the  distribution. 

Definition  3*34  The  mode  of  a  distribution  is  thut  value  of  ^  for  which 

is  a  maximum. 


PROBLEMS  37 

SPECIAL   PROBABILITY    DISTRIBUTIONS 

Certain  distributions  occur  so  often  in  statistical  problems  that  they 
merit  special  attention.  Some  of  these  are  tabulated  in  Tables  3.1  and 
3.2.  Since  most  applications  involving  these  distributions  require  the 
use  of  probabilities  associated  with  the  distributions,  it  is  convenient 
to  have  available  adequate  tables  of  such  probabilities.  Accordingly, 
tables  for  the  Poisson,  standard  normal,  chi-square,  "Student's"  t,  and 
F  distributions  are  presented  in  Appendices  2  through  6.  Each  of  these 
tables  is  given  in  cumulative  form,  that  is,  in  terms  of  the  cumulative 
distribution  function,  so  that  the  reader  will  have  to  learn  only  one 
method  of  reading  the  tables. 

Problems 

3.1  A  sample  of  3  TV  sets  is  selected  from  a  lot  of  30  sets.  If  there  are  5 
defective  sets  in  the  lot,  what  is  the  probability  the  sample  will  con 
tain  no  defectives?  3  defectives?  1  defective  and  2  nondefectives? 

3.2  A  buyer  will  accept  a  lot  of  10  TV  sets  if  a  sample  of  3,  selected  at 
random,  contains  no  defective  sets.  What  is  the  probability  of  accept 
ing  a  lot  of  10  that  contains  5  defectives? 

3.3  An  electrical  circuit  consists  of  4  switches  in  series.  Assume  that  the 
operations  of  the  4  switches  are  statistically  independent.  If  for  each 
switch  the  probability  of  failure  (i.e.,  remaining  open)  is  0.02,  what  is 
the  probability  of  circuit  failure? 

3.4  Rework  the  preceding  problem  for  the  case  where  the  circuit  consists 
of  4  switches  in  parallel. 

3.5  Defects  are  classified  as  type  A,  B,  or  C,  and  the  following  probabilities 
have  been  determined  from  available  production  data:  P(A)=0.20, 
P(#)=0.16,  jP(C)=0.14,  P(.AB)=0.08,  PC-AC)  =0.05,  P(BC)-0.04, 
and  P(ABC}  =0.02.  What  is  the  probability  that  a  randomly  selected 
item  of  product  will  exhibit  at  least  one  type  of  defect?  If  an  item 
exhibits  at  least  one  type  of  defect,  what  is  the  probability  that  it 
exhibits  both  A  and  B  defects? 

3.6  An  electrical  assembly  consists  of  two  parts  connected  in  series  in  the 
order:  A  followed  by  B.  The  probability  that  part  A  is  defective  is 
0.025  and  the  probability  that  part  B  is  defective  is  0.011.  What  is  the 
probability  of  having  a  defective  assembly?  A  nondefective  assembly? 
An  assembly  that  fails  only  because  part  B  is  defective? 

3.7  Suppose  the  probability  that  a  certain  piece  of  air-borne  electronic 
equipment  will  not  be  in  working  order  after  its  first  flight  is  0.40, 
and  the  probability  of  failure  drops  to  one-half   its  previous  value 
after  each  succeeding  flight.    (Assume  no  repair  and  replacement,) 
What  is  the  probability  the  equipment  will  be  in  working  order  after 
three  flights?  After  four  flights  given  it  has  survived  two  flights? 

3.8  Consider  a  four-engine  aircraft  (two  on  each  wing)  where  the  prob 
ability  of  an  engine  failure  is  0.05.  Assume  that  the  probability  of  one 
engine  failing  is  independent  of  the  behavior  of  the  others.   What  is 
the  probability  of  a  crash  if  the  plane  can  fly  on  any  two  engines? 
If  the  plane  requires  at  least  one  engine  operating  on  each  side  in  order 
to  remain  in  the  air? 


8 

I   ^                       ^ 

1 

fe  ^                                 ^                                  « 

^ 

^    S;                                        ft 

d 
S 
3 

Ci                                 ft                  x            "^ 

ii 

§ 

If) 
.2 

1 

-     S1 

j—  i 

.a 

s 

-=,     + 

P 

•2 

fe; 

^o 

jj* 

|            <J               ft 

•  i-H 

1 

\        ^    1?     ."     . 

. 

«J 

P4 

.+.     2>  Q,    I     1        ".    ^    '•             « 

^H        ft,             '.        VH        p^ 

O 

^      S  ,9  ^  ^     •*   v  °^          ^     ^ 

v    i      -r  v    | 

PH 

"        d     9       *•      •*       .-T1    r\       "<       —  ^.    —  T*       ^—  r 

<3             H       p      TH      YH           O       HH    i-H           C.5      O           O 

^     vH            O"       5X     ^H 

<D 

II         II      II      II      II         II     V     II        A     II         II 

V     II         II     V     H 

1 

H              «       *         *       fe;              «      0         8             X         »              H 

0      <=*         H     0      c* 

R 

1 

T                            & 

>»—  1 

1 

(J~               1                ^ 

CO 

fc 

ft                                                  PM 

S 

^3 
§ 

>,       SS                                                                  J4                                                                          ^ 

PQ 

<* 

H 

'£ 

u 

^ 

9"                        1            |        S 

o                            o               v>          u 

ii                           K                ||           j| 

k 

II 

wm^,^..™  ,, 

3                      5*            3"        S 

I 

*•>                                                                         o 

»°* 

'S                                                                       o 

»*< 

«                                                 i 

9                                                                 5 

%o 

M                                                                    ^OT                                          _                            •* 

•S 

g                           §                 *"O 

« 

{5*                      w            fS        te 

S 

t38] 


I 

cj 


-f 


+ 


^Qj     G 

I      I! 


+ 
^ 

fO 


H- 


+     " 


.3 


O 

I 


on 


tri 


8 
V 

H 

V 
8 
I 


V 

V 
8 


A    A 


A 


A 


o 
A 


V 

V 

o 


A   A 


O     O     0 

A    A   A 


& 

CN 


CS 

CO 


pp 


| 
O 

13 
g 


. 


t 


No 


1 
J3 

GO 


2 

i 


[391 


CSJ 

A 


Vari 


n 

3 


W 
oq 


§ 

•B 

3 
8 
rt 


8 
V 

V 


A    of 

^  ^r 

If 
ff 


I 


it- 


t 


+ 


t 

t 

g. 


11 

c 


y 
3 


I 
I 

[401 


PROBLEMS  41 

3.9  Suppose  3  defective  dry  cells  are  mixed  in  with  7  nondefectives,  and 
you  start  testing  them  one  at  a  time.  What  is  the  probability  that  you 
will  find  the  last  defective  on  the  sixth  test? 

3.10  Three  operators  (A,  B,  and  <7)  alternate  in  operating  a  certain  ma 
chine.  The  number  of  parts  produced  by  A,  B,  and  C  are  in  the  ratio 
3:4:3  and,  of  the  parts  produced,  1  per  cent  of  A's,  2  per  cent  of  B's, 
and  5  per  cent  of  C*B  are  defective.  If  a  part  is  drawn  at  random  from 
the  output  of  their  machine,   what  is  the  probability  it  will  be  de 
fective? 

3.11  Referring  to  Problem  3.10,  what  is  the  probability  that,  if  a  defective 
part  is  selected,  it  was  produced  by  A?  by  B?  by  C? 

3.12  Iif(x9  ?/)=exp   {  —  (rc+2/)}  for  x>0,  y  >0,  find: 

(a)    /,(x),    (b)    My},    (c)    Fi(x),    (d)    F*(y}9    (e)    f(y\x),    (f)    f(x\y), 
(g)  F(x,  2/),  (h)  F(y\x),  (i)  F(x\y).  . 

3.13  If /(a,  $/)  =3z  for  0  <y  <rc,  0  <x  <1,  find  the  same  functions  as  asked 
for  in  the  preceding  problem. 

3.14  If  f(x9  y)  =24?/(l  —  x  —  y}  over  the  triangle  bounded  by  the  axes  and 
the  line  x  +  y  =  1,  find  the  same  functions  as  asked  for  in  Problem  3.12. 

3.15  During  the  course  of  a  day,  a  machine  turns  out  either  0,   1,  or  2 
defective  items  with  probabilities  |,  f,  and  £,  respectively.  Calculate 
the  mean  and  variance. 

3.16  Given  that  the  number  of  accidents  occurring  at  a  particular  inter 
section  between  10:00  P.M.  and  midnight  on  Saturday  is  0,  1,  2,  3,  or  4 
with  probabilities  0.90,  0.04,  0.03,  0.02,  0.01,  respectively,  determine 
the  expected  number  of  accidents. 

3.17  Suppose  that  the  life  in  hours  of  a  certain  type  of  tube  has  the  p.d.f. 
/(#)  =  a/x*}  #>500,   and /0*0  =0,   x  <5QQ.  Find  the  c.d.f.   Determine 
the  mean  and  variance.  What  is  the  probability  a  tube  will  last  at 
least  1,000  hours? 

3.18  A  submarine  carries  three  missiles.  Assuming  the  only  error  is  in  one 
direction  (e.g.,  a  range  error  but  no  sideways  error)  and  that  a  hit 
within  40  miles  of  the  target  is  considered  a  success,   compute  the 
probability  of  a  successful  operation  (i.e.,  an  operation  in  which  at 
least  one  hit  is  a  success)  if  all  three  missiles  are  launched  and  the 
error  p.d.f.  is: 

/(#)  =  (lOO+aO/10,000  - 100  <x  <0 

«(100  — aO/10,000  0<x<100 

=  0  elsewhere. 

3.19  Referring  to  the  previous  problem,   the  submarine  can  carry  eight 
missiles  of  a  smaller  sisse.  However,  in  this  case  a  hit  must  be  within 
15  miles  to  be  successful.  Assuming  the  same  p.d.f.,  should  the  light 
or  heavy  missiles  be  used? 

3.20  A  service  station  will  be  supplied  with  gasoline  once  a  week.  Its  weekly 
volume  of  sales  in  thousands  of  gallons  is  predicted  by  the  p.d.f. 
f(x')  =  5(1  — x)4  for  0  <x  <1.  Determine  what  the  capacity  of  its  under 
ground   tank  should   be  if  the   probability  that  its   supply   will  be 
exhausted  in  a  given  week  is  to  be  0.01. 

3.21  Show  that  the  correlation  between  two  random  variables  is  0  if  they 
are  statistically  independent. 

3.22  Let  X  have  the  marginal  density  /i(#)  =1  for  —  %  <x  <£,  and  let  the 
conditional  density  of  Y  given  X  be 


42  CHAPTER    3,    THEORY    IN    PROBABILITY    AND    STATISTICS 


=  1;   —  x  <y  <1  —  x,  0  <x 
=  0;  elsewhere. 

Find  the  correlation,  between  X  and  F.  Discuss  the  relationship  be 
tween  eon-elation,  and  statistical  independence. 

3.23  A  process  is  producing  parts  that  are,  on  the  average,  1  per  cent  defec 
tive.   Ten   parts   are  selected   at  random   from   the   process  and   the 
process  is  stopped  if  one  or  more  of  the  ten  are  defective.  What  is  the 
probability  that  the  process  will  be  stopped? 

3.24  In  inspecting  1,000  welded  joints  performed  by  a  certain  welder  using 
n  specific  process,  150  defective  joints  were  discovered.  If  the  welder 
is  about  to  weld  5  joints,  what  is  the  probability  of  getting  no  defective 
joints?  of  one?  of  two?  of  two  or  more?  Discuss  any  assumptions  you 
make  in  solving  this  problem. 

3.25  A  large  number  of  rivets  is  used  in  assembling  an  airplane.    It  has 
been  determined  that  the  probability  distribution  for  the  number  of 
defective  rivets  is  Poisson  with  X  =  2.  Find  the  probability  that  the 
number  of  defective  rivets  in  a  plane  will  be  no  more  than  two. 

3.26  Suppose  there  is  an  average  of  1  typographical  error  per  10  pages  in 
a  certain  book.  What  is  the  probability  that  a  30-page  chapter  will 
contain  no  errors? 

3.27  A  telephone  vswitchboard  handles,  on  the  average,  600  calls  during  the 
rush  hour.  The  board  eaix  make  a  maximum  of  20  connections  per 
minute.  What  is  the  probability  the  board  will  be  overtaxed  in  any 
given  minute  during  the  rush  hour? 

3.28  Assuming  a  normal  distribution,  find: 

(a)  P(  —  3  <  Y  <  —  1) ;  given  ;x  =  0;  tr  =«=  1 . 

(6)  7>(  — 3<F<0.r>);  given  M»0,  <r«l. 

(C)  p(  —  8<F  <0);  Rivon/A  — 2,  <r*^4 

02)  P(4<F<50);  given  M«*  —0.1 ,  <r*«4 

O)  J^FSrS);  Rivon/A-0,  <ra«l 

CO  /*  ( Y  <  -  3)  ;  gi  ven  M  «  2,  cr*  -  4. 

3.29  AnBuming  a  clu-nquare  (liHtributum,  find: 

c«) 


)  for  v«lf> 
(r)    P(23.8  <xs  <3(K4)  for  v  «  24 
(d)  P(x 


3.30     ARBuminp  a  ^-cliHtri{)utu>n,  find: 
(a)   />C|«|  >  2.01  5)  for  p-5 
(6)    P(«>2X)15)  for  ^-5 
(c)    ;>(  —  1.341  «<2.  121)  for  p«1 
(rf)   /*(^<l.r>)  for  »«20. 

3*31      AHHuming  an  /^-diHtribution,  find: 

for  Vi-11,  vt*tt 


(r)    W  >  7,79)  for  ^  «  «,  va  -  11 
(rf)   7*  (0.221  </^S2.62)  for  n«6, 


REFERENCES    AND    FURTHER    READING  43 

3.32  The  finished  diameter  on  armored  electric  cable  is  normally  distributed 
with  mean  0.77  inch  and  standard  deviation  0.01  inch.  What  is  the 
probability  the   diameter  will  exceed  0.795  inch?   If  the  engineering 
specifications  are  0.78  ±0.02  inch,  what  is  the  probability  of  a  defec 
tive  piece  of  cable? 

3.33  If  the  p.d.f.  for  the  life  of  a  certain  type  of  component  is  /(#)  =  (1/100) 
exp    { — rc/lOO}    for  x  >0,   what  is   the   probability  that  a  randomly 
selected  component  will  last  400  hours?  That  it  will  last  400   hours 
given  that  it  has  already  survived  200  hours?   If  an  assembly  uses 
three  of  these  components   in  series,   what  is  the  probability  that  an 
assembly  incorporating  three  randomly  selected  components  will  not 
fail  because  of  component  failure? 

3.34  The  hazardrate  is  defined  as /(#)/{  1 —F  (or)  Mf/O)  =  (1/0)  exp  {  —  x/6} 
for  x  >0,  what  is  the  hazard  rate? 

References  and  Further  Reading 

1.  Anderson,  R.  L.,  and  Bancroft,  T.  A.  Statistical  Theory  in  Research.  McGraw- 
Hill  Book  Company,  Inc.,  New  York,  1952. 

2.  Brownlee,  K.  A.  Statistical  Theory  and  Methodology  in  Science  and  Engineer 
ing.  John  Wiley  and  Sons,  Inc.,  New  York,  1960. 

3.  Cramer,  H.  Mathematical  Methods  of  Statistics.  Princeton  University  Press, 
Princeton,  N.J.,  1946. 

4.  Derman,  C.,  and  Klein  M.  Probability  and  Statistical  Inference  for  Engineers. 
Oxford  University  Press,  New  York,  1959. 

5.  Hald7   A.  Statistical  Theory  with  Engineering  Applications.  John  Wiley  and 
Sons,  Inc.,  New  York,  1952. 

6.  Irluntsberger,   13.  V.  Elements  of  Statistical  Inference.  Allyn  and  Bacon,  Inc., 
Boston,  1961. 

7.  Kemeny,   J.    G.,    Snell,   J.   L,.,   and   Thompson,    G.   L.   Introduction  to  Finite 
Mathematics.  Prentice-Hall,  Inc.,  Englewood  Cliffs,  N.J.,  1957. 

8.  Mood,    A.    M.    Introduction   to  the    Theory   of  Statistics.    McGraw-Hill   Book 
Book  Company,  Inc.,  New  York,  1950. 

9.  Mosteller,  F.,  liourke,  H.  E.  K.,  and  Thomas,  G.  B.  Probability  and  Statistics. 
Addison- Wesley  Publishing  Company,  Inc.,  Reading,  Mass.,  1961. 

10.  National  Bureau  of  Standards.  Tables  of  the  Binomial  Probability  Distribution. 
Applied    Mathematics   Series   6.  U.S.  Govt.    Print.   Off.,  Washington,   D.C., 
1949. 

11.  rtomig,   H.   G.,   5Q—1OO  Binomial   Tables.  John   Wiley  and  Sons,   Inc.,    New 
York,  1953. 

12.  Wadsworth,  G.  P.,  and  Bryan,  J.  G.  Introd^lction  to  Probability  and  Random 
Variables.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1960. 

13.  Whitesitt,  J-  El.  Boolean  Algebra  and  its  Applications.  John  Wiley  and  Sons, 
Inc.,  New  York,  1961. 


CH APTE  R    4 

ELEMENTS  OF  SAMPLING  AND 
DESCRIPTIVE  STATISTICS 

IN  THIS  CHAPTER  we  shall  discuss  the  basic  ideas  of  sampling  and  the 
presentation  of  sample  data.  Certain  useful  statistics  of  a  summarizing 
nature  will  be  defined  and  efficient  methods  of  calculation  outlined. 

To  begin  the  discussion  of  sampling,  the  reason  for  taking  samples 
should  be  mentioned.  The  reason  is  usxially  one  of  the  following: 
(1)  Due  to  limitations  of  time,  money,  or  personnel,  it  is  impossible  to 
study  every  item  in  the  population;  (2)  the  population,  as  defined,  may 
not  physically  exist;  (3)  to  examine  an  item  may  require  that  the  item 
be  destroyed. 

Before  proceeding  to  the  actual  mechanism  of  obtaining  samples  and 
the  analysing  of  data  therefrom,  it  will  be  wise  to  define  some  terms 
frequently  encountered* 

4.1      THE  POPULATION  AND  THE  SAMPLE 

In  statistical  work  it  is  important  to  know  whether  we  are  dealing 
with  a  complete  population  of  observations  or  with  a  sample  of  ob 
servations  selected  from  a  specified  population. 

A  population  is  defined  as  the  totality  of  all  possible  values  (measure 
ments  or  counts)  of  a  particular  characteristic  for  a  specified  group  of 
objects.  Such  a  specified  group  of  objects  is  called  a  universe.  Obvkmsly 
a  universe  can  have  several  populations  associated  with  it.  Some  exam 
ples  of  universes  and  populations  are : 

(1)  The  employees  of  Arizona  State  University  as  of  5:00  P.M.  on 
December  4,  1902. 

(2)  Associated  with  the  preceding  universe  are  many  populations, 
for  example,  the  population  of  blood  types,  the  population  of 
weights,  the  population  of  heights,  etc, 

(3)  The  univerae  of  all  single-dwelling  units  in  Tempe,  Arizona,  on 
December  31,  1902, 

(4)  Associated  with  thin  universe  of  single-dwelling  tmits  are  such 
populations  as  the  number  of  rooms  per  unit,  the  number  of 
people  residing  in  each  unit,  and  so  on. 

(5)  A  universe  may  contain  only  one  object,  such  an  a  piece  of 
steel  pipe,  and  the  population  consists  of  all  possible  measure 
ments  of  its  inside  diameter. 

(6)  A  universe  might  consist  of  all  vacuum  tubes  of  a  specific  type 
manufactured  by  a  given  manufacturer  under  similar  condi 
tions, 

C44I 


4.2       TYPES    OF    SAMPLES  45 

(7)    Populations    associated    with    the    preceding    universe    are: 
lengths  of  life,  function  on  test,  etc. 

These  examples  should  suffice  to  impress  upon  the  reader  the  impor 
tance  of  clearly  defining  the  population  under  investigation. 

The  concept  of  a  sample,  as  opposed  to  a  population,  is  very  im 
portant.  A  sample  is  just  a  part  of  a  population  selected  according  to 
some  rule  or  plan.  The  important  things  to  know  are:  (1)  that  we  are 
dealing  with  a  sample  and  (2)  which  population  has  been  sampled. 

If  we  are  dealing  with  the  entire  population,  our  statistical  work  will 
be  primarily  descriptive.  On  the  other  hand,  if  we  are  dealing  with  a 
sample,  the  statistical  work  will  not  only  describe  the  sample  but  also 
provide  information  about  the  sampled  population. 

4.2     TYPES  OF  SAMPLES 

There  are  several  types  or  classes  of  samples  encountered  in  practice. 
The  characteristics  which  distinguish  one  type  from  another  are :  (1)  the 
manner  in  which  the  sample  was  obtained,  (2)  the  number  of  variables 
recorded,  and  (3)  the  purpose  for  which  the  sample  was  drawn.  The 
last  two  characteristics  listed  are  easily  understood  in  any  practical 
situation  although  No.  3  is  frequently  not  clearly  stated  and  perhaps 
even  forgotten.  The  manner  of  obtaining  the  sample  is  very  important 
and  will  be  discussed  further. 

Samples  may  be  grouped  into  two  broad  classes  when  their  method 
of  selection  is  considered,  namely,  those  which  are  selected  by  judg 
ment  and  those  which  are  selected  according  to  some  chance  mecha 
nism.  Samples  selected  according  to  some  chance  mechanism  are 
known  as  probability  samples  if  every  item  in  the  population  has  a 
known  probability  of  being  in  the  sample.  In  particular,  if  each  item  in 
the  population  has  an  equal  chance  of  occurring  in  the  sample,  then  the 
sample  is  known  as  a  random  sample. 

Why  are  random  samples  preferred  to  subjectively  selected  samples? 
An  answer  to  this  question  may  be  formulated  as  follows:  A  good 
sample  is  one  from,  which  generalizations  to  the  population  can  be 
made ;  a  bad  sample  is  one  from  which  they  cannot  be  made.  To  general 
ize  from  a  sample  to  a  population,  we  need  to  be  able  to  deduce  from 
any  assumptions  about  the  population  whether  the  observed  sample  is 
within  the  range  of  sampling  variation  that  might  occur  for  that  popu 
lation  under  the  given  method  of  sampling.  Such  deductions  can  be 
made  if,  and  only  if,  the  laws  of  mathematical  probability  apply.  The 
purpose  of  randomness  is  to  insure  that  these  laws  do  apply.  If  we  had 
equally  well-established  and  stable  laws  of  personal  bias,  subjective 
sampling  could  be  used. 

We  can  sample  from  different  populations  in  various  ways: 

(1)   A  random  sample  may  be  drawn  from  a  population  specified 
by  a  continuous  probability  density  function.  In  this  case,  the 


46  CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 

question  of  sampling  with  or  without  replacement  does  not 
arise* 

(2)  A  random  sample  may  be  drawn  from  an  infinite  population 
specified  by  a  discrete  probability  density  function.  Again,  the 
question  of  with  or  without  replacement  does  not  arise. 

(3)  A  random  sample  may  be  drawn   from  11  finite   population 
(specified  by  a  discrete  probability  density  function)  where  the 
sampling  is   performed   with  replacement.   Sampling  with  re 
placement  effectively  makes  the  population  infinite. 

(4)  If  sampling  from  a  finite  population  is  performed  without  re 
placement,   we  no  longer  have  a  random  sample  as  defined 
earlier.  Sometimes,  a  "random"  sample  for  this  situation  is  de 
fined  as  one  in  which  each  set  of  n  objects  has  an  equal  chance 
of  being  the  sample  of  size  n. 

Other  types  of  samples  of  a  specialized  type  are  sometimes  en 
countered.  Two  of  these  are: 

Stratified  Random  Sample 

The  population  is  first  subdivided  into  subpopulations  or  strata. 
Then  a  simple  i^andom  sample  is  drawn  from  each  stratum. 
Systematic*  Random  Sample 

Consider  the  N  units  in  the  population  to  be  arranged  in  some 
order.  If  a  sample  of  sixe  n  is  required,  take  a  unit  at  random  from 
the  fir«t  k^N/n  units  and  then  take  every  fcth  unit  thereafter. 

Having  defined  the  various  types  of  sampling  frequently  encountered, 
the  following  caution  is  noted:  The  methods  of  analysis  will  not  be  the 
same  for  each  type  of  sampling,  (treat  care  must  bo  exereised  to  use  the 
proper  method  of  analysis;  fuihtrc  to  do  HO  oan  load  to  serious  errors  in 
judgment  when  the  decision-making  stage  is  reached. 

4,3    SAMPLING;  FROM  A  SPECIFIED  POPULATION 

How  do  we  go  about  .selecting  a  sample  from  a  specified  population? 
Some  examples  will  serve  as  explanation:  (I)  Suppose  the  population 
consists  of  only  two  values*.  One  of  them  can  be  selected  at  random  by 
tossing  ati  uubiaHcd  coin.  (2)  Consider  a  population  consisting  of  100 
items.  One  hundred  numbered  tickets  (corresponding  to  our  population 
of  items)  can  be  placed  in  a  howl  and  tickets  selected  in  a  chance 
manner.  (3)  In  the  previous  example,  the  sample  values  could  have  been 
selected  using  a  table  of  random  numbers. 

To  Illustrate  the  use  of  a  table  of  random  numbers,  consider  the 
problem  of  obtaining  a  sample  of  n**r*  batteries  from  a  lot  of  A^  — 25, 
First,  number  the  batteries:  01,  02,  •  •  •  ,  25.  Second,  refer  to  a  table 
of  random  numbers  such  an  given  in  Appendix  7  and  proceed  through 
the  following  steps. 

(t)   Select  by  any  method  one  of  the  four  pages  of  tabled  values. 

(2)    Without  direction,  bring  a  pencil  point  down  on  the  printed 
page  BO  an  to  hit  a  random  digit. 


4.4       PRESENTATION    OF    DATA  47 

(3)  Read  this  digit  and  the  next  three  to  the  right,  for  example, 
2167. 

(4)  Let  the  first  two  of  these  specify  the  row  and  the  last  two  the 
column. 

(5)  Go  to  this  point  in  the  table  of  random  numbers  and  read  the 
specified  digit  and  the  next  one  to  the  right.  This  reads  73. 
However,  the  only  possible  numbers  of  use  in  the  specified 
problem  are  01,  02,  -  •  -  ,25.  Thus,  it  is  necessary  to  run  down 
the  column  until  five  suitable  numbers  are  observed.  In  order 
the  numbers  observed  are  73,  48,  54,  01,  18,  38,  60,  70,  44, 
30,  41,  86,  23,  64,  31,  71,  68,  64,  13,  12.  The  numbers  specifying 
the   five   batteries  to   be  included  in  the  sample   have  been 
underscored. 

(6)  Appropriate  changes  should  be  made  in  step  No.  5  to  handle 
different  problems. 

4.4      PRESENTATION    OF   DATA 

Having  obtained  a  random  sample  from  a  specified  population,  some 
way  of  reducing  it  to  an  understandable  form  is  called  for.  To  illustrate 
the  usual  techniques  for  presenting  such  data,  consider  the  data  in 
Table  4.1. 

In  this  form  the  data  are,  to  say  the  least,  confusing.  It  is  not  easy  to 
visualize  any  pattern  in  the  observed  values,  nor  is  it  easy  to  estimate 
the  average  function  time.  We  find  it  convenient,  therefore,  to  arrange 
the  values  in  a  frequency  distribution  as  in  Table  4.3.  To  accomplish 
this,  we  first  make  use  of  a  tally  sheet  as  shown  in  Table  4.2.  Inciden 
tally,  Table  4.3  provides  us  with  an  array,  that  is,  the  values  arranged 
in  order  of  magnitude. 

Upon  examination  of  Table  4.3,  we  note  that  all  the  observations  are 
greater  than  or  equal  to  59  milliseconds  and  less  than  or  equal  to  70.5 
milliseconds.  That  is,  we  have  established  the  range  of  our  data.  Fur 
ther,  we  can  roughly  estimate  the  average  function  time  to  be  65  milli 
seconds. 

However,  since  it  takes  too  long  to  scan  all  the  values  in  Table  4.3, 
the  data  are  still  in  rather  cumbersome  form.  To  remedy  this,  it  is 
customary  to  condense  the  data  even  more  by  tabulating  only  the  fre 
quencies  associated  with  certain  intervals,  usually  referred  to  as  class 
intervals.  To  set  up  class  intervals,  a  good  working  rule  is  to  have  no 
fewer  than  5  and  no  more  than  15  intervals.  Also,  the  limits  of  the  class 
intervals  should  be  chosen  so  that  ther^  is  no  ambiguity  in  assigning 
observed  values  to  the  classes.  This  latter  requirement  is  most  easily 
satisfied  by:  (1)  selecting  class  limits  which  carry  one  more  decimal 
place  than  the  original  data,  or  (2)  proper  use  of  inequality  and  equality 
signs.  We  shall  adopt  the  second  of  these  two  procedures  in  this  text. 
Using  class  intervals  of  width  1  millisecond,  we  get  the  data  in  the  form 
of  Table  4.4.  In  this  table  we  have  used  the  letter  X  to  represent  the 
various  function  times  in  milliseconds.  To  interpret  the  values  and 


48  CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 

TABLE  4,1-Function  Times  of  201  Explosive  Actuators 

Measured  in  Milliseconds 

(Hypothetical  Data) 


64.0 

61,5 

69.0 

65.25 

69.0 

66.0 

63.5 

65.25 

66.25 

67.25 

67.25 

62.5 

61.75 

63,5 

63.75 

66.5 

66.0 

65.5 

65.25 

66.5 

64,5 

67.75 

64.5 

68,0 

63  .  75 

68.0 

70.5 

68.0 

65.0 

62.0 

62  .  75 

61.5 

60.0 

65.75 

66.0 

62.0 

65  .  75 

60.75 

63.75 

62,0 

70.25 

64.75 

68.5 

65.0 

66,5 

64.0 

67.0 

67.0 

63.0 

64.0 

67.0 

63.25 

65.25 

67.5 

65,0 

67.5 

64.5 

68.0 

63  .  5 

68,75 

63.0 

66.25 

67.0 

65.25 

64,0 

65.25 

63.0 

67.0 

65,5 

62.0 

64.5 

66.25 

65.0 

63.75 

67.5 

65.5 

64.75 

67.0 

68.0 

59.0 

64.5 

67.0 

67,75 

63.25 

63.25 

65.5 

64.0 

67.0 

64.5 

67,5 

65,0 

61,0 

64.5 

63.0 

66.5 

66.0 

65.0 

61.25 

69.5 

64,0 

68,0 

64.5 

66.5 

64,25 

65.0 

62.25 

63.5 

63.0 

67,0 

65  .  25 

65,0 

65  .  0 

65.25 

65.25 

63,0 

65  .  5 

65,0 

62,0 

64,0 

62.5 

64,75 

61.5 

62  .  75 

68.5 

63.5 

63,0 

64,5 

67,0 

61.75 

66.25 

64.75 

65.5 

62.75 

68,5 

61,5 

63,0 

65,5 

65.5 

63.0 

65.5 

66.75 

69.5 

65.25 

63  .  5 

66.0 

62.25 

62,5 

61.5 

68,0 

63  .  75 

66.0 

64.0 

67,0 

67.75 

65.25 

67.75 

68.0 

63.5 

63,25 

63,0 

61.75 

69,0 

65.0 

62.5 

62,0 

64,75 

64.0 

66.75 

66,0 

64.5 

64.25 

62.5 

66.5 

66,75 

64.5 

60.0 

65,0 

66.0 

64.5 

66.25 

65*75 

65.5 

64,5 

62.0 

65,25 

64.25 

63.0 

64.0 

66.75 

65  .  25 

63,75 

67.0 

61.0 

70.0 

70.0 

65  ,  5 

65  „  25 

64.5 

67,5 

65.75 

70.0 

frequencies,  we  proceed  as  follows:  One  actuator  had  a  function  time 
of  more  than  58  milliseconds  but  less  than  or  equal  to  59  miHinoconda; 
two  actuators  had  a  function  time  of  more  than  59  miIli«econdB  tmt  le«B 
than  or  equal  to  00  milliseconds;  ami  so  on.  Pleane  note  that  we  have 
less  information  available  in  Table  4,4  than  in  Table  4.3,  This  IB  be 
cause  we  no  longer  know  the  individual  values  but  only  in  which  clans 
interval  they  fall.  But  the  lo«a  in  accuracy  is  balanced  to  &ome  extent 
by  the  gain  in  conciseness.  The  column  headed  "relative  frequency" 
tella  us  what  proportion  of  the  total  observations  fall  in  each  class.  The 
valuen  are  foxmd  by  dividing  each  elans  frequency  by  the  total  fre 
quency. 


4.4       PRESENTATION    OF    DATA 


49 


Conforming  to  the  adage  that  a  "picture  is  worth  ten  thousand 
words/ '  we  often  represent  our  distribution  by  a  chart  or  frequency 
histogram.  This  is  illustrated  by  Figure  4.1.  The  dotted  line  pictures  a 
frequency  polygon.  Note  that  the  frequency  histogram  is  formed  by 
erecting  rectangles  over  the  class  intervals,  the  height  of  each  rectangle 
agreeing  with  the  class  frequency  if  the  left-hand  scale  is  read,  and  with 
the  class  relative  frequency  if  the  right-hand  scale  is  read.  The  fre 
quency  polygon  is  formed  by  joining  the  midpoints  at  the  tops  of  the 
rectangles. 

It  is  also  to  be  noted  that  the  frequency  histogram  and  polygon,  as 
well  as  the  frequency  distribution,  give  us  not  only  an  estimate  of  the 
average  value  but  also  an  idea  of  the  amount  of  variability  present  in 
the  data. 

Another  convenient  way  of  tabulating  data  is  to  prepare  a  cumu 
lative  frequency  distribution  showing  the  number  of  observations  less 
than  or  equal  to  a  specified  value.  The  figures  are  obtained  by  adding,  in 
cumulative  fashion,  the  frequencies  recorded  in  Table  4.4.  This  is 
illustrated  in  Table  4.5.  The  graph  which  arises  from  this  table  is  shown 

TABLB  4.2-Tally  Sheet  for  Data  of  Table  4.1 


Function 
Time 
(MS) 

Tally 

.  Fre 
quency 

Function 
Time 
(MS) 

Tally 

Fre 
quency 

59.0 

1 

1 

65.0 

-t*44-  1~H4-  11 

12 

^Q    9  c; 

f£C      OC 

-4*4_li    -^-4_l_l      1111 

1  4- 

•D  y  .  Z.O 
^O     d 

\JO  .  Z.O 

1^^     ^ 

jrTTtHfc-     JL  JTJL  J.-     JL  JL  JL  JL 

-1-4-4J—  -4-4-U      1 

JLrr 

ov  .  o 
59.75 

oo  .  o 
65.75 

JL  JL  JL  ±  —    JL"I~±-A—     1 
1111 

4 

60.0 

11 

2 

66.0 

-H44.  Ill 

8 

60.25 

66.25 

-H44- 

5 

60.5' 

66.5 

•H44^1 

6 

60.75 

1 

1 

66.75 

1111 

4 

At    n 

1  1 

x:7    fk 

•+4-tJ_  -4-4>iJ      t  1 

1  ? 

O  JL  .  U 

61.25 

1  1 
1 

1 

\j  i  •  \j 
67.25 

JL  JL  Jl  JL-    XT.  JL  JU     J   1 
11 

JL  ^ 

2 

61.5 

-ir«4- 

5 

67.5 

-H44- 

5 

61.75 

111 

3 

67.75 

1111 

4 

62.0 

-H4-3L  11 

7 

68.0 

-Hb44-  111 

8 

62.25 

11 

2 

68.25 

62.5 

-H4-4. 

5 

68.5 

111 

3 

62.75 

Ill 

3 

68.75 

1 

1 

63.0 

-fr-H4~-H44~  1 

11 

69.0 

111 

3 

63.25 

1111 

4 

69.25 

63  .  5 

-W44-  11 

7 

69.5 

11 

2 

63.75 

-1HH4   1 

6 

69.75 

AA.  n 

•JM-4-l_-1H~4~4- 

10 

7O   0 

1  1  1 

3 

V>*x  •  VJ 

64.25 

i.   A   1    JL'—    1    IjfTTC 
111 

JL  \J 

3 

/  \j  .  \j 
70.25 

j  *  i 
1 

1 

f\A.     X 

-4J.J1    M-4-4-J.     1111 

1  4. 

7O    S 

i 

1 

Q~r  *  O 

64.75 

J/TTTC—  I—  .  J.  1  I"T      1  JL  1  JL 

-H4JL. 

JL  Tt 

5 

/  \J  .   vJ 

i 

TABLE  4.3-Frequency  Distribution  for  Data  of  Table  4.1 


Number  of  Actuators 

Function  Time 

Exhibiting  Given  Function 

Relative  Frequency 

(MS) 

Time  =  Frequency  (f) 

(r.  f.) 

59.0 

1 

0.005 

60.0 

2 

0.010 

60.75 

1 

0.005 

61.0 

2 

0.010 

61.25 

1 

0,005 

61.5 

5 

0,025 

61.75 

3 

0.015 

62.0 

7 

0.035 

62  .  25 

2 

0.010 

62.5 

5 

0.02S 

62  .  75 

3 

0.015 

63.0 

11 

0.055 

63.25 

4 

0.020 

63.5 

7 

0.035 

63  .  75 

6 

0  ,  030 

64.0 

10 

0.050 

64.25 

3 

0.015 

64.5 

14 

0.070 

64.75 

5 

0,025 

65.0 

12 

0.060 

65.25 

14 

0.070 

65  .  5 

11 

0.055 

65  .  75 

4 

0.020 

66.0 

8 

0,040 

66.25 

5 

0.025 

66,5 

6 

0.030 

66.75 

4 

0.020 

67,0 

12 

0,060 

67.25 

2 

0.010 

67.5 

5 

0,025 

67  .  75 

4 

0,020 

68.0 

8 

0.040 

68  .  5 

3 

0,015 

68.75 

1 

0.005 

69.0 

3 

0.015 

69.5 

2 

0,010 

70.0 

3 

0.015 

70.25 

1 

0.005 

70.5 

1 

0.005 

Totals 

201 

1.005* 

Total  exceeds  1 .000  because  of  errors  of  rounding* 


C503 


TABLE  4.4-Frequency  Distribution  (Using  Class  Intervals)  for  Data 

of  Table  4.1 


Function  Time 
(MS) 


Number  of  Actuators 
With  Function  Time 

In  Specified  Class 
Interval  =  Frequency  (f) 


Relative  Frequency 
(r.  f.) 


58  <  X  <  59 

1 

0.005 

59  <  X  <  60 

2 

0.010 

60  <  X  <  61 

3 

0.015 

61  <  X  <  62 

16 

0.080 

62  <  X  <  63 

21 

0.104 

63  <  X  <  64 

27 

0.134 

64  <  X  <  65 

34 

0.169 

65  <  X  <  66 

37 

0.184 

66  <  X  <  67 

27 

0.134  , 

67  <  X  <  68 

19 

0.095 

68  <  X  <  69 

7 

0.035 

69  <  X  <  70 

5 

0.025 

70  <  X  <  71 

2 

0.010 

Totals 

201 

1.000 

-4O 


30 


g20 


Ul 

ct 


1O 


-'V 


0.20 


0.15    >: 


Ul 

O.1O    2l 
ui 


0.05 


O.OO 


58  60  62  64  66  68  7O 

TIME  (IN     MILLISECONDS) 

FIG.   4.1— Frequency  histogram   and   polygon   plotted  from   Table  4.4. 


C511 


52  CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 

TABLE  4.5-Cumulative  Frequency  Distribution  Formed  From  Table  4.4 


Function  Time 
(-30 


Number  of  Actuators  With 

Function  Time  Less  Than 

or  Kqual  to  the  Specified 

Value  ==  Cumulative 

Frequency  (c.f.) 


Relative  Cumulative 
Frequency  (r.c.f.) 


58 

0 

0.000 

59 

1 

0.005 

60 

3 

0.015 

61 

6 

0.030 

62 

22 

0.109 

63 

43 

0.214 

64 

70 

0.348 

65 

104 

0.517 

66 

141 

0.701 

67 

168 

0.836 

*  68 

187 

0.930 

69 

194 

0.965 

70 

199 

0  .  990 

71 

201 

1  ,000 

in  Figure  4.2  and  is  quite  helpful  in  interpreting  the  observed  data. 
Note  the  cumulative  (ogive)  curve  is  plotted  by  joining  the  right-hand 
endpoints  at  the  tops  of  the  rectangles,  This  curve  (see  clotted  line)  in 
formed  as  just  mentioned  because  it  represents  the  cumulative  fre 
quency  up  to  and  including  the  upper  clans  limit. 

4.5     CALCULATION   OF  SAMPLE  STATISTICS 

If  a  satnple  is  to  be  described  in  any  reasonable  manner,  it  m  desirable 
to  calculate  certain  representative  values  which  {summarize  a  great  deal 


2OOr~ 


t2O 


u. 

> 


80 


2E    4O 
O 

O 


eo  62  64  66 

TIME   (IN   MHJJSECONOS) 


7O 


o.eoS 
& 

40.403: 
o 


FIG.  4.2— Cumulative  frequency  histogram  and  polygon 
plotted  from  Table  4*5* 


4.6       THE    ARITHMETIC    MEAN  53 

of  information.  Not  all  of  the  representative  values  to  be  described  in 
the  following  pages  are  of  equal  importance.  However,  we  have  gone 
into  considerable  detail  in  defining  them  all  so  that  the  reader  will  be 
aware  of  their  existence,  uses,  advantages,  and  disadvantages. 

4.6      THE  ARITHMETIC    MEAN 

It  is  not  surprising  that  the  ordinary  arithmetic  mean  is  the  most 
common  of  these  representative  values.  The  sample  mean,  denoted  by 
X  ,  is  defined  as  the  arithmetic  average  of  all  the  values  in  the  sample. 
The  formula  for  calculating  the  sample  mean  is 

X  «   (X,  +   -  -   -  +  Xn)/n  =   X  Xt/n  =   Z)  X/n  (4.1) 

i=l 

where  there  are  n  observations  in  the  sample, 

Example  4.1 

Given  the  sample  values  3,   4,    —2,    1,   and  4,    calculate  the   mean. 


The  above  example  illustrates  the  method  of  computing  the  arith 
metic  mean.  It  is  to  be  noted  that  the  arithmetic  mean  is  affected  by 
every  item  in  the  sample  and  is  greatly  affected  by  extreme  values. 
Two  interesting  properties  of  the  arithmetic  mean  are:  (1)  the  sum  of 
the  deviations  from  the  arithmetic  mean  is  zero,  and  (2)  the  sum  of  the 
squares  of  the  deviations  from  the  arithmetic  mean  is  less  than  the  sum 
of  the  squares  of  the  deviations  from  any  other  value. 

As  might  be  expected,  the  arithmetic  mean  has  both  advantages  and 
disadvantages.  Its  advantages  are:  (1)  it  is  the  most  commonly  used 
average,  (2)  it  is  easy  to  compute,  (3)  it  is  easily  understood,  and  (4)  it 
lends  itself  to  algebraic  manipulation.  The  one  major  disadvantage  of 
the  arithmetic  mean  is  that  it  is  unduly  affected  by  extreme  values  and 
may  therefore  be  far  from  representative  of  the  sample. 

Before  proceeding  to  a  second  representative  measure  for  describing 
samples,  it  will  pay  us  to  look  at  methods  of  calculating  the  arithmetic 
mean  when  our  data  are  in  the  form  of  a  frequency  distribution.  If  for 
each  different  value  of  X  we  have  a  frequency/,  then  the  sample  mean 
is  given  by 

2-  =  ' 


/I+/2+     •     •     *    +fd 


n 
a—i 
where  there  are  d  different  values  of  X. 


54  CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 

Example  4.2 

_  The  data  in  Table  4.3  are  of  the  typo  just  described.  Then 
X  =  {(1)(59.0)  +  (2)  (60.0)  +  -  -  -  +  (1)(70.5)  }/20l  =  13,071.75/201 
=  65.034  milliseconds, 

Many  times  oxir  data  appear  in  frequency  tables  where  we  no  longer 
know  the  actual  values  of  the  observations  but  only  to  which  class  in 
terval  they  belong.  In  these  instances,  the  best  we  can  do  is  to  approxi 
mate  the  sample  mean.  To  obtain  this  approximation  we  assume  that 
the  values  in  a  particular  class  interval  are  xmiformly  distributed  over 
the  interval.1  This  permits  us  to  use  the  midpoint  for  each  observation 
in  the  interval  when  calculating  the  mean.  Thus,  if  we  denote  the  mid 
point  of  the  tth  interval  by  £»*,  and  there  are  k  intervals,  the  sample 
mean  is  approximately 


"Vs   r^    ^^l    +*     '     •     •     +  /&£&      ^        .-,  ^      A.^^^  >.      «v 

A  =  -— -—      -  —  —--—-—.-  _  _-,-.._-_-_  _  ^-  ^4.o) 

y  i  -j-  -  *  •  +  jk  n  n 

Example  4.3 

Considering   tho   data   of   Table   4,4,    it  is  seen   that   3fs  { (I)(58.5) 
+  (2)  (59.rO  +    —  -    +  (2)(70.f>)  }/201    -  13,032.5/201   -  «4.«38  milli- 


Short-cut  methods  of  ealexilation  for  use  when  machines  are  not 
available  are  summarized  in  liquations  4.2a  and  4.tta: 


(w)  (4  ,  3a) 

n 

where  JTo  and  fo  are  arbitrary  origins,  #  —  A*"  —  X^  z"=  ({•—  fu)/«»,  and 
t(j  is  thci  width  of  a  class  intorvah 

Example  4.4 

-Y  /  ^  fZ 


to 

3 

_20 

20 

5 

—  10 

30 

8 

0 

40 

2 

10 

la 

_  ?,° 

-90 
In  the  above  table,  A'o8"^.  Therefore 


writorn  anHXtmt^  that  all  the  valuon  in  an  intihrval  are  concentrated  at 
the  midpoint. 


4.8       THE    MEDIAN  55 


Example  4.5 

Class  Interval  f  £  i  fi 


$<X<\.5 

3 

10 

—  2 

—  6 

15<X<25 

5 

20 

—  1 

—  5 

25<X<35 

8 

30 

0 

0 

35<X<45 

2 

40 

1 

2 

18 

—  9 

Thus  ^^30+ (  —  9/18)  (10)  =25. 

4.7  THE   MI  ORANGE 

Another  representative  value  of  importance,  especially  when  a  quick 
average  is  needed,  is  the  midrange.  The  midrange  is  defined  as 

^ -^-min    ~T"    -^-max  ,  ^       ,. 

MR  =  (4.4) 

where  X^^  is  the  smallest  (minimum)  sample  value  and  Xmax  is  the 
largest  (maximum)  sample  value.  It  must  be  realized  that  even  though 
the  midrange  is  quick  and  easy  to  compute,  it  is  often  inefficient  be 
cause  all  information  contained  in  the  intermediate  values  has  been 
ignored.  Also  it  can  be  quite  unrepresentative  if  either  the  smallest  or 
largest  value  is  decidedly  atypical  of  all  the  data. 

4.8  THE   MEDIAN 

A  representative  value  frequently  employed  as  an  aid  in  describing  a 
set  of  data  is  the  median.  The  sample  median,  denoted  by  M,  is  the 
[(n+l)/2]th  observation  when  the  values  are  arrayed  in  order  of  mag 
nitude.  Theoretically,  one-half  the  observations  should  have  a  value 
less  than  the  median  and  one-half  the  observations  should  have  a  value 
greater  than  the  median.  However,  in  practice  it  does  not  always  work 
out  quite  this  way  due  to  clustering  of  the  observations  (see  Example 
4.6).  Regardless,  the  median  is  important  as  a  measure  of  positioner 
location. 

Example  4.6 

If  we  consider  the  data  of  Table  4.3  where  n  =  201,  the  median  is  the 
(201  +  l)/2  =  101st  item  in  the  array.  Counting  down  the  frequencies  in 
Table  4.3  we  find  the  101st  item  to  be  65.  Thus  M  =  65  milliseconds. 

Example  4.7 

Given  the  sample  (2,  3,  4,  6,  6,  7),  the  median  is  the  (6  +  l)/2  =3. 5th 
observation  in  the  array.  To  avoid  ambiguity,  it  is  agreed  that  the 
median  will  be  halfway  between  the  third  and  fourth  observations  in 
the  array.  Thus  Af  =  5. 

When  data  are  grouped  in  class  intervals  as  in  Table  4.4,  the  median 


56  CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 

cannot  be  located  exactly.  However,  if  we  assume  that  the  observations 
in  each  class  interval  are  uniformly  distributed  over  the  interval,  a  close 
approximation  to  the  median  may  be  obtained.  The  first  step  is  to  lo 
cate  the  class  in  which  the  median  belongs :  This  is  done  by  adding  up 
the  class  frequencies  until  we  find  the  class  which  contains  the 
[(w-f-l)/2]th  observation.  Of  course,  if  a  cunrulative  frequency  distribu 
tion  has  been,  formed  as  in  Table  4.5,  the  median  class  is  easily  located. 
Then,  the  sample  median  may  be  approximated  using  the  equation 


M  & 


»+    1 

c 

2 


(4-5) 


where 

LM  =  lower  limit  of  the  median  class, 
n  =  number  of  observations  in  the  sample, 

,JJ=ssum  of  the  frequencies  in  all  classes  preceding  the  median  class, 
fM  ==  frequency  in  the  median  class,  and 
t#  =  width  of  the  median  class. 

Example  4,8 

Considering  the  data  of  Table  4.4,  we  sec  that 

7QV 

64.01  milliseconds. 


/ini  — .  7Q\ 

M  &  64  +  (  — — —  )  (1) 
\        34        / 


It  is  possible  to  approximate  the  median  graphically  from  a  cumula 
tive  frequency  (ogive)  curve  using  the  relative  curmiiative  frequency 
(r.c.f.)  scale.  This  will  be  illustrated  in  Section  4.9. 

The  median,  a  measure  of  position,  is  affected  by  the  number  of  items 
bxit  not  by  the  magnitxide  of  extreme  values.  Two  characteristics  of  the 
median  which  are  of  interest  are:  (1)  the  sum  of  the  absolute  values  of 
the  deviations  from  the  median  is  less  than  the  sum  of  the  absolute 
values  of  the  deviations  from  any  other  point  of  reference,  and  (2)  theo 
retically  the  probability  IB  §  that  an  observation  selected  at  random 
from  a  set  of  data  will  be  IOHH  than  (greater  than)  the  median. 

Some  advantages  and  disadvantages  of  the  median  with  which  one 
should  be  familiar  if  he  wants  to  make  proper  une  of  this  statistic  will 
now  be  mentioned.  The  advantages  are:  (1)  it  in  easy  to  calculate,  and 
(2)  it  is  often  more  typical  of  all  the  observations  than  is  the  arithmetic 
moan  sinco  it  is  not  affected  by  extreme  values.  The  disadvantages*  are: 

(1)  the  items  mtint  be  arrayed  before  the  median  can  be  obtained  and 

(2)  it  does  not  lend  itself  to  algebraic  manipulation* 

4.9      PERCENTILE,  DECILE,  AND  QUARTILE  LIMITS 

In  this  section  we  shall  consider  locating  various  values  which  divide 
i  he  population  or  sample  into  groupn  according  to  the  magnitude  of  the 


4.9       PERCENTILE,    DECILE,    AND    QUARTILE    LIMITS  57 

observations.  The  median  (see  Section  4.8)  was  obviously  one  such 
value  since  it  divided  the  array  into  two  groups,  each  containing  50  per 
cent  of  the  observations.  We  now  wish  to  determine  other  such  values. 
Let  us  consider  the  most  general  case  first.  If  we  want  to  locate  a 
value,  say  P#,  such  that  p  per  cent  (0  <p  <  100)  of  the  observations  are 
less  than  Pp  and  100  —  p  per  cent  of  the  observations  are  greater  than  Pp, 
we  call  Pp  the  upper  limit  of  the  pth  percentile,  and  approximate  P&  by 
the  [p(n+ l)/100]th  observation  in  the  sample  array  if  we  start  counting 
from  the  smallest  value.  For  example,  P$7  is  the  upper  limit  of  the  67th 
percentile  and  is  approximated  by  the  [67(n+l)/100]th  observation  in 
the  sample  array.  Similarly,  P6e  is  the  upper  limit  of  the  66th  percentile 
and  is  approximated  by  the  [66(n+l)/100]th  observation  in  the  sample 
array.  If  we  refer  to  the  67th  percentile,  we  mean  the  interval  from  P6e 
to  P67 — in  general,  the  pth  percentile  is  the  interval  from  Pp~i  to  Pp. 
(NOTE:  Percentile  limits  are  special  cases  of  the  fractiles  introduced 
in  Definition  3.32.) 

Example  4.9 

If  we  consider  the  data  of  Table  4.3,  what  is  the  upper  limit  of  the  80th 
percentile?  P8o  is  approximately  the  80(201  +  1)/100  =  161.6th  observa 
tion  in  the  array  which  is  67  milliseconds. 

Example  4.10 

What  is  the  upper  limit  of  the  35th  percentile  in  the  sample  given  in 
Example  4.7?  P3S  is  approximated  by  the  35(6  +  l)/100  =  2.45th  obser 
vation  in  the  array.  To  avoid  ambiguity,  we  agree  to  set  P$$  forty-five 
one  hundredths  of  the  way  from  the  second  to  the  third  observation 
in  the  array  when  we  count  from  the  smallest  value.  Thus  P3s  is  0.45  of 
the  way  between  3  and  4,  that  is,  jP35  =  3.45. 

The  reason  for  the  word  percentile  should  now  be  clear :  if  we  locate 
Pi>  Pz,  -  •  -  ,  PQQ,  we  have  (theoretically)  split  our  array  into  100  parts 
(percentiles) ,  each  containing  1  per  cent  of  the  observations. 

The  meaning  of  such  terms  as  decile  limits  and  quartile  limits  (see 
the  heading  of  this  section)  is  now  almost  obvious.  The  decile  limits 
DI,  D2;  -  •  -  ;  Z>9  theoretically  split  our  array  into  ten  parts  (deciles), 
each  containing  10  per  cent  of  the  observations.  The  quartile  limits 
Qij  Qa,  and  <2a  theoretically  divide  our  array  into  four  parts  (quartiies), 
each  containing  25  per  cent  of  the  observations.  No  particular  methods 
of  calculation  will  be  presented  for  decile  or  quartile  limits  since  they 
are  only  special  cases  of  percentile  limits.  This  is  clear  once  we  observe 
that 

Pio  *  Di  P4o  =  Dt  P75  =  Q3 

P2o  =   Dz  Pfeo  ==   £>6  =  Qa  =  M  Pso  =*   £>s 

P2s  =  Qi  Peo  =  £>e  Poo  =   Dg 

P30    =     £>3  ^70    —     #7 


58 


CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 


In  Section  4.8  we  mentioned  the  possibility  of  estimating  the  median 
from,  the  graph  of  the  relative  cumulative  frequency  distribution.  To 
illustrate  this  technique  we  shall  undertake  the  location  of  percentile 
limits  in  general.  Consider  the  cumulative  frequency  curve  of  Figure 
4,2  which  we  reproduce  here  as  Figure  4.3.  The  procedure  is  as  follows: 
Pp  being  the  upper  limit  of  the  pth  percentile  which  says  that  p  per  cent 
of  the  observations  are  less  than  or  eqxial  to  Pp,  all  we  have  to  do  is 
locate  p/100  on  the  relative  cumulative  frequency  scale,  draw  a  hori 
zontal  luie  from  this  point  to  the  ogive  curve,  and  from  here  drop  a 


vertical  lino  down  to  the  horizontal  axis, 
illustrated  in  Figure  4.3  for  P$i  and  P&$. 


200 


thus  locating  Pp.  This  is 


UJ 

9 

^  12O 

u. 

UJ 

>    80 


4O 


/v-bc 


QC 


6O  62  64  66  68  7O 

TIME    (IN   MILLISECONDS) 

FIG-  4*3— Cumulative  frequency  (ogive)  "curve*'   plotted  from  Table  4.5, 

4.1O     THE   MODE 

Another  valxio  of  aid  in  describing  a  sample  is  the  mode*  The  mode,  is 
defined  as  the  value  which  occurs  most  frequently  in  the  sample-  The 
mode  of  the  sample  will  he  denoted  by  MO,  It  should  ho  obvious  that, 
the  mode  will  not  always  be  a  central  value;  in  fact,  it  may  often  he  un 
extreme  value.  Then  too,  a  sample  may  have  more  than  one  mode. 
We  should,  at  this  time,  distinguish  i>et\veen  an  ahtwlute  made  und  a 
relative  mode.  An  absolute  mode  in  what  wo  defined  above  (there  may,  of 
coxmto,  be  more  than  one  absolute  mode) ;  a  relative  mode  is  a  value 
which  occurs  more  frequently  than  neighboring  values  even  if  it  is  not 
an  absolute  mode. 


Example  4.11 

<Hvon  a  sample  eotmiHting  of  the  values  ft,  7, 
nay  thore  IK  no  nuKlc  or  tluvrc  an*  fiw  modern 
only  once, 


1,  4,  and  *i»  we  may 
e  otich  vahu*  oecurn 


4.1 0       THE    MODE 


59 


Example  4.12 

Considering  the  data  of  Table  4.3,  we  see  there  are  two  absolute 
modes,  64.5  and  65.25  milliseconds,  since  each  of  these  values  occurs  14 
times  and  no  other  value  occurs  that  frequently. 

Example  4.13 

Given  a  frequency  histogram  like  that  shown  in  Figure  4.4,  we  would 
say  there  are  two  relative  modes:  one  in  Class  A  and  one  in  Class  J5. 
However,  the  mode  in  Class  A  is  the  only  absolute  mode. 


A  B 

FIG,  4.4— Example  of  a  bimodal  frequency  histogram. 

If  our  data  are  grouped  in  class  intervals,  it  will  be  impossible  to 
locate  the  mode  exactly.  Under  such  circumstances,  the  best  we  can  do 
is  to  approximate  the  value  of  the  mode.  As  was  the  case  when  approxi 
mating  the  median,  the  first  step  is  to  locate  the  modal  class.  This  is 
accomplished  quickly  by  picking  out  the  class  interval  which  shows  the 
highest  frequency.  The  sample  mode  is  then  approximated  by 


MO  ^  ZMO  + 


(4-6) 


where 


&MO  =  lower  limit  of  the  modal  class, 

di  =  the  difference  (sign  neglected)  between  the  frequency  of  the 

modal  class  and  the  frequency  of  the  preceding  class, 
^2  =  the  difference  (sign  neglected)  between  the  frequency  of  the 

modal  class  and  the  frequency  of  the  following  class,  and 
w=*  width  of  the  modal  class. 

Example  4.14 

Consider  the  sample  given  in  Table  4.4.  The  modal  class  is  from  65  to 
66  milliseconds  with  a  frequency  of  37.  Therefore, 


MO  =  65  +  C ")  (1)  =  65.23  milliseconds. 

\o  "~r~  It) / 


60  CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 

Since  the  mode  is,  by  definition,  the  most  typical  value,  it  is  often 
considered  the  most  descriptive  of  the  representative  values  discussed 
so  far.  However,  its  importance  diminishes  as  the  number  of  observa 
tions  becomes  limited. 

4-11      THE    RANGE 

All  the  representative  values  discussed  in  the  preceding  sections  have 
been  some  sort  of  average  or  measure  of  position.  It  must  be  clear, 
though,  that  they  are  not  sufficient  by  themselves  to  describe  most 
populations  or  samples  adequately.  This  statement  may  be  verified 
easily  if  we  consider  two  sets  of  data  which  have  the  same  mean,  the 
same  median,  and  the  same  mode  but  which  differ  greatly  in  the  amount 
of  variation  present  in  each  set  of  data.  It  would  seem  then  that  some 
measure  of  the  variation,  or  dispersion,  among  the  individual  values  is 
also  needed.  Several  such  measures  have  been  devised,  and  we  shall 
mention  four  of  these  in  this  and  succeeding  sections. 

A  measure  of  dispersion,  to  be  suitable,  should  be  large  when  the 
values  vary  over  a  wide  range  (and  there  are  quite  a  few  extreme 
values)  and  should  be  small  when  the  range  of  variation  is  not  too  great. 

The  simplest  measxire  of  variation  is  one  that  has  been  mentioned 
before,  that  is,  the  range.  If  we  denote  the  smallest  (minimum)  sample 
value  by  Xmin,  and  the  largest  (maximum)  sample  value  by  -Xn>ft*,  the 
sample  range  is  given  by 


The  sample  range,  though  easy  to  obtain,  is  often  termed  inefficient 
because  it  ignores  all  the  information  available  from  the  intermediate 
sample  values*  However,  for  small  samples  (n<  10),  the  efficiency  (rela 
tive  to  other  measures  of  variation  yet  to  be  defined)  is  quite  high.  For 
a  more  explicit  discussion  of  the  efficiency  of  the  range  relative  to  the 
standard  deviation  the  reader  is  referred  to  Section  4.12.  Thus  we  find 
the  sample  range  enjoying  a  favorable  reception  and  wide  use,  because 
of  ease  in  computation,  in  such  applications  as  statistical  quality  con 
trol  where  small  samples  arc  the  rule  rather  than  the  exception, 

Example  4.15 

For  the  sample  given  in  Example  4.7,  we  obtain  .R=*7-—  2  =  5. 

4-12     THE  STANDARD   DEVIATION   AND   VARIANCE 

Perhaps  the  best  known  and  most  widely  used  measure  of  variability 
is  the  standard  deviation.  Of  almost  equal  importance  is  the  square  of 
the  standard  deviation,  this  quantity  being  known  as  the  variance.  We 
shall  explain,  both  of  these  measures  by  defining  the  variance* 

The  sample  variance,  denoted  by  s2,  is  defined  by 


4.12       THE    STANDARD    DEVIATION    AND    VARIANCE  61 

-  Xy  +  •  •  -  +  (Xn  -  Xy}/(n  -  1)  (4  g) 

Sometimes  it  is  convenient  to  let  x^  =  Xi  —  X,  that  is,  it  is  simpler  to 
denote  deviations  about  the  mean  by  lower-case  letters.  Then, 


However,  the  best  form  for  machine  calculation  purposes  is 

n^      2  /    n         \  2 

L,  -Xt    —    C    2.4  x* 

-1  \    t—1  S 


n 


n  —  1 

The  sample  standard  deviation  is  then  defined  as  the  positive  square 
root  of  the  variance,  namely, 

s  =  v^5"".  (4.11) 

The  use  of  n  —  1  (instead  of  n)  when  defining  the  sample  variance 
may  seem  peculiar  to  the  reader,  since  we  implicitly  used  a  divisor  of  N 
when  defining  the  population  variance.  Our  reason  for  using  n  —  1  is 
this:  In  general,  one  prefers  unbiased  estimators2  to  biased  estimators, 
and  the  use  of  n—  1  gives  us  an  unbiased  estimator  of  <r2.  If  n  were  used, 
the  resulting  function  of  the  sample  observations  would  produce  biased 
estimates  of  the  unknown  population  variance  —  biased  because,  on  the 
average,  the  estimates  would  be  too  small.  Thus  the  student  of  statistics 
must  resign  himself  to  remembering  that,  while  the  population  variance 
is  defined  using  a  divisor  of  N,  the  sample  variance  requires  a  divisor 
of  n—  1.  Incidentally,  we  refer  to  n  —  1  as  the  degrees  of  freedom  associ 
ated  with  the  sample  variance  (and  standard  deviation)  , 

Example  4.16 

For  the  sample  (13,  5,  8,  5)  we  see  that 
$*  -  {(13  -  7.75)a  +  (5  -  7.75)2  +  (8  -  7.75)*  +  (5  -  7.75)*}  /3 
-  {(S.25)2  +  2(~  2,75)2  +  (0,25)2}/3  -  14.25     and 


thus  s  =  VI 4.25  =  3.775. 

If  we  had  used  the  formula  recommended  for  machine  calculation, 
the  same  value  of  s2  would  have  been  obtained: 

a  An  estimator  is  a  statistic,  that  is,  a  function  of  the  sample  values,  which  will 
provide  us  with  numerical  estimates  of  a  parameter. 


62  CHAPTER    4,    SAMPLING    ANO    DESCRIPTIVE    STATISTICS 

s*  =*  {  132  +  52  +  32  +  52  _  (13  +  5  +  3  +  5)2/4}  /3 
^  283  -  (31)  8/4  ^  283  —  240.25  =  42.75  = 

If  the  sample  data  appear  in  a  frequency  distribution,  the  following 
forms  are  appropriate  for  calculation.  When  no  class  intervals  are  in 
volved  (as  in  Table  4.3), 


n  —  1 
or 


n  —  1 
where  -ST0  is  an  arbitrary  origin,  n—  y^/,  and  %***X-~XQ* 

Example  4.17 

ConBidor  "Pablo,  4.6.  Utfirig  Kqxiation  4.12,  we  obtain 
$*  =«  {12,700  -  (45C))2/1&}/17  -  1450/17  -  85.3. 
Utnng  Kqxiation  4.12a, 
3*  **  {1900  -  (-  90)«/18}/17  -  85,3. 

TABLE  4.6-tll  ust  ration  of  the  Use  of  Equations  4.12  and  4J2a 


,      . 

\  ^"  •  •"•  *"*  <&tj 


-Y 

/ 

.AY 

/-Y* 

Z 

& 

f& 

10 

3 

30 

300 

—  20 

—  60 

1  ,  200 

20 

5 

100 

2,000 

—  10 

—  50 

500 

30 

8 

240 

7,200 

0 

0 

0 

40 

2 

80 

3,200 

10 

20 

200 

Totals  I  H  450  1  2  ,  700  .  .  —  <>0  1  ,  <)()0 

When  claMH  intervals  are  involvc^l,  tho  appropriut<i  formuta^  arc: 

,.«?^-<2:/»v. 

n  —  1 
and 


i 

In  which  w  ivS  tho  width  of  a  claws  interval  and  £«*  (£  —  f0)/t^  where  fn  in 
an  arbitrary  origin* 


4.12       THE    STANDARD    DEVIATION    AND    VARIANCE 

TABLE  4.7-Illustration  of  the  Use  of  Equations  4.13  and  4.13a 


Totals 


18 


450        12,700 


—  9 


63 


Class 

Interval 

/ 

* 

/* 

f? 

i 

fi 

/* 

5 

<  X  < 

:  15 

3 

10 

30 

300 

—  2 

—  6 

12 

15 

<  X  < 

:  25  

5 

20 

100 

?, 

,000 

—  1 

—  5 

5 

?.S 

<  X  < 

:  35  

8 

30 

240 

7 

,200 

0 

0 

0 

3.S 

<  X  < 

:  45  .   .    . 

2 

40 

80 

^ 

,200 

1 

2 

2 

19 


Example  4.18 

Consider  Table  4.7.  Setting  £0 
be  verified  by  evaluating 

s*  =  {12,700  —  (450)2/18}/17 


,  it  is  seen  that  $2  =  85.3.  This  may 


or 
sz  = 


[{19-  (- 


It  was  mentioned  in  Section  4.11  that,  for  small  n,  the  range  is 
reasonably  efficient  relative  to  the  standard  deviation.  By  this  state 
ment  was  meant  that  if  one  wishes  to  estimate  <r,  it  can  be  done  using 
either  R  or  s.  When  sampling  from  a  normal  population,  the  efficiency 
of  the  sample  range  relative  to  the  sample  standard  deviation  as  an 
estimator  for  the  population  standard  deviation  is  given  in  Table  4.8. 
As  an  example  of  the  use  of  this  table,  if  a  person  desires  to  use  R 

TABLE  4.8-Efficiency  of  Range  (R)  Relative  to  Standard  Deviation  O)  as 
an  Estimator  of  cr  for  a  Normal  Population 


Sample  Size  (n) 

Relative     Efficiency 

<r/JS(#) 

2  

1.000 

0.886 

3  

0.992 

0.591 

4  

0.975 

0.486 

5  

0.955 

0.430 

6    

0.933 

0.395 

7  

0.912 

0.370 

8  

0.890 

0.351 

9  

0.869 

0.337 

10.           

0.850 

0.325 

12  

0.815 

0.307 

14  

0.783 

0.294 

16           

0.753 

0.283 

18  

0.726 

0.275 

20  

0.700 

0.268 

30  

0.604 

0.245 

40             

0.536 

0.231 

50  

0.490 

0.222 

64 


CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 


rather  than  s,  he  would  estimate  a-  by  calculating 

#  =  (R)  { Value  of  a/E(K)  for  given  n] . 


(4.14) 


4.13      THE  COEFFICIENT  OF  VARIATION 

The  coefficient  of  variation  has  been  explained  by  statisticians  in 
different  ways.  However,  attention  usually  is  called  to  the  rather  obvi 
ous  fact  that  things  with  large  values  tend  to  vary  widely,  while  things 
with  small  values  exhibit  small — numerically  small,  that  is— variation. 
Thus,  to  afford  a  valid  comparison  of  the  variation  among  large  values 
and  t'he  variation  among  small  values,  such  as  the  variation  among 
salaries  of  industrial  executives  and  the  variation  among  the  wages  of 
day  laborers,  the  variation  is  expressed  as  a  fraction  of  the  mean,  and 
frequently  as  a  percentage.  This  measure  of  relative  variation  is  called 
the  coefficient  of  variation  and  is  defined  as 


CV  -  s/  T  . 
In  percentage  form,  this  becomes 

100  CV  «  100(V  X)  per  cent. 


(4.15) 


(4.16) 


TABLE  4.9-Special    Form  for  Calculating  and    Presenting   Sample    Statistics 


n 

£* 

x 

£*» 

—  ' 

(T.xy/n 

-  —  • 

,^          ,u^     „ 

.... 

I>9 

*2 

$ 

—  - 

•A  max 

•Am  In 

-  —  „ 

—  - 

R 

SPECIAL  NOTES 

FORMULAE 

s*  -  2>V(»  - 

R  BK  A"™**  — 


largest  observation—  smallest  observation. 


4.14       SUMMARY  65 

The  coefficient  of  variation  is,  of  course,  an  ideal  device  for  comparing 
the  variation  in  two  series  of  data  which  are  measured  in  two  different 
units;  e.g.,  a  comparison  of  variation  in  height  with  variation  in  weight. 

Example  4.19 

For  the  sample  given  in  Example  4.16  we  see  that  CV  =  3. 775/7. 775 
=  0.4871,  and  in  percentage  form  100  CV  =  48.71  per  cent. 

4.14      SUMMARY 

The  greater  part  of  this  chapter  has  been  devoted  to  outlining  meth 
ods  of  calculation  for  various  statistics;  i.e.,  functions  of  sample  values, 
which  are  useful  in  statistical  inference.  Not  all  of  the  statistics  dis 
cussed  are  used  in  everyday  applications.  However,  a  select  few  are  used 
so  often  that  it  is  convenient  to  have  a  standard  form  for  calculation 
and  presentation  of  results.  One  such  form  is  presented  in  Table  4.9. 


66  CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 


Problems 

4.1        Plot  a  frequency  histogram  and  polygon  for  the  following  data.  Make 
approximate  eye-estimates  of  the  arithmetic  mean,  median,  and  mode. 

WEEKLY  WAGES  OF  188  FEMALE  EMPLOYEES  OF 
A  SHOE  MANUFACTURING  COMPANY 


$20  .  15 

$25.00 

$40.39 

$25.49 

$25  .  70 

24.15 

22.54 

23.80 

29.60 

18.74 

25.62 

23.89 

28.37 

26.00 

16.70 

26.00 

27.82 

24.80 

26.52 

28.09 

27.84 

25.80 

25.88 

25.04 

24.98 

22.97 

23.20 

23.24 

29.00 

24.55 

25.48 

20.88 

21.70 

25.76 

26.20 

28.00 

28.92 

27.92 

25.80 

22.45 

28.24 

25.70 

22.75 

21.40 

27.10 

31.37 

26.77 

26.00 

18.64 

27.39 

24.53 

24.25 

28.28 

30.32 

23.00 

28.13 

26.23 

21.55 

28.04 

25  .  58 

22.78 

26.88 

26.64 

22.83 

23.45 

25  .  20 

29  .  29 

25.62 

23.40 

26.12 

27.08 

24.40 

25.49 

30.48 

27.03 

26.11 

21.80 

20.85 

26.79 

26.25 

22.04 

22.54 

21.85 

25.65 

27.50 

29.48 

25.20 

26.00 

22.69 

25  .  78 

21.77 

24.32 

26.00 

22.52 

17.50 

26.52 

20.48 

22.92 

23,96 

26.00 

22.00 

22.44 

26.00 

26.35 

25.64 

22.48 

27.25 

24.19 

23.75 

28.94 

21.85 

22.99 

22.33 

24.18 

25.65 

23.12 

22.71 

26.48 

23.23 

23.44 

31.00 

25.38 

25.83 

18.60 

33  .  80 

30.61 

22.00 

29.72 

23.28 

25  .  65 

23  .  80 

26.90 

24.55 

23.12 

29.24 

26.00 

22.68 

24.04 

32.60 

22.15 

25.15 

22.53 

25.12 

23.72 

22.99 

25.70 

27.98 

26.34 

23.08 

24.24 

28.00 

27,14 

23.13 

26.38 

24.00 

26.03 

31.60 

24.79 

24.73 

27.48 

30.23 

22.47 

34.99 

22.09 

19.30 

24.55 

26.67 

24.08 

25  .  78 

23.42 

30.60 

28.32 

22.28 

24.73 

25.65 

29.15 

27.74 

23.69 

28.83 

25.64 

22.48 

25.20 

23.84 

25.68 

28.24 

30.72 

28.92 

25.73 

PROBLEMS  67 


4.2        Plot  a  frequency  histogram  and  polygon  for  the  data  given  below. 
PER  CENT  SILICON  IN  236  SUCCESSIVE  CASTS  OF  PIG  IRON 


1.13 

1.00 

0.96 

0.67 

0.77 

0.65 

0.83 

0.92 

0.80 

0.94 

0.96 

0.76 

0.34 

0.60 

0.79 

0.73 

0.85 

0.62 

0.60 

0.66 

0.84 

1.00 

0.99 

0.96 

0.60 

0.32 

0.87 

0.89 

0.70 

0.91 

1.20 

1.00 

0.97 

1.00 

1.08 

0.85 

0.71 

0.72 

0.74 

0.96 

0.92 

1.00 

0.67 

0.77 

0.74 

1.32 

0.85 

0.94 

0.94 

0.89 

0.98 

0.87 

0.97 

0.94 

0.60 

0.72 

0.72 

0.65 

0.88 

1.00 

1.09 

0.60 

0.72 

0.88 

1.17 

1.00 

0.75 

0.73 

0.91 

1.11 

1.45 

1.45 

0.87 

0.64 

0.60 

1.00 

0.81 

1.14 

0.68 

0.74 

0.36 

0.85 

1.17 

1.00 

0.82 

0.77 

0.67 

0.70 

0.68 

0.89 

0.93 

1.13 

1.00 

0.80 

1.00 

0.86 

0.73 

0.66 

0.79 

0.51 

0.60 

0.89 

1.00 

1.18 

0.82 

0.60 

0.76 

1.07 

0.84 

0.93 

0.73 

0.60 

0.79 

0.61 

1.14 

1.33 

1.00 

0.80 

0.71 

0.95 

0.87 

0.83 

0.65 

0.64 

0.85 

0.78 

0.86 

0.60 

0.92 

0.87 

1.00 

0.91 

0.72 

0.79 

0.70 

1.00 

0.81 

0.80 

0.81 

0.87 

0.60 

0,86 

0.94 

1.00 

0.97 

0.70 

0.37 

1.00 

1.00 

0.99 

0.84 

0.72 

0.48 

1.50 

1.50 

1.00 

0.99 

0.80 

0.85 

0.84 

1.00 

0.91 

0.60 

0.68 

0.75 

0.47 

0.73 

0.97 

0.92 

0.60 

0.82 

1.14 

0.87 

0.70 

O.80 

0.95 

0.61 

1.02 

1.45 

0.93 

0.57 

0.60 

0.61 

0.69 

0.81 

1.00 

1.25 

0.90 

0.60 

0.82 

0.84 

0.92 

0.71 

0.94 

0.87 

0.84 

0.94 

0.97 

0.90 

0.99 

0.97 

1.06 

1.10 

0.89 

0.69 

0.86 

0.61 

0.38 

0.89 

0.97 

0.87 

0.71 

0.33 

0.80 

0.64 

0.26 

1.16 

1.25 

0.66 

0.56 

1.12 

0.73 

0.62 

0.78 

0.68 

0.61 

1.00 

1.11 

1.00 

0.81 

0.70 

0.85 

1.00 

1.50 

1.18 

0.94 

68  CHAPTER    4,    SAMPLING    AND    DESCRIPTIVE    STATISTICS 

4.3        A  random  sample  of  201   women  students   was  obtained  and  their 
heights  and  weights  were  recorded  as  follows: 

HEIGHTS  AND  WEIGHTS  or  201  WOMEN  STUDENTS  AGED  18 
UNIVERSITY  OF  BRITISH  COLUMBIA,  1944-45* 


5-4 

139 

5-5 

1414 

5-44 

99 

5-2 

1164 

5-1J 

1184 

5-3A 

5-4 

158 

5-4 

123 

5-84 

151 

5-4 

1104 

VJ    '-'  jj 

108 

S-Tk 

146& 

5-44 

152| 

5-64 

118 

5-64 

123 

5-54 

5-31 
5-2| 

1274 

1414 
127 

v   *  £ 

5-64 
5-7 
5-32 

1174 
1394 
122 

5-94 

5-5 

5-5 

142| 
119 
1351 

5-3 
5-94 
5-14 

1194 

1344 
114 

5-44 
5-54 
5-3  1 

115 
117 
1124 

«M*   •**  4^ 

5-51 

175 

4-11 

1101 

5-5 

1301 

5-71 

1484 

5-54 

149 

5-84 

148 

124 

5-2  1 

128 

5-3 

125| 

5-6 

128 

5-3 

130 

5-1 

1074 

5-1J 

1044 

5-41 

1124 

5-24 

1024 

5-5 

129 

5-14 

104 

5-1* 

107J 

5-24 

1234 

5-54 

1184 

5-3 

1104 

5-44 

1341 

5-61 

1184 

5-6 

142 

5-8 

152 

5-3 
5-5 
5-8 

1324 
1254 

5-54 
5-54 
5-14 

1371 
1084 
1261 

5-2^ 
5-7 
5-34 

120 
143 
126| 

5-2 
5-54 
5-54 

964 
122 
1494 

5-2 
5-2 
5-41 

114 
1441 
1334 

5-34 

126 

5-7 

1301 

5-2 

1101 

5-9 

134 

5-7 

152| 

v/  v  * 
5-5 

140 

5-84 

1304 

5-4J 

115 

5-74 

1214 

5-74 

1304 

5-5 

131 

5-54 

125J 

5-5 

111 

5-6 

128 

5-81 

151 

5-64 

5-7 

1224 
162 

5-21 

5-4 

110 
117 

5-44 
5-61 

128| 
1434 

5-44 

5-5 

1274 
129§ 

5-54 
5-64 

125 

1404 

5-3 

1324 

5-3J 

1164 

5-10 

1514 

5-6 

117 

5-7 

1394 

5-41 
5-4* 

112* 
125 

5-2  1 
5-44 

974 
126 

5-10 
5-54 

1104 
125J 

5-104 
5-7 

1404 
1381 

5-34 
5-74 

108 
1574 

•*»  *  *f 
5-21 

1754 

5-0 

126 

5-74 

1414 

5-54 

1274 

5-6 

1184 

5-3 

105| 

5-54 

121§ 

5-64 

1644 

5-3  4 

112 

5-44 

1464 

5-6 

115* 

5-4 

1154 

5-71 

122 

5-4 

110 

5-3 

108 

5-6 

<~  —  ^  jt 
173 

5-10 

1334 

5-8 

128 

5-44 

1284 

5-54 

120 

5-8 

1354 

5-5  1 

1494 

5-5  1 

123 

5-41 

1314 

5-24 

113 

5-5 

1261 

5-9 

1384 

5-2 

121  \ 

5-71 

1314 

5-3 

128 

5-6 

1154 

5-6  i 

139 

5-4 

1244 

5-44 

1304 

5-54 

1204 

5-44 

1344 

5-3  1 

142 

5-34 

112 

5-64 

1254 

5-54 

125 

5-52 

126 

5-44 

118 

5-8 

152J 

5-8 

1304 

5-34 

136 

5-3 

1494 

5-t()4 

154 

5-S4 

127 

5-34 

1254 

5-3J 

1114 

5-1 

1474 

5-0 

130 

5-2 

134 

5-54 

130 

5-71 

1174 

5-74 

1284 

5-31 

118 

5-54 

1114 

5-4 

122| 

5-9 

15l| 

5-14 

5-51 

1214 
1354 

5-64 

5-7 

1381 
145 

5-7 
5-7 

122 
138| 

5-34 
5-41 

117 
1372 

5-61 
5-6  1 

134 
120J 

**  * 
5-3* 

•* 
135 

5-44 

na| 

5-3 

1134 

5-54 

118| 

5-64 

1414 

,* 

5-64 

140 

5-7 

1134 

5-4 

1264 

S-54 

1554 

5-44 

1214 

* 
5-8 

147J 

5-54 

129 

5-24 

135 

5-8 

125 

5-7 

188 

5-14 

121J 

5-74 

151 

5-5 

122 

5-5J 

1294 

5-44 

1254 

s-f 

11  4  i 

*  hi*  *,*»««*..* 

^  ^xm-******-, 

Source:  U.B.C,  Students*  Health  Service, 

Plot  a  frequency  histogram  and  polygon  for  (a)  the  heights  and 
the  weights. 


PROBLEMS  69 

4.4  Plot   a   cumulative   frequency   curve   for   the   data  of   Problem   4.1. 
Estimate  the  median  from  this  curve. 

4.5  Plot  a  cumulative  frequency  curve  for  the  data  of  Problem  4.2.  Esti 
mate  the  median  from  this  curve. 

4.6  Plot  a  cumulative  frequency  curve  for  (a)  the  heights  and  (b)  the 
weights  given  in  Problem  4.3.  From  these  curves,  estimate  the  median 
height  and  median  weight. 

4.7  Given  the  samples  listed  below,  calculate  for  each  the  mean,  median, 
mode,  midrange,  range,  variance,  standard  deviation,  and  coefficient 
of  variation: 

(a)  5,  19,  -3,  7,  1,  1 
(i)  5,  -3,  2,  0,  8,  6 
(c)  6,  9,  5,  3,  6,  7 
(<Z)  1,3,2,  -1,5 
(«)  10,  15,  14,  15,  16 
(/)  0,  5,  10,  -3 
(£)  8,  7,  15,  -2,  0 

4.8  Suppose  that  F  =  100  and  s2  =  15.  What  would  the  values  of  7  and  s2 
become  if  each  original  observation  were  (a)  increased  by  10  units, 
(b)  multiplied  by  10  units? 

4.9  Given  the  observations: 


2 

10 

3 

10 

4 

20 

5 

50 

6 

30 

Calculate  the  same  statistics  as  asked  for  in  Problem  4.7. 

4.10  Calculate  the  same  statistics  as  asked  for  in  Problem  4.7  for  each 
of  the  following  sets  of  data: 

(a)  Problem  4.1 
(6)  Problem  4.2 
(c)  Problem  4.3. 

4.11  A  bag  of  potatoes  was  sampled  for  quality,  five  potatoes  being  selected 
at  random  from  the  bag.  Among  the  observations  recorded  were  the 
weights  of  the  potatoes:   17,   15,   10,   12,   and   11   ounces.   Calculate 
7,  s2,  s,  and  JR.  What  property  (using  the  word  rather  loosely)  is 
common  to  the  sample  range  and  the  sample  variance  (or  standard 
deviation)? 

4.12  Given  that  n~25,  222/2===600,  and  Y  =  204,   calculate  the  variance, 
standard  deviation,  and  coefficient  of  variation. 


C  H  APTE  R    5 

SAMPLING  DISTRIBUTIONS 

CERTAIN  SAMPLING  msTRtBUTioNS  pertinent  to  methods  to  be  pre 
sented  in  later  chapters  will  be  discussed  in  this  chapter.  The  law  of 
large  numbers,  TchebychefPs  inequality,  and  the  central  limit  theorem 
will  be  given.  Various  approximations  to  exact  sampling  distributions 
will  also  be  considered. 

5.1      SAMPLE   MOMENTS 

In  the  preceding  chapter,  the  calculation  of  several  different  sample 
statistics  was  outlined.  Of  particular  importance  are  those  statistics 
known  as  sample  moments.  They  are  defined  by 


(5.1) 
and 

n 

~  In  (5.2) 


where  7c  =  0,  1,  2,  •  •  -  .  In  particular,  it  is  noted  that,  wj  —  -ST.  The 
reader  will  see  the  similarity  of  the  above  definitions  to  the  definitions 
of  /4  <md  M*  given  in  Definitions  3,23  ami  3.25,  It  may  be  shown  that 
J£(m£)  «=  /4  for  all  k  but  K(mk)  docs  not  equal  pM  except  for  A*«=0,  1.  For 
this  reason,  mfk  is  known  as  an  unbiased  estimator  of  MiC^  —  O,  1,  •  -  •  ) 
while  m^  is  a  biased  estimator  of  /*A(fc=ss2,  3,  *  •  -  )•  In<ucleutally,  this 
last  remark  is  just  a  restate™ on t  of  the  reason  given  in  the  preceding 
chapter  for  using  $*  rather  than  w2  as  an  estimator  of  cr2. 


5.2     VARIANCE  OF  THE  SAMPLE  MEAN 

It  has  jxmt  boon  stated  that  K(^}—y..  That  in,  in  nampling  from  a 
specified  population,  the  expected  value  (or  average)  of  all  poasiblo 
Bample  meaiiB  is  the  popxtlation  mean.  However,  we  realise  that  it  ia 
equally  important  to  know  nomothing  about  the  variation  among  all 
possible  values  of  the  nample  mean.  To  investigate  thin  variation,  con 
sider  the  variance  of  the  sample  mean, 

n 

After  Borne  algebraic  manipulation,  it  in  wen  that 

2 
<rx  «   J 

1703 


£<*• -*>}>• 


5.3      TCHEBYCHEFF'S    INEQUALITY  71 


~  M)2  +  n(n  - 
{a-*2  +  (n  -  !)£[(*<  -  M)(*y  -  /*)]}/».  (5.4) 


Two  cases  must  now  be  distinguished:  (1)  random  sampling  and  (2) 
sampling  without  replacement  from  a  finite  population.  For  these  two 
cases,  we  obtain,  respectively, 

and 

^2     AT   *j 

(5.6) 


n    N  -  1 

A7"  being  the  size  of  the  population. 

The  preceding  result  is  very  important.  It  says  that,  no  matter  what 
the  population  (as  long  as  it  has  a  finite  variance),  the  distribution  of 
the  sample  mean  becomes  more  and  more  concentrated  in  the  neighbor 
hood  of  the  population  mean  as  the  sample  size  is  increased.  That  is, 
the  larger  the  sample  size,  the  more  certain  we  become  that  the  sample 
mean  will  be  close  to  the  (unknown)  population  mean.  This  result  will 
be  expressed  more  precisely  in  the  following  section. 

5.3     TCHEBYCHEFF'S    INEQUALITY 

A  useful  inequality  is  that  due  to  Tchebycheff,  namely 

P(  |   X  -  fJL  |    >   ka)    <    1/k*.  (5 .  7) 

This  inequality  is  often  expressed  in  the  following  alternative  form: 

P(\  X  —  »\   <  ka)  >  1  —  1/k2  (5 . 8) 

or 

P(fji  -  k<r  <  X  <  VL  +  k<r)   >   1   -   I/*2.  (5.9) 

TchebychefPs  inequality  shows  how  <r  may  be  used  as  a  measure  of 
variation.  It  can  be  applied  in  a  wide  variety  of  cases  for  it  assumes  only 
the  existence  of  M  and  <r2.  That  is,  no  assumption  is  made  concerning 
the  form  of  the  population  but  only  that  the  mean  and  variance  exist. 
If  we  restrict  our  attention  to  unimodal  distributions,  the  inequality 
may  be  sharpened.  Under  such  a  restriction,  we  obtain 

P(\  X  —  MO  |  >  kB)  <  4/9k*  (5.10) 

where  MO  is  the  mode  and  J52  =  cr2  +  (MO  —  ^)2.  An  alternative  form  is 


72  CHAPTER    5,    SAMPLING    DISTRIBUTIONS 

where  A  =  (/*  —  MO)/cr.  It  should  be  noted  that  if  the  distribution  is 
not  only  unimodal  but  also  symmetric,  that  is,  M  —  MO,  then  Equations 
(5.10)  and  (5.11)  reduce  to 


P(  |  X  —  AC  |  >  *<r)  <  4/9k*.  (5.  12) 

5.4      LAW  OF    LARGE  NUMBERS 

It  is  now  possible  to  give  precise  formulation  to  the  law  of  large 
numbers.  Invoking  TchebychefFs  inequality  with  respect  to  the  sample 
mean,  we  have 


(5,13) 
or 

P(ju  -  £cr/V^"  <  X  <  fjL  +  £<r/vX>   >   1  —  l/£2-  (5.14) 

Setting  K~k<r/*\/n,  it  is  seen  that 

(5.  15) 


Thus,  when  sampling  from  any  population  with  a  finite  variance,  the 
sample  size  may  be  chosen  large  enough  to  make  it  almost  certain  that 
the  sample  mean  will  be  arbitrarily  close  to  the  population  moan.  This 
is  what  is  known  as  the  law  of  large  numbers. 

In  reliability  and  quality  control  work,  much  attention  is  given  to 
the  number  of  defective  items  in  a  sample  of  size  n.  Thus  it  will  be  of 
interest  to  see  how  the  law  of  large  numbers  provides  information  in 
such  a  case.  If  we  assume  random  sampling  from  a  binomial  population 
m  which  M~P  and  <r2~pCL  —  p),  then,  as  n  gets  large, 

^1  (5*16) 

where  x  is  the  number  of  defective  items  observed  in  a  sample  of  size  n 
and  e  is  an  arbitrarily  small  positive  quantity.  That  is,  as  n  increases, 
we  become  more  and  more  certain  that  the  observed  fraction  defective 
will  be  a  good  estimate  of  the  true  fraction  defective  in  the  population. 

5,5     CENTRAL   LIMIT  THEOREM 

Without  doubt,  the  most  important  theorem  in  statistics  is  the  can- 
tral  limit  theorem.  It  is  important  not  only  from  the  theoretical  point 
of  view  but  also  becaxise  of  its  impact  on  statistical  methods.  Since  a 
proof  of  this  theorem  Is  beyond  the  scope  of  this  text,  it  will  be  stated 
without  proof.  Hero,  then,  is  the  theorem  : 

//  a  population  has  a  finite  variance  of  <r%  and  mean  M?  then  the  dis 
tribution  of  the  sample  mean  approaches  the  normal  distribution  with 

the  variance  <r*/n  and  mean  jj,  as  the  sample  size  n  increases* 

t 
Note  that  nothing  Is  said  about  the  form  of  the  sampled  population. 


5.7      THE    HYPERGEOMETRIC    DISTRIBUTION  73 

That  is,  no  matter  what  the  form  of  the  sampled  population,  provided 
only  that  it  has  a  finite  variance,  the  sample  mean  will  be  approxi 
mately  normally  distributed.  This  is  indeed  a  remarkable  theorem. 

5.6      RANDOM  SAMPLINC  FROM  A  SPECIFIED 
POPULATION 

Suppose  a  random  sample  of  n  observations  is  obtained  from  a  given 
population.  The  joint  probability  density  function  for  (JSTi,  X2,  -  -  -  ,  Xn} 
will  be  represented  by  g(x^  •  •  •  ,  x«).  Now,  it  will  be  remembered  that 
a  random  sample  implies  statistically  independent  observations.  Fur 
ther,  for  statistically  independent  variables,  the  joint  probability 
density  function  may  be  expressed  as  the  product  of  the  marginal 
densities.  Thus, 


'     •    gnOn)  (5-17) 

and,  since  each  observation  came  from  the  same  population, 


where  f(x)  is  the  probability  density  function  describing  the  sampled 
population.  It  should  be  noted  that  Equation  (5.18)  gives  the  joint 
probability  density  function  of  the  sample  in  the  order  drawn.  Inci 
dentally,  the  function 

n  /(**•) 

i=i 
is  often  referred  to  as  the  likelihood  function. 

5.7     THE   HYPERGEOMETRIC   DISTRIBUTION 

In  many  instances,  the  type  of  sampling  performed  in  industrial  ap 
plications  is  the  selection  of  a  sample  of  n  items  out  of  a  lot  of  N  items. 
This  selection  is  usually  done  in  such  a  manner  that  the  sampling  is 
without  replacement.  Thus  we  have  a  "random"  sample  only  in  the 
specialized  sense  that  every  possible  group  of  n  items  in  the  lot  has  the 
same  chance  of  comprising  the  sample.  In  such  a  case,  if  x  represents 
the  number  of  defective  items  in  the  sample, 


/(*)   -  C(D7  x)-C(N  -  D,n-  x)/C(N,  n)  ; 
a-a,a+l,  ...,»-!,& 
a  =  max  (0,  n  —  N  +  2?) 
b  =  min  (D,  n) 

where  D  represents  the  number  of  defective  items  in  the  lot. 

The  distribution  specified  by  Equation  (5.19)  is  known  as  the  hyper- 


74  CHAPTER    5,    SAMPLING    DISTRIBUTIONS 

geometric  distribution.  It  is  the  distribution  underlying  practically  all 
acceptance  sampling  by  attributes  where  an  item  of  product  is  classified 
as  either  defective  or  nondefective.  The  reader  should  become  very 
familiar  with  the  hypergeometric  distribution  and  be  competent  in 
evaluating  probabilities  associated  with  it. 

Using  the  theory  of  earlier  chapters,  it  is  seen  that 

M  =  E[X]  =  nD/N  (5.20) 

and 

<r*  =  E[(X  —  M)2]  =  nD(iV  —  D)(N  —  n)/N~(N  —   1).     (5.21) 


Thus,  as  expected,  the  average  number  of  defective  items  in  a  sample 
is  equal  to  the  size  of  the  sample  multiplied  by  the  fraction  of  defective 
items  in  the  lot. 

5,8     THE   BINOMIAL   DISTRIBUTION 

Suppose  that  a  random  sample  of  size  n  is  selected  from  an  infinite 
binomial  population  described  by 

*(y)  -  P"(i  -  p)1-*;    y  =  o,  i  (5.22) 

0  <  p  <  1 

or  that  a  random  sample  of  wize  n  is  selected  (using  sampling  with  re 
placement)  from  a  finite  population  of  N  items,  D  of  which  arc  defec 
tive.  In  the  latter  case  we  can,  therefore,  let  p~f)/N. 

Both  of  the  sampling  situations  described  above  lead  to  the  same 
sairxpling  distribution  of  j?  where  x  represents  the  number  of  defective 
items  in  a  sample  of  n  items.  This  distribution  is  described  by  the  p.f  - 

/(*)  «  C(»,  *)p*(l  -  p  "-*;     *  »  0,  1,  •  •  •  ,  n  (5.23) 

0  <  p  <  1. 

Using  theory  already  developed,  it  can  easily  be  shown  that 

M  ^  /<;[A*1  «  np  (5,24) 

and 

<r*  «  /4(-Y  -  M)^J  »  np(l  —  ^)  «  npg  (5,25) 


where  #«  1  ~-  p.  Those  results  will  prove  usefxil  in  later  work, 

Probabilities  associated  with  the  binomial  density  of  Equation  (5*23) 
or  with  its  cumulative*  form  have  boon  published  by  the  National 
Bureau  of  Standards  (4),  Robertson  (7),  and  Hornig  (8),  Reference  will 
be  made  to  nueh  tables  as  the  need  arises. 

5,9      BINOMIAL   APPROXIMATION    TO   THE 
HYPERGEOMETRIC 

Under  certain  conditions  it  in  permismhle  to  UHC  the  binomial  distri 
bution  aa  an  approximation  for  the  hypergeometrie  contribution.  This 


5.10       POISSON    APPROXIMATION    TO    THE    BINOMIAL  75 

approximation  is  usually  invoked  to  simplify  numerical  calculations. 
To  see  how  the  approximation  is  justified,  consider  the  hypergeometric 
distribution 

/(#)  =  C(D,  x)-C(N  —  D,n  —  x)/C(N,  n).  (5.26) 

Writing  this  out  in  detail,  we  obtain 


IN    N  —   1  N  —  (x  -  1)     N  —  x 

N  -  D  -  1  N  -  D  -  (n  -  x  -  I)' 

N  —  x  —  1  j\f  ~  (n  —  i) 


(5.27) 


Setting  D  =  pN  and  dividing  the  numerator  and  denominator  of  each 
factor  inside  the  braces  by  N,  it  is  seen  that 


1  -   l/N  1  —  O  -  1)/7V   1   —  x/N 

q  —  l/N  q  —  (n  ~  x  —  1)/N\ 


1  -  (n  - 
where  g=  1  —  p.  Letting  A7"  get  very  large,  it  is  clear  that 

/O)  —  >  C(n,  oc)p*q"-*.  (5  .  29) 

That  is  ,  if  JV  is  large,  the  hypergeometric  distribution  may  be  ap 
proximated  by  the  binomial  distribution.  The  question  of  how  large  N 
should  be  relative  to  n  before  using  the  binomial  approximation  is  one 
which  must  be  answered.  First,  since  tables  of  logarithms  of  factorials 
are  not  available  for  k  greater  than  2000,  calculation  of  the  hyper 
geometric  will  be  extremely  tedious  for  such  cases.  Second,  and  perhaps 
more  to  the  point,  Burr  (1)  has  said  that  if  the  lot  size  N  is  at  least 
eight  times  the  sample  size  n,  it  will  be  satisfactory  to  use  the  binomial 
as  an  approximation  to  the  hypergeometric.  However,  since  Burr's 
statement  is  only  a  general  comment  with  no  reference  to  the  magni 
tude  of  the  error  involved,  it  seems  only  fair  to  say  that  each  individual 
case  must  be  considered  on  its  own  merits. 

5.10   POISSON  APPROXIMATION  TO  THE  BINOMIAL 

In  instances  when  we  do  not  have  access  to  published  tables  of  the 
binomial  distribution,  it  becomes  necessary  to  find  some  way  of  ob 
taining  the  required  probabilities  without  excessive  calculation.  In 
such  cases,  we  usually  seek  some  form  of  approximation  to  the  binomial 
which  involves  less  computation  or  is  associated  with  tables  that  are 
more  readily  available.  Two  such  approximations  involve,  respectively, 
the  Poisson  and  normal  distributions.  The  first  of  these  will  be  dis 
cussed  in  the  present  section,  while  the  normal  approximation  will  be 
examined  in  Section  5.11. 


76  CHAPTER    5,    SAMPLING    DISTRIBUTIONS 


If  p  is  very  small  (less  than  0.1)  and  n  is  quite  large  (greater  than  50), 
it  is  sometimes  convenient  to  approximate  the  binomial  p.f.  by  the 
Poisson  p.f.  in  which  jj.  —  np.  To  see  how  this  approximation  is  justi 
fied,  consider  the  following  argument.  In 


(5.30) 


f(x)  =  C(n, 

n(n  —!)-••(«  —  x  -f-  1) 


set  p  =  jji/n.  Then, 

n(n  -  1)   -  -  •  (n  -  x 


xl 


/  M  \y  /*  y 

f  _  J  f  !   _  _  1 

\  n  /   \  n  / 


/  w  \  /n  —  1\  /n  —  x+  1\    ^ 

""  w/  v  w  /  "  *  *  v^    «     /  ~^ 


* 


n 


If  we  let  n—  *>co  and  p—  ^0  such  that  np^p.  remains  constant, 

/(*)  -+  (1)(1)  -  -  (t)  -™-  e-*(O  -  ^  Y-  (5.32) 

»\  xl 

which  is  the  Poisson  probability  function. 

Therefore,  if  np  is  large  relative  to  p  and  n  i&  largo  relative  to  np?  the 
Poisson  may  bo  used  as  a  reasonable  approximation  to  the  binomial. 
All  that  is  necessary  is  to  net  p  in  the  Poisson  distribution  equal  to  np 
of  the  binomial  distribution  we  are  attempting  to  approximate.  In 
other  words,  the  means  of  the  two  distributions  have  been  equated. 

5.11      NORMAL  APPROXIMATION  TO  THE  BINOMIAL 

The  binomial  distribution  may  also  be  approximated  by  the  normal 
distribution.  As  in  the  preceding  section,  the  sample  »i2o  should  be 
reasonably  large  before  the  approximation  IB  employed, 

To  illustrate  the  nature  of  the  approximation,  ccmsidor  Figure  5.1. 
Hero,  the  binomial  distribution  for  n^  10  and  p^^  is  pictured  by  the 
ordinates  at  the  various  values  of  x.  If  rectangles  of  width  one  are 
erected  as  shown,  the  area  of  the  histogram  equala  1,  This  is  just  an 
alternative  way  of  expressing  the  fact  that  the  sum  of  the  ordinates 
equals  1.  Umng  areas  under  the  normal  curve,  probabilities  associated 
with  various  x  values  may  be  closely  approximated. 


5.11        NORMAL    APPROXIMATION    TO    THE    BINOMIAL 


01          23456          789         1O 
X=  NUMBER    OF   FAILURES   IN  SAMPLE  OF    SIZE    1O 

FIG.  5.1— Binomial  distribution  for  n  =   10  and  p  =    V2    (solid  line 
ordinates),    area    representation    (dotted    line    rectangles) 
and    the    normal    approximation. 

In  order  to  evaluate  probabilities  associated  with  a  normal  distribu 
tion,  the  mean  and  variance  must  be  known.  To  specify  the  mean  and 
variance  of  the  approximating  normal,  let  v  =  np  and  cr2  =  np(l  — p) 
=  npq  where  np  and  npq  are  the  mean  and  variance,  respectively,  of 
the  binomial  distribution  to  be  approximated.  Then,  for  any  integers  a 
and  b  (a<6)  in  the  closed  interval  (0,  n),  the  approximation  takes  the 
form : 


P{a  <:  X  <  b} 
P{a  <  X  <b] 
P{a  <  X  <  b} 


.,{ 


(a  —  £)  —  np 


P<- ^= —^Z 

-v/npg 


(a 


—  np 


•\/Vfcp<? 

O  —  £)   —  ^P 
-\/npq 


<  ^ 
^  Z 


\o  -f-  tJ   —  ^P 

(* 

^/npq 
+  i)  ~ 

np 

* 

-Vnpq 

np 

-^/npq 

or 


-  np 


-\/npq 


(5 . 33) 
(5.34) 
(5.35) 

(5 . 36) 


Other  illustrations  could  be  given,  but  the  foregoing,  together  with  the 
examples  which  follow,  should  be  sufficient.  The  important  thing  to 
note  is  that  ^  is  added  to  or  subtracted  from  the  limit  so  as  to  include 
or  exclude  a  or  6,  the  proper  choice  being  indicated  by  the  nature  of  the 
inequality.  This  adding  or  subtracting  of  £  is  often  referred  to  as  a 
"correction  for  continuity. " 

Example  5.1 

A  random  sample   of   100    observations   is   drawn   from   a   binomial 
population  in  which  p=0.2.  Evaluate  P  {10^X^25}.  We  say  that 


78  CHAPTER    5,    SAMPLING    DISTRIBUTIONS 


4  "~~  4 

-  P{-  2.62  <  Z  <  1.38} 

=  G(1.38)  —  G(  —  2.62)  -  0.91621  -  0.00440 
=  0,91181. 

Example  5.2 

Referring  to  Example  5.1,  evaluate  P  {  10  <  A"<25  }  .  We  have 


4         ~     4     > 
=  P{-  2.37  <  Z  <  1.38}  =  G(1.38)  -  G(-  2.37) 
=  0.91621  —  0.00889  =*  0.90732. 

Example  5.3 

Referring  to  Example  5,1,  evaluate  P  {X  >2G  }  .  Proceeding  as  before, 
P{X>26}   c* 


—  1  —  G(1.62)  *  1  —  0,94738  =  0.05262. 

It  is  reasonable  to  ask  what  error  is  involved  in  using  the  approxima 
tion  just  described.  Mood  (3)  has  said  that,  if  npq>25,  the  error  is  less 
than  Q.l5/^/npq.  However,  we  should  realize  that,  for  a  given  n,  the 
normal  curve  gives  a  better  approximation  when  p  is  close  to  §  than 
when  p  is  close  to  0  or  1.  On  the  other  hand,  if  n  is  large  enough  (say 
100  or  more),  the  approximation  will  be  satisfactory  for  most  values  of 
p.  If  p  is  very  close  to  0  or  1,  the  approximation  will  be  lews  reliable  in 
the  tails  than  near  the  center  of  the  distribution.  Thus,  in  reliability 
work,  where  very  small  values  of  p  are  frequently  encountered,  the 
normal  approximation  may  not  be  too  good  and  one  should  use  either 
the  Poisson  approximation  or  calculate  exact  probabilities. 

5.12     THE   MULTINOMIAL   DISTRIBUTION 

If  a  random  sample  of  size  n  is  taken  from  the  multinomial  popula 
tion  described  by 


0  <  pi  <  1  (5.37) 

0,  1 


t—  1  »—l 

a  multinomial  distribution  is  obtained.  This  distribution  is  defined  by 

==  Pi*»2>««  •  •  •  p*rs      0  <      i  <  I  5 


Pi  -  1 


i**l 


5.13       THE  NEGATIVE  BINOMIAL  DISTRIBUTION  AND  THE  GEOMETRIC  DISTRIBUTION       79 

Xi  =  0,  1,  -  -  -  ,  n 

k 


where  x+  is  the  number  of  items  occurring  in  the  class  associated  with  pt-. 
The  number  pt  is  the  probability  of  any  item  being  assigned  to  the  ith 
class  and  it  is,  of  course,  the  fraction  of  the  total  population  belonging 
to  the  fth  class.  For  example,  an  item  of  product  may  be  assigned  to 
one  of  four  classes:  good,  minor  defect,  major  defect,  or  critical  defect. 
Then,  the  n  sample  items  would  be  classified  into  the  four  groups  upon 
inspection.  The  number  falling  in  the  first  group  would  be  denoted  by 
o?i,  the  number  in  the  second  group  by  x2,  and  so  on. 

5.13     THE  NEGATIVE  BINOMIAL  DISTRIBUTION  AND 
THE  GEOMETRIC   DISTRIBUTION 

A  sampling  distribution  encountered  fairly  often  in  industrial  appli 
cations  is  that  known  as  the  negative  binomial  distribution.  Suppose  p  is 
the  probability  of  a  defective  item  and  g  =  1  —p  is  the  probability  of  a 
nondefective  item.  If  random  sampling  is  being  carried  out,  it  is  fre 
quently  of  importance  to  know  the  probability  that  the  rth  defective 
unit  will  occur  on  the  (x+r)th  unit  sampled. 

To  obtain  the  probability  just  described,  it  is  noted  that:  (1)  the  last 
unit  must  be  defective  and  (2)  in  the  preceding  x+r—  1  units  sampled 
there  must  be  exactly  r  —  I  defective  units.  Then, 


=   {C(x  +  r  -  l,r  -  l)p-HT}  -p 

*;      *  -  0,  1,  •  •  •  .         (5.39) 


Another  way  of  saying  this  is  that  the  probability  of  the  rth  defective 
unit  occurring  on  the  mth  unit  sampled  is 

s(m)  =  C(m  —  1,  r  —  l)pr^-r;      m  =  r,  r  +  1,  -  -   •  .       (5.40) 


It  is  sometimes  of  interest  to  know  the  probability  of  the  rth  defective 
unit  occurring  on  the  rth  or  (r  +  l)st  or  ...  or  nth  unit  sampled.  This  is 
given  by 

n  n 

X)  C(m  —  l,r  —  l}pTqm~r  =   ^  C(n,  w)pmgw-m  (5  .41) 

w«r  mT 

and  the  last  expression  may  be  found  by  consulting  tables  of  the  cumu 
lative  binomial  distribution. 

If  in  Equation  5.39  we  let  r=l,  the  negative  binomial  distribution 
simplifies  to  the  geometric  distribution. 


SO  CHAPTER   5,   SAMPLING    DISTRIBUTIONS 

5.14      DISTRIBUTION    OF    A    LINEAR    COMBINATION    OF 
NORMALLY   DISTRIBUTED  VARIABLES 

Suppose  we  consider 

U  =   i^aiXi  (5.42) 

i—1 

where,  for  the  moment,  all  that  is  known  is  that  the  a*  are  constants 
and  the  Xi  are  variables.  It  is  clear  that 


(5.43) 

<— 1 

and 


where  yu*  is  the  mean  of  X+,  of  is  the  variance  of  -X",-,  and  o\-/  is  the  co- 
variance  of  Xi  and  X3.  If  all  the  -X\-are  mutually  (pairwise)  independent, 

4  -!>?*•*  (5-45) 

since  cr^  equals  0  if  X^  and  X$  are  statistically  independent. 

Consider  now  the  case  where  X*  is  a  random  sample  from  a  normal 
population  with  mean  n*  and  variance  erf  (i=  1,  -  -  -  ,  n).  In  this  case  it 
may  be  shown  that  U  is  also  normally  distributed.  If  each  X^  is 
randomly  selected  from  the  same  normal  population,  that  is,  from  a  popu 
lation1  A^(ju,  <r)>  then  U  is  normally  distributed  with  mean 


and  variance 

«r«  £  «'«. 

i—1 

5.15      DISTRIBUTION   OF  THE  SAMPLE   MEAN    FOR 
NORMAL   POPULATIONS 

In  Sections  5.1,  5.2,  and  5.5,  it  was  stated  that; 


(2)   <r|  «  <r^/n,  and 

1  Tho  notation  N(t*f  <r)  stands  for  "normally  distributed  with  mean 
standard   deviation  <r/' 


5.17       CHI-SQUARE    DISTRIBUTION  81 

(3)  regardless  of  the  form  of  the  sampled  population  (provided  it 
has  a  finite  variance),  the  distribution  of  the  sample  mean  is 
asymptotically  normal  with  mean  //  and  variance  a2/n. 

In  the  present  section  it  is  stated  (without  proof)  that  if  a  random 
sample  is  taken  from  a  population  N(p,,  <T),  then  the  sample  mean  will 
be  distributed  N(n,  o-f^/n}  for  all  values  of  n.  It  should  be  clear  that 
the  probability  density  function  for  ~X  is 

*\/  jyif 
~  -  e-n(x-M)2/2^  (5.46) 


and  that  Z=  Vn(X  —  jLt)/cr  is  N(Q,  1). 

5.16      DISTRIBUTION  OF  THE   DIFFERENCE  OF 
TWO  SAMPLE    MEANS 

If  a  random  sample  of  n^  observations  is  obtained  from  a  population 
with  mean  ^  and  variance  <r\  and  if  a  random  sample  of  n%  observa 
tions  is  obtained  from  a  population  with  mean  _M_2  and  variance  v\,  what 
can  be  said  about  the  distribution  of  U  =  Xi  —  X^  where  Xi  is  the  mean 
of  the  first  sample  and  -XT2  is  the  mean  of  the  second  sample? 

Regardless  of  the  form  of  the  populations  sampled,  it  is  true  that 

^V-*2  =  A**!  "~  Vx2  =  MI  —  M2  (5.47) 

and 

2  2 


=  <4  +  4  =  —  +  —  •  (5-48> 

xl  -T-    xt 


If,  howeyer,_the  sampled  populations  are  both  normal,  it  is  also  true 
that  U  =  Xi  —  X%  is  normally  distributed  with  mean  and  variance  given 
by  Equations  (5.47)  and  (5.48).  In  this  situation, 

(5.49) 


/    o~i 

V  ~^ 


is  normally  distributed  with  mean  0  and  variance  1. 

If  the  populations  are  not  normal  but  both  sample  sizes  are  suffi 
ciently  large,  the  central  limit  theorem  may  be  invoked  to  achieve  an 
approximate  normal  distribution  for  the  difference  of  two  sample  means. 

5.17      CHI-SQUARE   DISTRIBUTION 

One  particular  distribution  arises  quite  frequently  in  applied  work 
and  is  known  as  the  chi-square  distribution.  When  referring  to  the  chi- 
square  distribution,  the  parameter  v  is  called  the  degrees  of  freedom. 
The  probability  density  function  for  chi-square  with  v  degrees  of  free- 


82  CHAPTER    5,    SAMPLING    DISTRIBUTIONS 

dom  is  given  by 

ffu\  =  !LLl f_L_  ;     u  >  0  (5  .  50) 

-^   J  2*/*rO/2) 

where  w  is  used  rather  than  x2  (chi-square)  for  ease  in  writing.  The 
cumulative  chi-square  distribution  is  tabled  in  Appendix  4  for  all  in 
tegral  values  of  v  from  1  through  100. 

5.18      DISTRIBUTION     OF     THE     SUM     OF     SQUARES     OF 
INDEPENDENT   STANDARD    NORMAL   VARIATES 

If  a  random  observation  is  obtained  from  a  normal  population  with 
mean  M  and  variance  <r2,  then  the  variable 

Z2  =  (X  —  /-OVV2  (5.51) 

is  distributed  as  chi-square  with  1  degree  of  freedom.  Now,  consider 
the  variable 

u  =  :fc  csr<  -  M*)Vo-«  (5-52) 


where  the  J5T*  are  independently  and  normally  distributed  with  means  /x» 
and  variances  erf.  Then,  U  is  distributed  as  ehi-«quare  with  k  degrees 
of  freedom. 

It  is  clear  that,  if  a  random  sample  of  size  n  is  obtained  from  a  nor 
mal  popxilatkm  with  mean  M  and  variance  a-2, 


U  -          (X*  -  ju)2A*  (5.53) 

t^i 

is  distribtited  as  ohi-Hquaro  with  n  degrees  of  freedom. 

5.19      DISTRIBUTIONS  OF  THE  SAMPLE  VARIANCE 
AND   STANDARD   DEVIATION    FOR   NORMAL 
POPULATIONS 

It  can  be  proved  that  the  mean  and  variance,  3T  and  s2,  of  a  random 
sample  from  a  normal  population  are  statistically  independent*  Further, 
it  is  readily  shown  that  the  variable 


IT  -   (n  -  1W*  -  S  (-Y,  -  ??)*/<?*  (5-54) 

*'«»n 

is  <listribiited  ns  (^hi-scixiare  with  v^n~~*\  degrees  of  freedom. 

From  the  preceding  result,  the  distributions  of   m^  \/m^  «2,  and  s 
can  be  obtained.  These  are: 


1  /    w,   \  («•— i)/2 

(  )  „,,(«-»>/»«-»»•/«-'  (5.55) 


5.21        DISTRIBUTION    OF    F  83 

1  /  n  V 

(wiz)   =  • — • — 

^      y  /VJ  1\ 

r(V) 

\  ,          ,, v  „  ^,      /o^2  /  r-       c*/C\ 

-1  (Vw2)n~2e~'im2/2<r  ,  (5.56) 

V 


n      <«— 


\2cr 


1  /^7    _    1  \  (n—  1)/2 

st)     =    _  -  -  (   -  -  -)  (52)(n-3)/V-<"-l>*2/2*2,  (5.57) 

}  n  -  1\      \    2o-2    / 


r 

and 


/^-    1\ 


_    1  \  Cn 

(5.58) 


5.20      DISTRIBUTION   OF   "STUDENT'S"    t 

Consider  two  independent  random  variables,  Z  and  U,  where  Z  fol 
lows  a  standard  normal  distribution  and  U  follows  a  chi-square  dis 
tribution  with  v  degrees  of  freedom.  Form  the  ratio 

t  =  Z/VW^-  (5.59) 

Then,  the  probability  density  function  of  t  is 

_OT<,<0o       (5.60) 


and  it  is  referred  to  as  the  ^-distribution  with  v  degrees  of  freedom.  This 
distribution  is  extremely  useful  in  many  problems  of  statistical  in 
ference.  A  table  of  cumulative  percentage  points  of  t  is  given  in  Ap 
pendix  5. 

5.21      DISTRIBUTION   OF  F 

Given  two  independently  distributed  chi-square  variates,  U  with  vi 
degrees  of  freedom  and  V  with  ?,  degrees  of  freedom,  it  may  be  shown 
that 

(5.61) 


is  distributed  as  F  with  vi  and  v*  degrees  of  freedom.  The  probability 
density  function  is 


84  CHAPTER    5,    SAMPLING    DISTRIBUTIONS 

+  ; 


/ 
\ 


f(F)  =  -  -  -  -  -  (  —  )       --  .     (5.62) 
J  ""2 


Of  particular  interest  in  applied  statistics  is  the  fact  that  when  two 
random,  samples  are  obtained,  one  from  each  of  two  normal  populations, 
the  ratio 


is  distributed  as  F  with  ^i  =  ni  —  1  and  v^  —  us  —  1  degrees  of  freedom. 
This  will  find  application  when  analyses  of  variance  are  discussed  later. 
Appendix  6  gives  certain  percentage  points  of  the  ^-distribution. 

5.22     ORDER  STATISTICS 

Observations  on  a  chance  variable  usually  occur  in  random  order. 
However,  in  certain  cases,  observations  ordered  according  to  magnitude 
are  encountered.  This  can  happen  in  two  ways:  (1)  the  observations 
were  obtained  in  random  order  but  were  subsequently  reordered  ac 
cording  to  magnitude,  and  (2)  the  observations  naturally  became 
available  in  order  of  magnitxide.  As  an  example  of  the  latter,  consider 
the  life  testing  of  a  group  of  vacuum  tubes.  The  first  observation  to 
arise  is  that  associated  witlx  the  weakest  tube  (i.e.,  the  txibe  with  the 
shortest  life),  the  second  observation  is  associated  with  the  next  weak 
est  tube,  and  so  on.  Since  such  data  occur  fairly  often  in  xudxistrial 
applications,  some  sampling  distributions  associated  with  order  sta 
tistics  will  now  be  discussed. 

Consider  a  population  specified  by  /(a;),  a<x  <  6.  Denote  the  smallest 
and  largest  values  in  a  random  sample  of  n  observations  from  this 
population  by  u  and  v,  respectively.  Then  it  may  be  shown  that 


v)  -  /?(*)>-*;     a^u^i^b.        (5  ,  63) 
The  marginal  p*d,f  /s  of  u  and  v  are 

gl(u)  «  n/O)[l  —  *X«0]n~l;     a  ^  u  £  b        '  (5.64) 

and 

]n~l;     a  ^  v  ^  b.  (5.65) 


These  distributions  are  very  useful  when  dealing  with  problems  involv 
ing  extreme  value®. 

Order  statistics  are  also  valuable  when  dealing  with  the  sample  range, 
ft  ea  D  —  u  =  -JTmax  —  -X°min-  If  Ht  Equation  (5*68)  we  let  v  ««  u+  R>  we  obtain 


,  K)  «  n(n  -  !)/(«)/(«  +  R)[F(u  +  -R)  -  F(u)}»^.      (5-66) 


PROBLEMS  85 

Then 

&— R 


=    f 

+J  n 


g(u,  K)du;     0  <  R  <£  b  —  a.  (5.67) 


It  should  be  noted  that  if,  instead  of  dealing  with  the  joint  distribution 
of  the  range  and  the  smallest  sample  value,  we  deal  with  the  joint  dis 
tribution  of  the  range  and  the  largest  sample  value,  namely, 


g(v,  R)  -  n(n  -  !)/(*  -  K)f(v)[P(i^  -  F(v  -  r}]—\       (5.68) 
then 

/* 
g(v,  R)dv;     0  <  R  <  b  -  a.  (5.69) 

u-f-72 

Equations  (5,67)  and  (5.69)  will,  naturally,  produce  the  same  result. 

Example  5.4 

If  /(re)  «=!,  0  <£  <1,  then  g(R)  =n(n  —  l)jRn~2(l  —  R)  where  0  <JB<1. 


Example  5.5 

There  is  no  simple  expression  for  the  distribution  of  the  range  when 
sampling  from  a  normal  population.  Pearson  (5)  gave  the  values  of  the 
mean  arid  standard  deviation  for  ranges  from  a  standardized  normal 
distribution.  Pearson  and  Hartley  (6)  evaluated  the  probability  integral 
of  the  range  for  sample  sizes  of  2  to  20.  Incidentally,  the  mean  and 
standard  deviation  of  the  range  when  a  standard  normal  population  has 
been  sampled  are  denoted  by  d%  and  ds,  respectively.  That  is,  when 
sampling  from  any  normal  population,  (JLR  =  d%crx  and  crR  =  d^orx.  Selected 
values  of  d*  and  da  are  given  in  Appendix  8. 

Problems 

5.1  How  large  a  sample  should  be  taken  if  we  want  to  be  95  per  cent  sure 
that  3T  will  not  fall  farther  than  cr/2  from  ju? 

5.2  A  book  of  400  pages  contains  400  misprints.  Estimate  the  probability 
that  a  page  contains  at  least  three  misprints. 

5.3  A  lot  contains  1400  items.  A  sample  of  400  items  is  selected.  If  no 
more  than  two  defective  items  appear  in  the  sample,  the  lot  will  be 
accepted.   Evaluate   the   probability  that  the  lot   will  be   accepted, 
assuming  that  the  lot  is  1  per  cent  defective. 

5.4  The  width  of  a  slot  on  a  forging  is  normally  distributed  with  mean 
0.900  inch  and  standard  deviation  0.003  inch.  The  specifications  are 
0,900  ±0.005  inch.  What  percentage  of  forgings  will  be  defective? 

5.5  Referring  to  Problem  5.4,  samples  of  size  5  are  obtained  daily  and 
their  means  computed.  What  percentage  of  these  sample  averages  will 
be  outside  specifications? 

5.6  The  diameters  of  some  shafts  and  some  bearings  are  each  normally 
distributed  with,  standard  deviation  equal  to  0.001  inch.  If  the  shaft 
has  a  mean  diameter  of  0.500  inch  and  the  bearing  has  a  mean  diame 
ter  of  0.503  inch,  what  is  the  probability  of  inter  ference? 


86  CHAPTER    5,    SAMPLING    DISTRIBUTIONS 

5.7  Three  resistors  are  connected  in  series.  Their  nominal  ratings  are  10, 
15,  and  20  ohms,  respectively.  If  it  is  known  that  the  resistors  are 
normally  distributed  about  the  nominal  ratings,  each  having  a  stand 
ard  deviation  of  0.5  ohm,  what  is  the  probability  that  an  assembly 
will  have  a  resistance  in  excess  of  46.5  ohms? 

5.8  Rework  Problem  5.7  assuming  that  the  standard  deviation  is  5  per 
cent  of  nominal  in  each  case. 

5.9  A  "1-poimd"  box  of  candy  is  machine  packed  to  contain  32  pieces  of 
candy.  If  the  weights  of  the  pieces  of  candy  are  normally  distributed 
with  a  mean  of  0.5  ounce  and  a  standard  deviation  of  0.05  ounce,  what 
are  the  probabilities  that  a  customer  receives: 

(a)  loss  than  1  pound,  (b)  less  than  15  ounces,  (c)  more  than  1  pound, 
(d)  more  than  16.2  ounces,  (c)  exactly  1  pound? 

5.10  Referring  to   Problem  5.9  and  assuming  the  standard  deviation  re 
mains  unchanged,  how  should  you  change  the  mean  of  the  process 
so  that  only  1  customer  in  100  will  receive  lews  than  the  advertised 
weight? 

5 J  1  A  factory  assembles  stoves  at  the  rate  of  500  per  week.  On  the  average, 
5  per  cent  of  the  stoves  are  found  to  be  defective,  when  inspected 
following  final  assembly.  What  is  the  probability  that  next  week's 
production  will  contain  less  than  20  defective  wtovew? 

5.12  Review  all  parts  of  the  book  pertaining  to  the  Pomstm,  normal,  chi- 
square,  ty  and  F  distributions,  and  be  certain  that  you  know  how  to 
une  the  tables  in  Appendices  2  through  (K 

References  and  Further  Reading 

1.  Burr,    T.    W.   Engineering  Statistics  and  Quality  (Control.    McGraw-Hill   Book 
Company,  Inc.,  New  York,  1953, 

2.  Ijiebermau,  CK  J»,  ami  Owen,  I").  B,  Tables  of  the  Hyper  geometric  Distribution. 
Stanford  University  ProwH,  Stanford,  Calif.,   19W), 

3.  Mood,  A.  M.  Introduction  to  the  Theory  of  Statistics.  McGraw-Hill  Book  Com 
pany,  Inc.,  Now  York,  1950. 

4.  National  Bureau  of  Standards  Tables  of  the  Binomial  Probability  Distribution* 
Applied  MathmoticH  HerioH  6.  U.S.  Govt,  Print.  Off.,  Washington,  D.C.,  1949, 

5.  Pearson,  K.  S.,  The  percentage  limits  for  the  dintributkm  of  range  in  Hamplen 
from  a  normal  population,  ttiomctrika,  24:  404  -17,  Nov.,   1932. 

0.   -  ,  and  Hartley,  II.  ().,  The  probability  integral  of  the  range  in  wampleH 

of  TV  observation**  from  a  normal  population.    Biometrika^  32:301-10,    April, 
1942. 

7.  Robertson,    W»    II,    Tablvs   of  the   Binomial    Distribution   Function  for  tftmall 
Value®  of  p.  Bandia  Corporation  Monograph  fcSCR-443,  Albuquerque,  N*  Mex«, 
Jan.,  I960. 

8.  Ilomig,   H,   G.   S&-10Q   Binomial   Table*.  John   Wiley  and  Bono,    Inc.,    New 
York,  1953. 


C  H  APTE  R    6 

STATISTICAL  INFERENCE:  ESTIMATION 

IN  THIS  CHAPTER,  general  concepts  associated  with,  that  part  of  statisti 
cal  inference  referred  to  as  "estimation  and  prediction"  will  be  ex 
amined.  Examples  dealing  with  particular  populations  frequently 
encountered  in  applied  work  will  also  be  given. 

6.1      SOME   PRELIMINARY   IDEAS 

In  general,  we  do  not  know  the  values  of  the  parameters  of  the  dis 
tribution  function  or  the  values  of  the  population  mean  and  variance. 
In  practice  we  obtain  a  random  sample  from  the  specified  population, 
assuming  that  we  know  the  form  of  the  distribution  (normal,  binomial,' 
etc.).  From  the  sample  we  attempt  to  estimate  the  true  but  unknown 
values  of  the  population  parameters.  At  this  point  criteria  should  be 
stated  by  which  we  may  judge,  or  evaluate,  different  estimators  of  a 
parameter.  First,  let  us  define  an  estimator  as  some  function  of  the 
sample  values  which  will  provide  us  with  an  estimate  of  the  parameter 
in  question.  Now,  let  us  set  down  certain  desirable  properties  of  a  good 
estimator  which  may  be  used  as  criteria  to  distinguish  between  good 
and  bad  estimators.  Other  criteria  may  be  found  in  the  literature,  but 
the  three  given  here  are  perhaps  the  most  important  from  a  practical 
point  of  view. 

(1)  An  estimator  is  said  to  be  unbiased  if  the  expected  value  of  the 
estimator  is  equal  to  the  population  quantity  being  estimated. 
That  is,  if  §  is  an  estimator  of  0,  §  is  said  to  be  unbiased  if  the  aver 
age  of  all  possible  values  of  §  is  Q. 

(2)  Let  @  be  an  estimator  of  6  calculated  from  a  random  sample  of 
size  n.  If,  as  n  gets  very  large  (i.e.,  approaches  N  where  N  is  the 
number  of  items  in  the  population),  the  probability  that  §  will 
be  very  close  to  6  approaches  1,  or  certainty,  then  0  is  called  a 
consistent  estimator  of  6.  In  other  words,  if  we  take  a  larger  and 
larger  sample,  we  expect  to  get  an  estimate  which  is  very  close 
to  the  true  value,  and  the  probability  that  we  will  do  so  is  very 
great. 

(3)  If  0i  and  §2  are  two  different  (but  both  unbiased)  estimators  of  6 
with  variances  o^  and   <r$2,   respectively,  and  if   <r\<<r\,  then 
we  prefer  <?i  to  $2.  That  is,  in  general,  we  prefer  the  estimator 
(out  of  the  class  of  all  unbiased  estimators)  which  has  the  mini 
mum  variance. 

Estimates  of  the  type  discussed  above  are  of  a  special  kind  known  as 
point  estimates.  There  is  a  second  class  of  estimates,  however,  known  as 


C873 


88  CHAPTER    6,    STATISTICAL    INFERENCE:    ESTIMATION 

interval  estimates.  These  are  very  important  in  statistical  methodology 
and,  if  at  all  possible,  we  obtain  an  estimate  of  this  type.  Let  us  illus 
trate  the  difference  between  point  and  interval  estimates  by  a  short 
example. 

Example  6.1 

If  I  wish  to  estimate  the  average  weight  of  the  people  in  a  class 
room,  I  could  take  a  random  sample  of  five  people,  record  their  weights, 
and  average  them.  The  resulting  average  (suppose  it  turned  out  to  be 
160  pounds)  would  be  my  point  estimate.  However,  this  is  not  suffi 
cient  for  our  purposes.  If  *I  say  that  the  true  average  weight  of  all  the 
people  in  the  room  (they  are  my  population)  is  between  0  and^  300 
pounds,  I  am  very  confident  of  myself — in  fact,  I  am  almost  certain  of 
my  statement.  Rut  if  I  make  my  interval  much  smaller — and  in  prac 
tice  the  interval  should  be  as  small  as  possible— my  degree  of  confidence 
m  my  interval  estimate  will  become  less.  For  example,  if  I  say  1  think 
that  the  average  weight  of  all  people  in  the  room  is  between  158  and 
162  pounds,  my  degree  of  confidence  may  bo  quite  small, 

If  I  wish  to  be  ahlo  to  evaluate  my  degree  of  confidence  for  any 
interval  estimate,  it  is  customary  to  make  certain  assumptions  con 
cerning  the  distribution  of  the  observations  being  obtained.  Several 
examples  of  such  confidence  intervals  -will  be  studied  later  in  this 
chapter. 

6.2      METHODS  OF  OBTAINING  POINT  ESTIMATORS 

Several  principles  of  estimation,  leading  to  routine  mathematical 
procedures,  have  been  proposed  for  obtaining  "good'*  estimators. 
These  include: 

(1)  The  principle  of  momenta 

(2)  Minimum  chi-sqxiaro 

(3)  The  method  of  leant  squares 

(4)  The  principle  of  maxinuim  likelihood. 

The  application  of  these  principles  in  particular  canes  will  lead  to 
estimators  which  may  differ  and  hence  poHsesn  different  attributes  of 
"goodness."  A  principle  much  in  line,  yielding  estimators  with  many 
desirable  attributes  of  "goodness"  and  obtained  by  easily  applied 
routine  mathematical  procedures,  is  that  of  maximum  likelihood  de 
vised  by  H.  A,  Fisher  (7,  8}*  This  important  principle  of  estimation 
will  be  used  in  the  remainder  of  the  chapter. 

The  procedure  for  determining  the  maximum  likelihood  estimate  of 
a  population  parameter  6  is  as  follows; 

(1)   Determine  the  density  function  of  the  sample,  0(*Yt,  -X"*»  -  *  -  » 
Arn;  0).  Note  that  in  Section  5.0, 


4**  I 


6.4       CONFIDENCE    INTERVALS:    GENERAL    DISCUSSION 


89 


was  referred  to  as  the  likelihood  function* 
(2)    Determine 


=  log 


I—  1 


This  step  is  not  essential.  However,  since  likelihood  functions 
are  products,  and  since  sums  are  usually  more  convenient  to  deal 
with  than  products,  it  is  customary  to  maximize  the  logarithm 
of  the  likelihood  rather  than  the  likelihood  itself, 
(3)  Determine  the  value  of  6  which  will  maximize  L  by  solving  the 
equation 


6.3      MAXIMUM    LIKELIHOOD    ESTIMATORS 

Rather  than  burden  the  reader  with  the  details  of  obtaining  maxi 
mum  likelihood  estimators,  the  results  for  four  of  the  more  common 
distributions  are  presented  in  Table  6.1. 


TABLE  6.1-Maximum  Likelihood  Estimators  Associated  With  Certain 

Distributions 


Distribution 

Parameter 

Maximum  Likelihood  Estimator 

Binomial  

7-> 

p  =f/n  =  observed  relative  frequency 

Poisson  

X(=M) 

X=jLt=jr=  53  x/n 

Normal                 

a(=l*) 

a  =  £=X  =   21  X/n 

Exponential  

b\  =  a*) 
0(=/i) 

^=^^m^   ^Z(X~T^/n 
0=£=X=   ]£  X/n 

6.4      CONFIDENCE    INTERVALS:   GENERAL    DISCUSSION 

A  point  estimate  of  a  parameter  is  not  very  meaningful  without  some 
measure  of  the  possible  error  in  the  estimate.  An  estimate  §  of  a  param 
eter  6  should  be  accompanied  by  some  interval  about  §,  possibly  of  the 
form  6  —  d  to  §+d,  together  with  some  measure  of  assurance  that  the 
true  parameter  6  does  lie  within  the  interval.  Estimates  are  often  given 
in  such  form.  Thus,  the  activated  life  of  a  thermal  battery  may  be 
estimated  to  be  300  ±  20  seconds  with  the  idea  that  the  life  is  unlikely 
to  be  less  than  280  seconds  or  greater  than  320  seconds.  The  develop 
ment  engineer  engaged  in  research  on  capacitors  may  estimate  the 
mean  life  of  a  certain  type  of  capacitor  under  stated  conditions  to  be 
300  ±  50  hours  with  the  implication  that  the  correct  average  life  very 
probably  lies  between  250  and  350  hours.  The  failure  rate  for  a  specific 
component  might  be  estimated  as  being  less  than  0.02  with  the  feeling 


90  CHAPTER    6,    STATISTICAL   INFERENCE:    ESTIMATION 

that  the  true  failure  rate  is  most  likely  no  greater  than  the  stated  limit. 
In  this  last  case,  the  point  estimate  might  have  been  anywhere  between 
0  and  0,02. 

Confidence  intervals  enable  us  to  obtain  a  useful  type  of  information 
about  population  parameters  without  the  necessity  of  treating  such 
parameters  as  statistical  variables.  It  should  be  clearly  understood  that 
we  are  merely  betting  on  the  correctness  of  the  rule  of  procedure  when 
applying  confidence  interval  techniques  to  a  given  experiment.  It  will 
be  observed  in  the  following  sections  that  this  technique  may  be  ap 
plied  to  various  familiar  population  parameters  such  as  the  mean  and 
variance. 

An  examination  of  the  following  sections  will  reveal  that  the  method 
for  finding  confidence  intervals  consists  in  first  finding  a  random 
variable,  call  it  Z,  that  involves  the  desired  parameter  0  but  the  dis 
tribution  of  which  does  not  depend  upon  any  other  unknown  param 
eters.  Next,  two  numbers,  Z\  and  Z*9  are  chosen  such  that 

P{Z1  <  Z  <  Z*}   =  T  (6.1) 

where  y  *s  the  desired  confidence  coefficient,  such  as  0.95.  Then  the 
two  inequalities  are  manipulated  so  that  the  probability  statement 
assumes  the  form 

P{L  <  0  <  U]   -  y  (6.2) 

whore  L  and  U  are  random  variables  depending  on  Z  but  not  involving 
0.  Finally,  we  substitute  the  sample  values  in  fj  and  U  to  obtain  a 
numerical  interval  which  is  the  desired  confidence  interval,  It  is  clear 
that  any  number  of  confidence  intervals  can  be  constructed  for  a  pa 
rameter  by  choosing  %i  and  #2  differently  each  time  or  by  choosing  dif 
ferent  random  variables  of  the  Z  type, 

The  above  has  been  concerned  with  what  is  called  a  two-sided  con 
fidence  interval.  However,  we  sometimes  do  not  care  how  much  our 
estimate  may  err  in  one  direction  provided  that  it  is  not  too  far  off  in 
the  other.  For  example,  xve  may  he  estimating  a  standard  deviation 
which  we  hope  will  be  small.  We  would  be  concerned  only  about  an 
upper  limit  and  hence  would  want  an  interval  of  the  form 

p[e  <  U}   «  y.  (6.3) 

The  theory  of  one-sided  intervals  is  basically  the  same  as  for  two-sided 
intervals. 

6.5     CONFIDENCE  INTERVAL  FOR  THE  MEAN  OF  A  NOR- 
IVIAL   POPULATION 

All  the  necessary  statistics  are  now  available  to  make  possible  an 
excellent  scheme  for  estimating  the  mean  of  a  normal  population.  It  has 
already  been  Haul  that  3T  in  an  xmhianed  estimate  of  M-  However,  it  is 
possible  to  learn  a  little  more  about  the  estimate,  namely,  whether  3T 


6.5       CONFIDENCE    INTERVAL:    MEAN    OF    NORMAL    POPULATION  91 

is  close  to  M  or  likely  to  be  far  removed  from  ju-  Making  use  of  the  f-dis- 
tribution,  the  following  statement  can  be  made:1 

P\X   —    £o.975(n-l)S^-    <    M    <    X   +    ^ 
=*     P<X   —    *0.975(n-l)-4:=    <   »    <X  +   /0.975Cn~l)    ~=          =    0.95  (6.4) 

v  -\/n 


where  io.arscn—  o  is  a  numerical  quantity  extracted  from  the  table  in  Ap 
pendix  5  under  the  column  labeled  0.975  and  for  n  —  1  degrees  of  free 
dom.  The  above  statement  (Equation  6.4)  is  read:  the  probability, 
before  the  sample  is  drawn,  that  the  random  interval 


will  cover,  or  include,  the  true  population  mean  /z,  is  equal  to  0.95. 
Thus,  if  a  random  sample  is  obtained  from  a  normal  population  with 
mean  M  and  variance  cr2,  and  the  two  quantities 

L    =    X    —    2o.975(«—  D-Sjg-  (6.5) 

and 

U    =    X   +    £o.975(n-l)Sjc  (6.6) 


are  computed,  it  can  be  said  that  one  is  95  per  cent  confident  that  the 
true  mean  ju  will  be  in  the  interval  (Z/,  C7).  One  does  not  say  that  the 
probability  is  0.95  that  M  lies  between  L  and  U  but  only  that  one  is 
95  per  cent  confident  that  AC  does  lie  between  L  and  U.  This  distinction 
is  made  because  M  either  does  or  does  not  fall  between  L  and  U;  the 
probability  is  either  0  or  1  for  /UL  is  a  constant  and  does  not  possess  a 
probability  distribution.  The  distinction  made  above  is  a  subtle  one 
and  the  concept  may  not  be  fully  appreciated  at  this  time.  However, 
it  is  a  distinction  that  must  be  made. 

Example  6.2 

Consider  the  estimation  of  the  mean  breaking  strength  of  some  par 
ticular  material.  We  take  at  random  a  number  of  samples,  for  this 
example,  six,  and  subject  them  to  test,  recording  the  pressure  at  which 
they  break.  These  values  might  be  as  follows: 

2206  Ibs.  2203  Ibs. 

2209  Ibs.  2206  Ibs. 

2205  Ibs.  2207  Ibs. 

Averaging  these  values,  we  obtain  a  point  estimate,  that  is,  one  value, 
of  2206  Ibs.  This  means  that,  from  our  sample,  a  reasonable  estimate  of 
the  true  (population)  average  breaking  strength  of  the  material  is 
2206  Ibs.  However,  we  do  not  have  any  measure  of  our  degree  of  con- 

1  The  symbol  sg  is  known  as  the  standard  error  of  the  mean,  and  it  is  clearly  an 
estimator  of  <rg  as  defined  by  Equation  (5.5). 


92  CHAPTER    6,    STATISTICAL    INFERENCE:    ESTIMATION 

fidence  in  this  estimate.  If  we  are  willing  to  assume  a  normal  distri 
bution,  we  can  find: 

L    =5    X   —    £o.975(«-l)<Sj£. 

«  2206  —  2.571(0.8165)  =  2203.9  Ibs. 
U  =  X  +  £0.975(71-1)^ 

=  2206  +  2.571(0.8165)  =  2208.1  Ibs., 

where  £0.975(5)  =2.571  was  obtained  from  the  table  in  Appendix  5  and  all 
other  values  were  calculated  from  our  sample.  We  can  now  say  that  we 
are  95  per  cent  confident  that  the  true  population  mean  breaking 
strength  lies  between  2203.0  Ibs,  and  2208.1  Ibs.  A  99  per  cent  confi 
dence  interval  can  be  found  in  a  similar  manner  using  £0.995(5)  =4,032. 
It  should  be  noted  that,  in  general,  a  WOy  per  cent  confidence  interval, 
0  <y  <1,  may  be  obtained  by  using  tfui-foo/ajc™—  D  *n  Equation  (6,4). 


Rather  than  proceed  as  in  Example  6.2,  we  might  have  wanted  only 
a  lower  confidence  limit.  That  is,  we  might  have  no  interest  in  an  upper 
limit  on  breaking  strength,  since  ordinarily  no  harm  can  result  from  the 
material  being  too  strong.  The  statement  needed  (assuming  a  0,95  con 
fidence  coefficient)  is  then 


<  M     =  0.95.  (6.7) 

Note  that  here  io.osu-i)  is  used  instead  of  £o.975(u~»i)  since  we  want  the 
entire  0.05  error  risk  to  he  on  one  side  of  the  limit  rather  than  to  be 
split  equally  beyond  two  limits.  Thus,  we  would  obtain 

L    —    X    —    2o.95<»—  l)$j£ 

«  2206  —  2.015(0.8165)  «=  2204.4,  (6.8) 

and  could  then  state  that  we  are  95  per  cent  confident  that  the  true 
population  mean  breaking  strength  is  above  2204.4  pounds*  In  general, 
the  lower  limit  woxild  be  //  —  3T  —  £7(n~i>S£. 

It  should  be  clear  that  if  only  a  lOOy  per  cent  upper  confidence  limit 
is  desired,  the  procedure  woxald  be  to  calculate 

U    «    J£  +   *-y  («-!)$£.  (6.9) 


6.6     CONFIDENCE  INTERVAL  FOR  THE  MEAN  OF  A  NON- 
NORMAL  POPULATION 

A  question  which  might  logically  arise  is  "What  can  we  do  if  we  want 
a  confidence  interval  estimate  of  the  mean  of  a  nonnormal  population?" 
The  central  limit  theorem  discussed  in  Chapter  5  provides  us  with  an 
answer  which  is  often  satisfactory.  That  is,  unless  the  distribution  is 
rmxch  different  from  normal  and  the  sample  si#e  in  extremely  small,  the 
distribxition  of  sample  means  will  be  nearly  normal  so  that  the  normal 
theory  may  be  applied  with  only  a  small  error, 

However,  if  the  error  introduced  by  the  approximate  procedure  sxig- 
gestod  in  the  preceding  paragraph  cannot  be  tolerated,  we  always  have 
recourse  to  exact  methods  associated  with  the  partictilar  population 
distribution  involved.  No  attempt  will  be  made  to  list  all  the  different 


6.7       CONFIDENCE    INTERVAL:    VARIANCE    OF    NORMAL    POPULATION  93 

situations.  Rather,  we  shall  state  only  that  the  basic  approach  is  always 
the  same  as  outlined  in  Section  6.4.  If  the  need  arises  for  an  exact  answer 
for  a  nonnormal  distribution,  the  reader  is  referred  to  many  such  ex 
amples  in  the  literature.  If  the  particular  case  in  question  cannot  be 
located  in  this  manner,  a  mathematical  statistician  should  be  con 
sulted. 

6.7      CONFIDENCE   INTERVAL   FOR  THE  VARIANCE  OF  A 
NORMAL   POPULATION 

Using  a  technique  similar  to  that  outlined  in  Section  6.5,  a  confidence 
interval  for  estimating  the  variance  of  a  normal  population  can  be 
found.  This  time,  however,  the  chi-square  distribution  will  be  used  to 
obtain  the  confidence  interval  specified  by 


\ 

J 

' 


{ 


Y 

(n—  1)  •X[(l~T)/2]  (n—  1) 

=T  (6.10) 


(n—  1)  [U—  Y)/2]  (n—  1 

and  this  is  read:  the  probability,  before  the  sample  is  drawn,  that  the 
random  interval  (Z/3  Z7),  where 

__  (6_u) 


-1)  X[  (l4_y)  /2] 

and 


U-  -      .  ,  (6.12) 

*  [  Cl-Y)  /2]  (n-1)  *  C  (1—  Y)  /21  Cn—  1) 

will  include  the  true  population  variance  a-2  is  equal  to  y.  Or,  as  it  is 
more  often  phrased,  we  are  100-y  per  cent  confident  that  the  true  popu 
lation  variance  cr2  will  be  in  the  interval  (I/,  C7)  .  For  the  example  used 
in  Section  6.5,  we  find  the  90  per  cent  confidence  interval  for  a-2  to  be 
(1.8,  17.5). 

As  with  means,  we  can  determine  a  one-sided  confidence  interval. 
This  would  be  defined  by 


-\ 

L 

' 


T  (6.13) 

1—  T)  (n—  1)   ' 

if  an  upper  limit  is  desired.  Although  a  lower  limit  is  conceivable,  it 
would  seldom  be  of  interest. 

If  we  are  interested  in  a  confidencejuaterval  for  estimating  cr  rather 
than  cr2,  the  confidence  limits  Z/  =  VZ  and  Uf=\/U,  where  L,  and  U 
are  the  confidence  limits  for  cr2,  may  be  computed.  It  should  be  noted 
that  this  is  not  the  exact  solution.  However,  it  is  sufficiently  accurate 
for  most  purposes, 


94  CHAPTER    6,    STATISTICAL    INFERENCE:    ESTIMATION 

6.8     CONFIDENCE    INTERVAL    FOR  p,    THE    PARAMETER 
OF  A   BINOMIAL   POPULATION 

It  has  already  been  suggested  in  Section  6.3  that  the  best  point  esti 
mate  of  p  is 

p  =  f/n  =  observed  relative  frequency.  (6.14) 

If  a  two-sided,  lOOy  per  cent  confidence  interval  estimate  of  p  is  de 
sired,  the  following  two  equations  must  be  solved  for  p: 

CO,  x)p*(l  -  p)--*  -  (1  -  T)/2  (6.15) 


(6.16) 

The  solution  of  Equation  (6.15)  is  L,  while  the  solution  of  Equation 
(6.10)  is  ?/.  If  /==0,  L  is  taken  to  be  0;  if  /=n,  J7  is  taken  to  be  I.  You 
may  then  state  that  you  are  IQOy  per  cent  confident  that  the  true 
value  of  p  is  between  Ij  and  £7. 

Example  6.3 

Consider  an  industrial  process  producing  parts  which  arc  classified 
either  as  defective  or  mmdefeetive.  In  a  random  sample  of  200  items, 
6  arc  found  to  be  defective.  Thus  p  =0/200  —  0.03.  To  obtain  L  and  U 
such  that  we  would  have  95  per  cent  confidence  in  the  limits,  we  would 
substitute  200  for  n,  0  for  /,  and  0.95  for  y  in  liquations  (6.15)  and 
(6.16)  and  solve  lining  tables  of  the  binomial  distribution*  However, 
due  to  the  nature  of  the  tables,  only  approximate  answers  would  be 
possible.  That  IB,  the  tables  are  not  published  for  small  enough  incre 
ments  of  p  to  permit  an  exact  solution.  Interpolation  would  ho  necessary. 

Because  the  computation  involved  in  solving  Equations  (0.15)  and 
(0,10)  in  tedious,  several  attempts  have  boon  made  to  provide  conven 
ient  tablet*  uncl  graphs  for  the  research  worker  to  use.  For  example, 
Uald  (12)  has  published  comprehensive  tables  for  certain  sample  sixes. 
More  condensed  tables  are  given  in  Rnodooor  (20).  (Copper  atul  Pear 
son  (5)  published  charts  which  arc  very  helpfuL  (1alvert  (4)  gives 
charts  and  nomographs  from  which  one-sided  tipper  eonficlenee  limits 
can  be  read  with  reasonable  accuracy.  Mxieiich  (15)  has  constructed  a 
compact  and  easily  used  slide  rule  which  extends  the  charts  provided 
by  Oalvert. 

As  pointed  out  in  Sections  5.10  and  5.1  1,  it  is  often  possible  to  ap 
proximate  the  binomial  distribution  by  the  Poisson  or  normal  distri- 
btition.  These  approximations  can  sometimes  be  used  to  advantage  in 
establishing  confidence  intervals  for  binomial  probabilities*  However, 
the  details  will  not  be  discussed  here. 


6.9       CONFIDENCE    INTERVAL:    DIFFERENCE    OF   TWO    MEANS  95 

6.9      CONFIDENCE  INTERVAL  FOR  THE  DIFFERENCE  BE 
TWEEN  TH  E  M EANS  OF  TWO  NORMAL  POPU  LATIONS 

Many  practical  problems  in  statistics  involve  the  comparison  of  two 
sample  means.  When  the  two  random  samples  from  which  the  means 
are  computed  can  be  assumed  to  have  come  from  normal  populations, 
confidence  limits  for  the  true  difference,  that  is,  for  the  difference 
between  the  means  of  the  two  populations  may  be  computed. 

Case  I  :  erf  =  cri 

If  it  can  be  assumed  that  the  two  normal  populations  have  equal 
variances,  that  is,  if  we  can  assume  a  common  variance  a2,  then  the 
ratio 

~y        "V  "V"        rV:  /~^V"  V    \.          /  N.    1  /9 

-A.  1  -A-2  -"^  1    -^L  2  /  .A  l     -^-2\      /          W-l^Z-2         X4-'* 


+  nj  (6.17) 


is  distributed  as  Student's  t  with  ni+n2—  -2  degrees  of  freedom  if  s2  is 
calculated  by  means  of  the  formula 

(6.18) 

In  Equation  (6.18)  the  expressions  T^aff  and  2^x1  represent  the  sum 
of  the  squares  of  the  deviations  about  the  means  in  the  first  and  sec 
ond  samples,  respectively.  Also,  s2  is  often  referred  to  as  the  pooled 
estimate  of  variance. 

Under  the  assumptions  stated  above,  100^  per  cent  confidence  limits 
for  /xi  —  ^2  may  be  found  by  calculating 


•+•  n2  —  2  HI  +  n%  —  2 


*  \, *  +js  •         '   i,  \  A~r  f  / 1  **t    \'f  i~T~'"Z       **  /  ~  Ji,  i  —Ji.  2  *"  \         "  s 

Example  6.4 

There  are  two  methods  of  measuring  the  jnoisture  content  of  heat- 
processed  beef.  For  Method  1  we  obtain  ^"1  =  8^6,  s?  — 109.63,  and 
ni  — 41.  For  Method  2  the  comparable  results  are:  Jf2  =  85.1,  s!  =  65,99, 
and  ri2==31.  Thus, 

J2  =  (40(109.63)  +  30(65.99)  }/70  =  90.93, 
and  ^-xa  M  {(90.93/41)  +  (90.93/31) }  ^  =  2.27. 

Finally,   assuming  an  80  per   cent  confidence  interval  is  desired,   we 
obtain  L=*  3. 5-  (1.294)  (2. 27)  ^0.6  and  J7  =  3. 5  + (1.294)  (2. 27)  ^6.4. 


Case  II  : 

If  there  is  reason  to  believe  that  the  two  populations  have  different 
variances,  the  procedure  just  discussed  is  not  appropriate.  What,  then, 
can  be  done?  If  we  are  willing  to  assume  that  Si  —  crf  and  sl  =  crl,  an 


96  CHAPTER    6,    STATISTICAL    INFERENCE:    ESTIMATION 

approximate  100-y  per  cent  confidence  interval  may  be  found  by  calcu 
lating 

i/»i  +  Jl/»2)1/a  (6.20) 

where  3[(14_T)/;>]  is  the  100(1  +  ^)72  fractile  of  the  standard  normal 
distribution.  However,  because  of  the  doubtful  validity  of  the  assump 
tion  that  the  sample  variances  equal  the  population,  variances,  this 
procedure  provides  only  a  very  crude  estimate  of  the  true  mean  differ 
ence.  Consequently,  the  procedure  should  be  used  with  extreme  cau 
tion. 

Case  III:  Paired  Observations 

If  two  samples  of  equal  size  are  obtained  (that  is,  if  ?&i  =  n2==n)  and 
if  the  observations  in  one  sample  can  logically  be  paired  with  the  ob 
servations  in  the  other  sample,  a  modified  procedure  applies.  By  pair 
ing,  it  is  meant  that  the  observations  (Xi,  X%,  •  •  •  ,  Xn)  and  the 
observations  (Fi,  Y2,  -  -  •  ,  Fn)  are  associated  as  follows: 

Xi  is  related  to  Y\ 
Xi  is  related  to  F$ 


Xn  is  related  to  Yn* 

In  the  language  of  a  later  chapter,  the  variables  X  and  Y  are  said  to  be 
correlated.  When  such  a  correlation  in  assumed  to  exist,  an  appropriate 
procedure  is  to  calculate  the  differences,  /->  —  X  —  Y,  and  then  estimate 
/jiD^fj>x~~VY»  Confidence  limits  are  then  given  by 


where  4=*4>/n  and  $%**  {  22/>2-  (  J^D^/n}  /(n-l). 


Example  6.5 

It  IB  desired  to  compare  the  prices  of  Delicious  and  Melntoeh  apples. 
On  a  certain  day>  prices  (per  box)  were  obtained  from  a  random  selec 
tion  of  eleven  markets.  ABBUming  (I)  prices  to  be  normally  distributed 
and  (2)  the  price  of  one  variety  in  a  market  would  he  influenced  by  the 
price  of  the  other  variety  in  the  name  market,  the  method  of  paired 
observations  will  he  tincd*  The  data  ure  given  in  Table  6.2*  Calculation 
yields  75^CU4,  *£  -0.0018,  and  ^-(U)018/1K  Therefore,  a  95  per 
cent  confidence  interval  for  ^/>  is  specified  by 


_  0.14=F  (2.228)  (.OOlS/i  !}»'*£*  ($Q.tI,  $0.17). 

Although,  not  Illustrated  in  Kxample  6.5,  it  should  be  obvious  that  some 


6. TO       CONFIDENCE    INTERVAL:    VARIANCES    OF    TWO    POPULATIONS 


97 


of  the  differences  could  be  negative.  This  is  so  because  the  differences 
are  defined  as  D  =  X—  Y,  not  as  D  =  \X—Y\.  Actually,  it  does  not  mat 
ter  whether  X—Y  or  Y  —  X  is  used  as  long  as  the  same  choice  is  used 
throughout  a  given  problem. 

Rather  than  burden  the  reader  with  excessive  repetition,  we  shall 
only  remind  him  that  one-sided  confidence  limits  are  also  possible.  All 
that  is  necessary  is  a  change  in  the  value  of  t  in  Equations  (6.19)  and 
(6.21),  or  a  change  in  the  value  of  z  in  Equation  (6.20). 


TABLE   6.2-Price  per  Box  of  Delicious  and  Mclntosh  Apples 


Market 

Delicious 

Mclntosh 

Difference 

1  

$2.  15 

$2   32 

$0   17 

2  

2.16 

2   34 

0.18 

3  

2.13 

2.30 

0.17 

4  

2.25 

2.40 

0.15 

5.  .    .     . 

2.20 

2   34 

0.14 

6  

2.18 

2.20 

0.02 

7  

2.27 

2.42 

0.15 

8    .  . 

2.21 

2,36 

0.15 

9  

2.23 

2.36 

0.13 

10  

2.16 

2.30 

0.14 

11    .                . 

2.20 

2.34 

0.  14 

6.10      CONFIDENCE    INTERVAL   FOR  THE    RATIO  OF  THE 
VARIANCES  OF  TWO   NORMAL   POPULATIONS 

The  problem,  of  estimating  the  ratio  of  two  population  variances  (or 
standard  deviations)  is  also  frequently  encountered.  If  the  two  popu 
lations  are  normal,  the  /^-distribution  may  be  used  to  provide  the  de 
sired  confidence  intervals.  The  procedure  is  to  calculate 


and 


U 


=(f) 
-(3) 


r-l.na— 1), 


(6.22) 


(6.23) 


and  these  limits  define  a  100-y  per  cent  confidence  interval  for 
Should  only  aix  upper  (or  lower)  limit  be  desired,  it  can  easily  be  found 
by  using  Fy  in  Equation  (6.23)  or  FI_T  in  Equation  (6.22).  One  other 
useful  result  is  the  following:  If  only  an  abbreviated  F-table  is  avail 
able  (e.g.,  one  that  contains  only  the  upper  percentage  points),  the 
identity 


98  CHAPTER    6,    STATISTICAL    INFERENCE:    ESTIMATION 

-^("i^s)  —  ~  -  (6.24) 

&  U-~?OO2,*'l) 

permits  the  calculation  of  F-values  at  the  left-hand  tail  of  the  distri 
bution. 

Example  6.6 

Using  the  data  given  in  Example  6.4,  99  per  cent  confidence  limits 
for  oi/V?  are  found  to  be: 


L  =  (65.99/109.63)  (0.416)  =  0.25 
U  =  (65.99/109.63)  (2.52)  =  1.52. 

If  a  confidence  interval  for  the  ratio  of  the  standard  deviations  of 
two  normal  populations  is  desired  (that  is,  if  we  wish  to  estimate 
cra/o-x),  it  is  appropriate  to  calculate  Z/=vX  and  tf'===vT7  where  L 
and  U  arc  defined  by  Equations  (G.22)  and  (6.23). 

6-11      TOLERANCE    LIMITS:   GENERAL   DISCUSSION 

One  common  method  used  by  engineers  to  specify  the  q  uality  of 
manufactured  product  is  the  method  of  tolerance  limits.  When  such 
limits  are  quoted,  it  is  expected  that  a  certain  percentage  of  the  product 
will  have  a  quality  between  the  stated  limits.  For  example,,  suppose 
electrical  gaps  are  judged  by  the  characteristic;,  "  transfer  time."  It  is 
then  desirable  to  be  able  to  quote  two  limits,  A  and  /?,  such  that  we 
are  fairly  certain  that,  say,  98  per  cent  of  all  gaps  produced  will  exhibit 
transfer  times  between  A  and  1$>  Such  limits  (dearly  provide  us  with  a 
measure  of  the  quality  of  the  product  under  consideration.  For  certain 
weapon  applications,  it  is  convenient  to  be  able  to  set  a  one-sided  toler 
ance  limit.  An  example  of  such  a  case  is  the  following:  90  per  cent  of  all 
Type  XYZ  batteries  will  yield  an  activated  life  of  at  least  200  seconds. 
In  general,  then,  tolerance  limits  are  limits  within  which  we  are  highly 
confident  will  lie  a  certain  percentage  of  the  individuals  of  a  statistical 
population. 

To  apply  tolerance  limits  in  u  satisfactory  manner,  certain  conditions 
must  be  met.  In  summary,  the  conditions  upon  which  tolerance  limits 
are  based  are  the  following: 

(1)  All   assignable   causes  of  variability   must   be   detected   and 

eliminated  BO  that  the   remaining   variability  may  be   con 
sidered  random. 

(2)  Certain  assumptions  must  be  made  concerning  the  nature  of 
the  statistical  population  tinder  study* 

6.12     TOLERANCE   LIMITS  (TWO-SIDED;  ONE-SIDED) 
FOR  NORMAL  POPULATIONS 

Tolerance  limits  considered  in  this  section  are  based  on  the  assump 
tion  that  the  parent  population  may  bo  described  by  a  normal  dintri- 


6.12       TOLERANCE    LIMITS    FOR    NORMAL    POPULATIONS  99 

but  ion.  If  the  true  mean  and  standard  deviation  are  known,  tolerance 
limits  are  formed  by  adding  to  and  subtracting  from  the  mean  some 
multiple  of  the  standard  deviation.  That  is,  if  &  and  o-  are  known,  toler 
ance  limits  take  the  form  jj,  ±  zo-  where  z  is  selected  from  Appendix  3 
and  depends  only  on  the  proportion  of  the  population  to  be  included 
within  the  calculated  limits.  For  example,  the  limits  M±  1.645<r  include 
90  per  cent  of  a  normal  population  with  mean  JJL  and  standard  devia 
tion  cr.  One-sided  tolerance  limits  may,  of  course,  be  obtained  by  con 
sidering  n+za  or  JUL—ZO-  as  the  problem  requires.  

In  a  practical  situation,  ^  and  cr  are  unknown.  Only  estimates,  ~X 
and  s,  are  available.  While  it  was  true  that  the  limits  /z±1.645<r  will 
include  90  per  cent  of  the  population,  the  same  statement  cannot  be 
made  concerning  J^it  1.645s.  Just  what  proportion  of  the  population 
will  lie  between  X  ±  Ks  depends  on  how  closely  "X  and  s  estimate  M  and 
cr.  Note  that  K  is  used  here  to  represent  the  constant  used  with  'X  and 
5  in  contrast  with  the  z  used  with  ^  and  cr. 

Since  X  and  s,  and  hence  ~X±Ks}  are  random  variables,  it  is  im 
possible  to  state  with  certainty  that  j£  +  Ks  will  always  contain  a 
specified  proportion,  P,  of  the  population.  That  is,  it  is  impossible  to 
choose  K  so  that  the  calculated  limits  will  always  contain  a  specified 
proportion,  P,  of  the  population.  However,  it  is  possible  to  determine 
K  so  that  in  many  random  samples  from  a  normal  population  a  certain 
fraction  y  of  the  intervals  Qc±.Ks)  will  contain  100P  per  cent  or  more 
of  the  population.  When  this  notation  is  used,  P  is  referred  to  as  the 
coverage  and  y  as  the  confidence  coefficient.  This  terminology  is  used 
since  we  are  lOOy  per  cent  confident  that  the  tolerance  range  specified 
by  j£±Ks  will  include  at  least  100P  per  cent  of  the  normal  population 
sampled.  

Intuitively,  it  is  reasonable  to  expect  that  values  of  K  used  with  3T 
and  s  will  be  larger  than  values  of  z  used  with  M  and  <r.  It  is_also  clear 
that  if  K  is  taken  large  enough,  then  the  probability  that  X  ±  Ks  will 
contain  at  least  100P  per  cent  of  the  population  may  be  made  very 
close  to  1.  However,  the  smaller  K  is  taken,  the  more  meaningful  and 
useful  the  tolerance  range  becomes.  The  engineer  is  thus  faced  with  a 
decision:  make  broad  statements  with  little  risk  of  error  or  make  pre 
cise  statements  (i.e.,  a  narrow  tolerance  range)  with  greater  risk  of 
error.  The  problem,  statistically  speaking,  becomes  that  of  finding  the 
smallest  value  of  K  consistent  with  a  specified  confidence  coefficient  y, 
proportion  P,  and  sample  size  n. 

We  must  not  forget  that  one-sided  tolerance  limits  are  frequently 
more  appropriate  than  two-sided  tolerance  limits.  That  is,  it  is  often 
desirable  to  specify  a  single  limit  such  that  a  given  percentage  of  the 
population  will  be  less  than  (or  greater  than)  this  limit.  Such  a  limit 
is  known  as  a  one-sided  tolerance  limit  and  is  usually  of  the  form 
~X+Ks  (or  IX  —  Ks"),  Both  one-sided  and  two-sided  tolerance  limits 
for  normal  populations  will  be  discussed  in  the  following  paragraphs. 

Table  6,3  is  an  abbreviated  table  of  K  factors  for  two-sided  tol- 


too 


CHAPTER    6,    STATISTICAL    INFERENCE:    ESTIMATION 


erance  intervals.  Values  of  K  taken  from  this  table  give  a  95  per 
cent  confidence  that  at  least  a  fraction  P  will  be  included  in  the 
interval  X  +  Ks.  Table  6.4  is  an  abbreviated  table  of  K  factors  for 
one-sided  tolerance  intervals.  Values  of  K  taken  from  this  table  give 
a  95  percent  confidence  that  at  least  a  fraction  P  will  be  above  (below) 
2T—  K$(~X  +  Ks).  Much  more  extensive  tables  can  be  found  in  Bowker 
and  Lieberman  (2),  Eisenhart,  Hastay,  and  Wallis  (21),  Owen  (19), 
and  Weissberg  and  Beatty  (23). 

Example  6.7 

Using  the  data  of  Example  6.2,  find  tolerance  limits  such  that  you 
arc  95  per  cent  confident  of  including  at  least  99  per  cent  of  the  sampled 
population.  These  limits  are  given  by 

3T  ±  Ks  «  2206  ±  5.775(2)  ~  (2194.45,  2217.55). 

TABLE  6.3-Two-Sided  Tolerance  Factors 

(Factors  K  such  that  the  probability  is  0.95  that  at  least  a  proportion  P 
of  the  distribution  will  be  included  between  l?±Ks  where  3T  and  $  are 
computed  from  a  sample  of  size  n.} 


P 

Yl 

0,7500 

0.9000 

0  .  9500 

0.9900 

5  

3.002 

4.275 

5.079 

6.634 

6  

2,604 

3.712 

4.414 

5.775 

7  

2.361 

3  .  369 

4.007 

5  .  248 

8  

2.197 

3.136 

3.732 

4.891 

9  

2  ,  078 

2,967 

3  ,  532 

4.631 

10  

1.987 

2  ,  836 

3,379 

4.433 

17  

1.679 

2.400 

2.858 

3,754 

37  

1  .  450 

2,073 

2.470 

3  ,  246 

145  

1  .280 

1  .829 

2.179 

2.864 

oo  ,  

1  .150 

1  .  645 

1,960 

2.576 

The  foregoing  discussion  of  tolerance  limits  and  the  K  factors  given 
in  Tables  6.3  and  6.4  depend  squarely  on  the  assumption  of  a  random 
sample  from  a  normal  population.  If  tolerance  limits  are  calculated 
using  these  tables  when  the  sampled  population  is  definitely  non- 
normal,  considerable  error  is  possible. 

6-13      D1STRIBUTION-FREE  TOLERANCE   LIMITS 

Sometimes  it  is  desirable  to  Bet  tolerance  limitn  that  do  not  depend 
on  the  assumption  of  normality.  That  is?  we  recognize  that  it  in  riot 
always  possible  to  justify  the  assumption  of  a  normal  distribution.  If 
we  are  dealing  with  a  statistical  variable  that  can  be  described  by  a 
continuous  distribution,  one  very  simple  set  of  dwtribution^fr^a  toler 
ance  limits  is  specified  by  Xmin  and  -Sfmax,  the  smallest  and  largest 


PROBLEMS  1O1 

TABLE  6.4-One-Sided  Tolerance  Factors 

(Factors  K  such  that  the  probability  is  0.95  that  at  least  a  proportion  P 

of  the  distribution  will  lie  above  (below)  ~X  —  Ks(X+Ks)  where  X  and 

s  are  computed  from  a  sample  of  size  n.} 


P 

0.7500 

0  .  9000 

0.9500 

0.9900 

5  

2.150 

3.412 

4.212 

5.751 

6  

1.895 

3.008 

3.711 

5.065 

7  

1.733 

2.756 

3.400 

4.644 

8  

1.618 

2.582 

3.188 

4.356 

9     .... 

1  .532 

2.454 

3.032 

4.144 

10  

1.465 

2.355 

2.911 

3.981 

17  

1,220 

2.002 

2.486 

3.414 

37 

1  .014 

1.717 

2.149 

2.972 

145  

0.834 

1.481 

1.874 

2.617 

CO    .... 

0.674 

1.282 

1.645 

2.326 

values  in  a  random  sample  of  size  n.  Clearly,  the  confidence  in  such 
limits  will  depend  on  n.  Persons  interested  in  reading  further  on 
this  topic  are  referred  to  Murphy  (16),  Ostle  (17),  Owen  (18),  and 
Wilks  (24). 

Problems 

6.1  As  a  physicist  or  chemist,  you  would  soon  become  acquainted  with 
such  "constants*'  as  Planck's  constant  and  Euler's  constant.  To  con 
sider  a  specific  case,  Planck's  constant  is  defined  as  "the  quantum  of 
energy  radiated  from  black  bodies  -5- frequency  of  radiation.33  Suppose 
you  were  attempting  to  find  the  value  of  this  constant  by  experi 
mental  methods.  You  ran  6  experiments  and  obtained  the  following 
estimates  of  h  (Planck's  constant)  : 

6.53  X  10~27 

6.54  X  10~27 
6.58  X  10~27 
6.56  X  10~27 

6.55  X  10~27 
6.55  X  10-27 

What  inferences  can  you  make  about  the  true  value  of  h?  Be  careful 
to  state  explicitly  any  assumptions  you  make. 

6.2  Given  that  n  =  9,  7  =  20,  and 

JC  y*  =  288, 

calculate  a  95  per  cent  confidence  interval  for  ju-  on  the  assumption 
that  you  have  a  random  sample  from  a  normal  population.  Interpret 
this  confidence  interval. 


102  CHAPTER    6,    STATISTICAL    INFERENCE:    ESTIMATION 

6.3  From  a  random  sample   of   100  aptitude  test  scores   drawn  from  a 
normal  population,  the  95  per  cent  confidence  interval  for  ju  is  calcu 
lated  to  be  (45,  55).  Fifty  other  random  samples,  each  of  size  100,  are 
drawn  from  the  same   population,   but   only   10   of  their   means   fall 
within  the  above  limits.  Is  it  not  correct  to  expect  95  per  cent  of  such 
sample  means  to  be  between  45  and  55?  Explain  your  answer. 

6.4  A  random  sample  of  25  observations  from  a  normal  population  had 
a  mean  of  20  and  a  sum  of  squares  of  the  deviations  from  the  mean  of 
2400,  Compute  and  interpret  the  90  per  cent  confidence  interval  for 
the  population  mean. 

6.5  It   has   been   reasonably   well  established   that   a   particular   machine 
produces  nails  whose  length  is  a  random  variable  with  a  normal  dis 
tribution-  A  random  sample  of  5  nails  yields  the  following  results: 

1 . 14  inches 

1 . 15  inches 
1 . 14  inches 
1.12  inches 
1.10  inches 

Calculate  99  per  cent  confidence  limits  for  /*. 

6.6  The  density  of  each  of  27  explosive  primers  was  determined,  with  the 
sample  average  being  1 .53  and  the  wampic  standard  deviation  being 
0.04,  [Determine  a  90  per  cent  upper  confidence  limit  for  /*. 

6.7  The   firing   of    101    rockets  yielded   an   average   range    (i.e.,    distance 
flown)  of  3000  yards  and  a  standard  deviation  of  40  yards.  Determine 
an  85  per  cent  lower  confidence  limit  for  M. 

6*8  Using  the  data  of  Problem  0,1,  compute  a  05  per  cent  confidence 
interval  for  or2. 

6.9  Using  the  data  of  Problem  4.11,  compute  a  90  per  cent  confidence 
interval  for  <r. 

6.10  If  in  a  sample  of  14   holts  the  estimate  of  the   population  standard 
deviation  of  their  lengths  was  #  —  ,,021,  what  are  the  OS  per  cent  con 
fidence  HmitR  for  the  standard  deviation  of  the  population  (<r)?  What 
assximptionB  nmnt  he  made  to  determine  these  limits? 

6.11  Using;  the  data  of  Prohlem  6.10,  determine  a  00  per  cent  upper  con 
fidence  limit  for  <r* 

0.12  Uninpr  the  data  of  Problem  0.5,  determine  a  07,5  per  cent  upper  confi 
dence  limit  for  cr. 

6A3  Using  the  data  of  Problem  0.0,  determine  a  00  per  cent  upper  confi 
dence  limit  for  <r, 

6.14  In   1054  the   mean  earnings  of  68   physicians  in   communities   from 
10,000  to  25,000  was  $13,944,  with  #«$40ia.   Find  the  00  per  cent 
confidence  limits  for  the  population  standard  deviation.  State  your 
assumptions* 

6.15  In  a  random  natnple  of  400  farm  operators*  05  per  cent  were  owners 
arid  35  per  cent  were  mmownern.   Determine  95  per  cent  confidence 
limits  for  the  true  percentage  of  farm  owners  in  the  population  of 
operators?  sampled. 

0,10     In  a  random  sample  of  600  light  bulbs,   12  were  defective.  Determine 

a  95  per  cent  upper  confidence  limit  for  the  true  fraction  defective. 
6.17     Uning  the  data  of  Problem  4.1,  and   the   results   found   in    Prohlem 


PROBLEMS  1 03 

4.10,  determine:  (a)  95  per  cent  confidence  limits  for  /x  and  (b)  80  per 
cent  confidence  limits  for  a2.  State  all  assumptions. 

6.18  Using  the  data  of  Problem  4.2  and  the  results  found  in  Problem  4.10, 
determine:  (a)  a  99  per  cent  upper  confidence  limit  for  M  and  (b)  a  95 
per  cent  upper  confidence  limit  for  <r.  State  all  assumptions. 

6.19  Using  the  data  of  Problem.  4.3  and   the   results    found   in    Problem 
4.10,  determine  50  per  cent  confidence  limits  for  each  of  the  means 
and  50  per  cent  upper  confidence  limits  for  each  of  the  standard  devia 
tions.  State  all  assumptions. 

6.20  You  are  engaged  as  a  testing  engineer  in  an  electrical  manufacturing 
plant.  One  of  the  products  being  produced  is  an  electric  fuse,  and  the 
most  important  characteristic  of  this  fuse  is  the  length  of  time  before 
it  "blows"  when  subjected  to  a  specified  load.  A  testing  program  was 
undertaken  and  the  following  sample  data  (in  seconds)  were  obtained. 


Day  1  Day  2 


42 

69 

45 

109 

68 

113 

72 

118 

90 

153 

Place  a  90  per  cent  confidence  interval  on  the  true  difference  between 
the  means  of  the  two  different  days'  productions.  Assume  that  each 
day's  production  may  be  represented  by  a  normal  population.  State 
all  other  assumptions  which  you  make  and  interpret  your  numerical 
answer. 

6.21  Given  that 

7i  -  75,     »!  «  9,     £  ylt  -  1482,      F2  «  60,     n*  -  16,       £  ytj  =  1830, 
i-i  j-i 

and  assuming  that  the  2  samples  were  randomly  selected  from  2 
normal  populations  in  which  erf  =  erf,  calculate  an  80  per  cent  confi 
dence  interval  for  JLAI-— jus- 

6.22  Two  barley  varieties  have  been  grown  at  a  number  of  locations  over 
several  years  in  an  area  and  their  general  adaptability  is  under  dis 
cussion.  Which  variety  would  you  select  for  the  area  on  the  basis  of  the 
following  yields  in  bushels  per  acre? 

Trebi— 41.2,  19.3,  45.5,  63.9,  63.8,  44.2,  42.5,  53.0. 
Svanota— 39.4,  30.8,  44.5,  51.5,  41.1,  26.5,  35.7. 

Place  confidence  limits  on  the  difference  between  the  means. 

6.23  Two  varieties  of  tomato   were  experimented  with  concerning  their 
fruit-producing  abilities.  The  study  was  done  in  a  greenhouse  and, 
because  of  extreme  variations  (among  locations  within  greenhouses) 
of  temperature,  light  quality,  and  light  intensity,  the  experimental 
plants  were  placed  in  pairs  (one  of  each  variety)  at  several  locations. 
The  following  data  were  obtained: 


1O4  CHAPTER    6,    STATISTICAL    INFERENCE:    ESTIMATION 

WEIGHTS  or  RIFE  FRXTITS  FOR  Two  VARIETIES  OF  TOMATO 

(in  pounds) 


Location 

Variety 

Difference  — 

A 

B 

A  —  B  =  D 

1  

3  .  03 
3.10 
2.35 
3.86 
3.91 
2.65 
1.72 
2.30 
2.70 
3  ,  60 

2.28 
2.68 
2.17 
3.56 
3  .  73 
1.48 
1.85 
1.86 
2.76 
2.68 

.75 
.42 
.18 
.30 
.18 
1.17 
—     .13 
.44 
—     .06 
.92 

2  

3 

4  

5  

6.  

7  

8  

9  

10.  .     .... 

Total 

29.22 

25.05 

4.17 

0.24 
6.25 


6.26 
6.27 


6.28 


6*29 

6.30 


6,31 


Determine  90  per  cent  confidence  limits  for  the  true  difference  between 
the  expected  weights  of  the  two  varieties.  State  all  assumptions. 
ITwing  the  data  of  Problem  6.20,  obtain  95  per  cent  confidence  limits 
for  orj/crl. 

Using  the  data  of  Problem  6.21   and  ignoring  the  assumption  used 
there,  namely,  that  cr?«»cri?  obtain  99  per  cent  confidence  limits  for 


Using  the  data  of  Problem  6,22,  obtain  95  per  cent  confidence  limits 

for  <rs/<rT. 

Using  the  data  of  Problem  6.2,  determine  with  95  per  cent  confi 

dence:  (a)  95  per  cent  tolerance  limits  and  (b)  an  upper  99  per  cent 

tolerance  limit. 

Using  the  data  of  Problem  6,4,  determine  with  95  per  cent  confidence: 

(a)  76  per  cent  tolerance  limits  and  (b)  a  lower  90  per  cent  tolerance 

limit* 

Using  the  data  of  Problem  6.6,  determine,   with  95  per  cent  confi 

dence,  a  99  per  cent  upper  tolerance  limit  on  the  densities. 

Using  the  data  of  Problem  6*7,  determine,   with  95  per  cent  confi 

dence,  a  90  per  cent  lower  tolerance  limit  on  the  ranges, 

Consider  tho  following  definitions: 

(a)  If  the  expected  value  of  an  estimator  does  not  equal  the  true  value 
being  estimated,  tho  difference  between  tho  expected  value  and 
the  true  value  ia  known  aa  the  bias  of  the  estimator, 

(6)    If  an  estimator  has  0  bias,  it  IB  Baid  to  bo  accunde* 

(c)  If  an  estimator  has  a  small  bias,  it  is  Baid  to  bo  relatively  accurate. 

(d)  If  an  estimator  has  a  large  bias,  it  i&  said  to  bo  inaccurate* 

($)  The  precision  of  an  estimator  is  a  measure  of  the  repeatability  of 
the  estimator.  Therefore,  precision  may  be  expressed  in  terms  of 
the  variance  of  an  estimator,  with  a  large  variance  signifying  lack 
of  precision  and  a  small  variance  signifying  high  prtuttaton*  Obvi 
ously,  absolute  precision  implies  a  0  variance,  an  ideal  seldom  (if 


REFERENCES    AND    FURTHER    READING  105 

ever)  achieved.  (JNTOTE:  Sometimes  a  measure  of  precision  is 
referred  to  as  a  measure  of  reliability.  Because  the  word  "relia 
bility"  has  another  meaning  in  engineering,  this  is  unfortunate. 
However,  as  with  many  expressions,  the  phrase  is  now  a  part  of 
the  language  of  statistics  and  will  therefore  continue  to  be  used.) 

It  should  be  observed  that  an  estimator  may  be:  (1)  both  precise  and 
accurate,  (2)  neither  precise  nor  accurate,  (3)  precise  but  not  accu 
rate,  or  (4)  accurate  but  not  precise. 

(a)    Discuss  the  foregoing  concepts  and  definitions  relative  to  the  con 
tents  of  Section  6.1. 

(6)    Discuss  these  ideas  taking  cognizance  of  costs  and  other  economic 
and  physical  limitations  which  continually  plague  the  researcher, 
(c)    Discuss  the  accuracy  and  precision  of  the  various  estimators  that 
have  been  introduced  so  far  in  this  text. 

References  and  Further  Reading 

1.  Anderson,  R.  L.,  and  Bancroft,  T.  A.  Statistical  Theory  in  Research.  McGraw- 
Hill  Book  Company,  Inc.,  New  York,  1952. 

2.  Bowker,  A.  H.,  and  Lieberman,  G.  J.  Engineering  Statistics.  Prentice-Hall, 
Inc.,  Englewood  Cliffs,  N.J.,  1959. 

3.  Brownlee,  K.  A,  Statistical  Theory  and  Methodology  in  Science  and  Engineer 
ing.  John  Wiley  and  Sons,  Inc.,  New  York,  1960. 

4.  Calvert,  R.  L.  The  Determination  of  Confidence  Intervals  for  Probabilities  of 
Proper,     Dud,    and    Premature    Operation.    Sandia    Corporation    Technical 
Memorandum  SCTM  213-55-51,  Sandia  Corp.,  Albuquerque,  N,  Mex.,  Oct. 
17,  1955. 

5.  Clopper,  C.  J.,  and  Pearson,  E,  S,  The  use  of  confidence  or  fiducial  limits 
illustrated  in  the  case  of  the  binomial.  Biometrika,  26:  404—13,  1934. 

6.  Dixon,  W.  J.,  and  Massey,  F.  J.  Introduction  to  Statistical  Analysis.    Second 
Ed.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1957. 

7.  Fisher,    R.    A.    On   the   mathematical  foundations   of   theoretical   statistics. 
Philosophical  Transactions  of  the  Royal  Society.  Series  A,  Vol.  222,  1922. 

gp  _  Theory    of    statistical    estimation.    Proceedings     of    the     Cambridge 

Philosophical  Society,  Vol.  22,  1925. 
9.   Freund,  J.  E.  Modern  Elementary  Statistics.  Second  Ed.  Prentice-Hall,  Inc., 

Englewood  Cliffs,  N.J.,  1960. 
IQ    1  Livermore,  P.  E.,  and  Miller,  I.  Manual  of  Experimental  Statistics. 

Prentice-Hall,  Inc.,  Englewood  Cliffs,  N.J.,  1960. 

11.  Hald,  A.  Statistical  Theory  With  Engineering  Applications.  John  Wiley  and 
Sons,  Inc.,  New  York,  1952. 

12.  .  Statistical   Tables  and  Formulas.  John  Wiley  and  Sons,   Inc.,   New 

York,  1952. 

13.  Huntsberger,  D.  V-  Elements  of  Statistical  Inference.  Allyn  and  Bacon,  Inc., 
Boston,  1961. 

14.  Mood,    A.   M.   Introduction  to  the   Theory   of  Statistics.   McGraw-Hill   Book 
Company,  Inc.,  New  York,  1950. 

15.  Muench,  J.   O.   A    Confidence  Limit   Computer.  Sandia  Corporation   Mono 
graph  SCR-159,  Sandia  Corp.,  Albuquerque,  N.  Mex.,  April,  1960. 

16.  Murphy,  R.  B.  Non-parametric  tolerance  limits.  Ann.  Math.  Stat.,  19:581-89, 
Dec.,  1948. 

17.  Ostle,    B.    Some   remarks   on    the   problem   of   tolerance   limits.    Industrial 
Quality  Control,  13(No.  10):11~13,  April,  1957. 

18.  Owen,  D.  B.  Distribution-Free  Tolerance  Limits.  Sandia  Corporation  Techni 
cal  Memorandum  SCTM  66A-57-51,  Sandia  Corp.,  Albuquerque,  N.  Mex., 
June,  1957. 


106  CHAPTER    6,    STATISTICAL    INFERENCE:    ESTIMATION 

19.   .  Tables  of  Factors  for  One-Sided  Tolerance  Limits  for  a  Normal  Dis 
tribution.    Sandia    Corporation    Monograph    SCR-13,    Sandia    Corp.,    Albu 
querque,  N.  Mex.,  April,  1958. 

20.  Snedecor,  G.  W.  Statistical  Methods.  Fifth  Ed,  The  Iowa  State  University 
Press,  Ames,  1956. 

21.  Statistical    Research    Group,    Columbia    University.    Selected    Techniques    of 
Statistical  Analysis.    (Edited  by   C.   Eisenhart,    M.   W.   Hastay,   and   W.   A, 
Wallis.)  McGraw-Hill  Book  Company,  Inc.,  New  York,  1947. 

22.  Wadsworth,  G.  P.,  and  Bryan,  J.  G.  Introduction  to  Probability  and  Random 
Variables.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1960. 

23.  Weiswborg,    A.,    and    Boatty,    G.    H.    Tables    of   Tolerance- Limit    Factors  for 
Normal  Distributions.   Battelle  Memorial  Institute,   Columbus,   Ohio,    Dec., 
1959. 

24.  Wiika,  S.  S.  Statistical  prediction  with  special  reference  to  the  problem  of 
tolerance  limits.  Ann.  Math.  Stat.y  13:400-409,  1942, 


C  H  APTE  R    7 

STATISTICAL  INFERENCE: 
TESTING  HYPOTHESES 

7.1      GENERAL  CONSIDERATIONS 

A  HYPOTHESIS  is  defined  by  Webster  as  "a  tentative  theory  or  supposi 
tion  provisionally  adopted  to  explain  certain  facts  and  to  guide  in  the 
investigation  of  others."  A  statistical  hypothesis  is  a  statement  about 
a  statistical  population  and  usually  is  a  statement  about  the  values  of 
one  or  more  parameters  of  the  population.  For  example,  the  following 
could  be  taken  as  hypotheses:  (1)  the  probability  of  a  1  on  a  toss  of  a 
certain  die  is  f ,  (2)  the  mean  height  of  American  adult  males  is  5  feet 
8.4  inches,  (3)  the  mean  length  of  a  certain  brand  of  6-inch  rulers  is 
5.99  inches  and  the  standard  deviation  is  0.02  inch. 

It  is  frequently  desirable  to  test  the  validity  of  such  hypotheses.  In 
order  to  do  this,  an  experiment  is  conducted  and  the  hypothesis  is 
rejected  if  the  results  obtained  from  the  experiment  are  improbable 
under  this  hypothesis.  If  the  results  are  not  improbable,  the  hypothesis 
is  accepted.  For  example,  we  might  test  hypothesis  (1)  above  by  toss- 
ing  the  die  600  times.  Intuitively,  it  is  evident  that  if  600  1's  are  ob 
tained,  the  result  is  improbable  under  the  hypothesized  probability  of 
^,  and  the  hypothesis  should  be  rejected.  On  the  other  hand,  if  100  1's 
were  observed,  this  result  would  not  be  improbable  and  the  hypothesis 
would  undoubtedly  be  accepted.  When  results  such  as  these  are  ob 
tained,  intuition  (combined  with  common  sense)  is  sufficient  to  decide 
whether  to  accept  the  hypothesis.  However,  in  actual  practice,  experi- 
meiital  results  do  not  usually  lead"lu  au^li  obvi0'usi?5ilt5tn^iong;  hence 
the  leraedTlfo^^  It  shoulcTBeT  pointed  out 

that  although  we  accept  or  rejec^aT^ypothesIs^we^have  not  proved  or 
disproved  the  hypothesis. 

In  testing  hypotheses,  there  are  two  types  of  errors  which  can  be 
made.  These  are  called: 

Type  I  error — the  rejection  of  a  hypothesis  which  is  true. 

Type  II  error — the  acceptance  of  a  hypothesis  which  is  false. 
To  aid  the  reader  in  comprehending  the  nature  of  statistical  hypotheses, 
decisions,  and  the  various  types  of  error,  Table  7.1  has  been  found 
helpful. 

When  setting  up  an  experiment  to  test  a  hypothesis,  it  is  desirable 
to  minimize  the  probabilities  of  making  these  errors.  In  order  to  make 
it  easier  to  talk  about  these  errors  and  their  probabilities,  the  proba 
bility  of  making  a  Type  I  error  is  designated  as  a.  and  the  probability 
of  making  a  Type  II  error  is  designated  as  /S.  It  should  also  be  noted 

[1071 


108  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

that  100 a  (in  per  cent)  is  commonly  referred  to  as  the  significance  level. 
What  constitutes  suitably  small  values  of  a  and  £?  This  is  not  a  ques 
tion  which  can  be  answered  unequivocally  for  all  situations.  Obviously 
the  values  of  a  and  /3  should  depend  on  the  consequences  of  making 
Type  I  and  II  errors,  respectively.  For  example,  if  we  are  considering 
the  purchase  of  a  lot  of  batteries  (or  some  other  very  critical  item)  for 
use  in  weapons,  we  might  hypothesize  that  the  lot  is  of  satisfactory 
quality.  (Actually  we  should  state  this  hypothesis  in  more  precise 
terms.)  If  this  hypothesis  is  true  and  we  reject  it,  no  great  harm  has 
been  done  since  we  can  always  wait  for  the  next  lot  (assuming  that  \ve 
are  not  in  a  hurry).  Consequently  a.  can  be  relatively  large  (perhaps 
0.25  or  larger).  Oil  the  other  hand,  if  the  hypothesis  is  false  and  we 
accept  it,  the  result  may  be  a  large  number  of  dud  weapons.  Since  this 
is  very  undesirable,  0  should  be  quite  small,  (maybe  0.01  or  leas).  It 
should  be  pointed  out  that  the  supplier  might  feel  differently  about 
these  probabilities* 

TABLE  7.1-Definition  of  the  Types  of  Errors  Associated  With  Tests 

of  Hypotheses 


Decision 

True  Situation 

Hypothesis  is  true 

Hypothesis 

is  false 
error 

Accept  the  hypothesis  

No  error 
Type  I  error 

Type  II 
No  error 

Reject  the  hypothesis*  ,  

An  important  consideration  in  discussing  the  probabilities  of  Type 
II  errors  is  the  "degree  of  falseness"  of  a  false  hypothesis.  In  a  given 
experiment,  if  the  hypothesis  is  false  but  is  nearly  true  (such  as  hypoth 
esising  that  a  probability  is  J  when  actually  it  is  1.0001/2),  ft  could  be 
quite  large.  However,  if  the  hypothesis  is  grossly  false,  (such  as  hy 
pothesizing  that  a  probability  is  f  when  it  is  actually  1),  /9  should  be 
much  smaller.  For  a  given  experiment  testing  a  specific  hypothesis, 
the  value  of  1  —  /3  is  known  as  the  power  of  the  test.  Since  the  power 
depends  on  the  difference  between  the  value  of  the  parameter  specified 
by  the  hypothesis  and  the  actual  value  of  the  parameter  where  the 
latter  is  unknown>  1—/9  should  be  expressed  as  a  function  of  the  true 
parameter.  Such  a  function  is  known  as  a  power  function  and  is  ex 
pressed  as  1  — /8(0)  where  6  represents  the  true  parameter  value.  The 
complementary  function,  /3(0),  is  known  as  the  operating  characteristic 
(OC)  function. 

Before  proceeding  further  with  the  details  of  testing  hypotheses,  a 
few  more  remarks  of  a  general  nature  are  in  order-  It  is  good  practice 
not  only  to  state  the  hypothesis  to  be  tested  (denoted  by  //)  but  also 
to  state  the  alternative(s)  to  //  (denoted  by  A).  This  ia  not  only  good 
procedure;  it  also  aids  in  the  determination  of  the  regions  of  acceptance 


7.1        GENERAL    CONSIDERATIONS 


1O9 


and  rejection  when  considering  the  sample  space  of  all  possible  values  of 
the  test  statistic.  Incidentally,  the  rejection  region  is  frequently  referred 
to  as  the  critical  region.  Using  the  notation  of  this  paragraph,  it  is  seen 
that  <x  =  P  (reject  H\H  is  true)  and  @  =  P  (accept  H\  A  is  true). 

Example  7.1 

Consider  a  simple  hypothesis,  H:IJ,=/JLQJ  against  a  single  alternative* 
A:fjL=fjLij  where  we  are  dealing  with  a  normal  population  with  known 
variance,  cr2.  Let  the  decision  to  reject  H  (accept  A)  or  to  accept  H 
(reject  A)  be  based  on  a  single  observation  obtained  at  random  from  the 
population  under  examination.  If  the  random  observation  is  less  than 
C  (see  Fig.  7.1),  H  will  be  accepted;  if  the  random,  observation  is  greater 
than  or  equal  to  (7,  H  will  be  rejected.  That  is,  X^C  constitutes  the 
rejection  or  critical  region.  The  probabilities  a.  and  /3  are  represented  by 
the  shaded  and  cross-hatched  areas,  respectively.  Clearly,  besides  de 
pending  on  the  choice  of  C,  a  depends  on  the  hypothesis  under  test 
(frequently  called  the  null  hypothesis)  while  /3  depends  both  on  the  null 
hypothesis  and  on  the  alternative  hypothesis. 


DISTRIBUTION 
ASSUMING  I-T 
IS  TRUE 


DISTRIBUTION 
ASSUMING  A. 
IS  TRUE 


ACCEPT   J± 


REJECT    ±L- 


FIG.   7. 1  —  Graphical   illusfraHon   of  the  acceptance  and 
rejection    regions   in    Example   7.1. 

Example  7.2 

Modify  Example  7.1  to  the  following  extent:  Consider  H:JJL=JJLO 
versus  the  composite  alternative  A  :/z  >MO-  In  this  situation  a  is  the 
same  as  before  but  j3  is  now  better  denoted  by  /3(/i)  =P  (accept  H\JJL). 
Clearly  /3(/x)  changes  as  we  think  of  the  "alternative  distribution"  in 
Figure  7.1  taking  all  possible  positions  for  which  ni  >/xo-  Thus,  an  OC 
curve  similar  to  the  one  shown  in  Figure  7.2  is  generated. 

Example  7.3 

Consider  a  further  modification  of  Example  7.1,  namely,  £T:/*  =  Mo 
versus  the  alternative  Arpt^Ma.  The  acceptance  and  rejection  regions 
might  be  as  shown  in  Figure  7.3,  namely,  reject  if  3C<Ci=Mo  — -ka  or 
if  -Xr>C2==Mo  +  &<r  and  accept  if  Ci  <X  <C2.  Only  the  distribution  of 
the  test  statistic  under  H  is  shown.  The  distribution  under  A  may  be 
visualized  if  the  reader  thinks  of  sliding  the  distribution  shown  to  the 
left  and  to  the  right.  For  this  situation,  an  OC  curve  similar  to  the  one 
in  Figure  7.4  would  result. 


FIG.  7.2— Type  of  OC  curve  to  be  expected   in   situations 
similar  to  Example  7.2, 


DISTRIBUTION 
ASSUMING  Ji 
IS  TRUE 


Ci 

P*  tirr*T    i-i-*--* 

/So 

Arv^fTPT  w              .... 

ca 

.    .,         .»_-*..    ocr.ipr'T    w  ..„.    *— 

X 


FIG.   7,3— Graphical  illustration   of  the  acceptance 
and  rejection  regions  in  Example  7,3, 


FIG.  7-4— Type  of  OC  curve  to  be  expected  in  situations 
similar  to  Example  7.3* 


T.2       ESTABLISHMENT    OF    TEST    PROCEDURES  111 

7.2      ESTABLISHMENT  OF  TEST   PROCEDURES 

When  establishing  a  test  procedure  to  investigate,  statistically,  the 
credibility  of  a  stated  hypothesis,  there  are  several  factors  that  must 
be  considered.  Assuming  a  clear  statement  of  the  problem  has  been 
formulated  and  that  an  associated  hypothesis  has  been  stated  in  mathe 
matical  terms,  these  are : 

(1)  The  nature  of  the  experiment  that  will  produce  the  data  must 
be  defined. 

(2)  The  test  statistic  must  be  selected.  That  is,  the  method  of  ana 
lyzing  the  data  should  be  specified. 

(3)  The  nature  of  the  critical  region  must  be  established. 

(4)  The  size  of  the  critical  region  (that  is,  ot)  must  be  chosen. 

(5)  A  value  should  be  assigned  to  0(8)  for  at  least  one  value  of  0 
other  than  the  value  of  8  specified  by  H.  This  is  equivalent  to 
stating  what  difference  between  the  hypothesized  value  of  the 
parameter  and  the  true  value  of  the  parameter  must  be  de 
tectable,  and  with  what  probability  we  must  be  confident  of 
detecting  it. 

(6)  The  size  of  the  sample  (i.e.,  the  number  of  times  the  experi 
ment  will  be  performed)  must  be  determined. 

It  should  be  clear  that  these  steps  will  not  always  be  taken  in  the  order 
listed.  Not  all  of  the  steps  are  independent,  and  frequently  it  is  neces 
sary  to  reconsider  (several  times)  the  various  steps  until  a  reasonable 
test  procedure  is  formulated.  More  will  be  said  on  this  subject  later. 
For  now,  some  explanatory  examples  will  probably  be  of  more  value 
than  additional  generalizations. 


Example  7.4 

With  respect  to  a  specific  coin  we  have  H:P  (heads)  =p  =  0.5  and 
A  :p5*=0,5.  The  experiment  will  consist  of  tossing  the  coin  some  number 
of  times,  counting  the  number  of  times  heads  occurs,  and  rejection  of 
//  will  take  place  if  either  a  very  small  or  very  large  proportion  of  heads 
are  observed.  Let  a.  =  0.05.  Ignore  /?(£>)  for  the  moment.  Consider  n  =  5 
and  the  rejection  region  to  consist  of  either  no  heads  or  all  heads.  Then, 
P  (rejection | p=0. 5)  =tSr  =  0.0625.  Since  this  is  greater  than  ct  =  0.05, 
a  larger  number  of  tosses  is  required.  Let  us  try  n  =  6,  keeping  the  same 
rejection  region.  Now  we  have  P  (rejection  |p  =  0.5)  =  0.03125  which  is 
less  than  a.  =  0.05.  Thus  an  acceptable  test  procedure  has  been  devel 
oped.  (NOTE:  The  probabilities  of  rejection  given  were,  of  course,  cal 
culated  using  f(x)  =  C(n, 


Example  7.5 

In  Example  7.4  we  derived  the  test:  "Toss  the  coin  six  times  and 
reject  //  if  either  zero  or  six  heads  occurs;  otherwise,  accept  H."  Clearly, 
other  rejection  regions  might  have  been  chosen  together  with  different 


112  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

values  of  n,  as  long  as  P  (rejection  \p  =  0.5)  <  ex.  What  we  found  in 
Example  7.4  was  the  smallest  value  of  n  for  the  specified  rejection 
region.  The  reader  should  investigate  some  of  these  other  possibilities. 

Example  7.6 

Now  consider  @(p)  for  the  test  derived  in  Example  7.4.  For  selected 
values  of  p  the  approximate  values  of  /3(p)  =  l—  p6  —  (1—  p)6  are  given 
in  Table  7.2.  (NOTE:  Only  approximate  values  are  given  because  the 
exact  answers  involve  an  unnecessary  number  of  decimal  places.  For 
example,  for  p  =  0.5, 

1  -P  (rejection  |  p  =  0.5)  =  1  -0.03125  =  0.96875^0.97.) 


If  one  did  not  consider  the  derived  test  to  be  discriminating  enough  (as 
evidenced  by  the  OC  curve),  the  discriminatory  power  could  be  in 
creased  by:  (1)  changing  the  sample  size  and  the  definition  of  the  crit 
ical  region  or  (2)  concocting  an  entirely  different  test  procedure  and  test 
statistic.  It  is  clear  that  we  are  faced  with  just  this  situation  in  the  pres 
ent  case.  The  test  derived  in  Example  7.4  is  good  for  detecting  two- 
headed  or  two-tailed  coins  (nearly  as  good  as  looking  at  both  sides  of 
the  coin)  but  is  poor  for  detecting  slightly,  or  even  moderately,  biased 
coins.  Thus  a  modified  or  new  test  is  required. 

TABLE  7.2-Selected  Values  of  the  OC  Function  for  Example  7.6 


P 

Approximate  Values  of 

£(p) 

0 

0 

0.1 

0.47 

0.2 

0.74 

0.3 

0.88 

0.4 

0.95 

0.5 

0.97 

0,6 

0.95 

0.7 

0.88 

0.8 

0.74 

0.9 

0.47 

1,0 

0 

Example  7.7 

Consider  the  following  modification  of  Example  7.4,  namely, 
H  :p>0.5  and  A  :p  <0.5.  The  experiment  will  remain  the  same  but  the 
regions  of  acceptance  and  rejection  will  change*  Obviously,  the  occur 
rence  of  many  heads  does  not  tend  to  deny  //,  so  the  rejection  region 
will  be  only  that  region  in  which  few  heads  occur.  That  is,  a  one-tailed 
test  (like  a  one-sided  confidence  limit)  is  required.  Proceeding  as  before, 
it  is  found  that  a  possible  test  is:  "Toss  the  coxa  five  times.  If  no  heads 
occur,  reject  //;  otherwise,  accept."  This  gives  <**» 0.031 25. 


7.4       NORMAL    POPULATION;    H^^po   VERSUS   Ai}*,  > 

7.3      NORMAL  POPULATION;  H:M~Mo  VERSUS 

Suppose  that  we  wish  to  know  if  a  random  sample  could  be  from  a 
normal  population  with  mean  ^o-  More  specifically,  assuming  normal 
ity,  the  hypothesis  ^J:M  =  Mo  wiu  be  tested  relative  to  the  alternative 
A  IM^/XQ.  For  a  chosen  a9  the  procedure  is  to  compute 


t  =  (X  -  MO)  As  =  Vn(X  -  MO)  A  (7.1) 

and  reject  H  if  t<  —  «a-.«/2>cw-i)  or  if  <>«a--«/2Xn--i>;  otherwise,  accept  H. 
Example  7.8 

A  metallurgist  made  four  determinations  of  the  melting  point  of  man 
ganese:  1269°,  1271°,  1263°,  and  1265°C.  Are  these  in  accord  with  a 
hypothesized  value  of  1260°C?  Here  the  hypothesis  is  H:M«1260,  the 
alternative  is  A  :ju^l260,  and 

/  =  ^  """  Mo  =  1267  —  1260  __ 

Jjf       "~          1.862          ~~  3"83 

is  computed.  Since  £0.375(3)  ==3.182  (a  5  per  cent  significance  level  is 
assumed),  H  is  rejected  and  it  is  concluded  that  the  hypothesized  value 
is  incorrect.  By  using  a  5  per  cent  significance  level,  it  is  recognized  that 
the  probability  of  Type  I  error  will  be  no  greater  than  0.05.  That  is, 
there  is  a  maximum  risk  of  5  per  cent  in  rejecting  the  hypothesis  that 
M  —  1260  if  the  hypothesis  is  really  true. 

It  would  also  be  of  interest  to  examine  the  OC  curve  for  the  test  pro 
posed  above.  However,  it  would  be  necessary  to  prepare  OC  curves  for 
many  values  of  <x  and  n.  Also,  the  formula  for  $  associated  with 
"Student's"  fr-test  involves  the  noncentral  it-distribution  which  must 
be  considered  beyond  the  scope  of  this  text.  Thus,  the  reader  is  referred 
to  examples  of  such  curves  given  in  Bowker  and  Lieberman  (3). 

7.4      NORMAL   POPULATION;  JET:,z</*o  VERSUS  A:M>/x0,  OR 

H:M>A*O  VERSUS  A:jm<fjL* 

A  more  common  situation  than  that  considered  in  Section  7.3  is  the 
case  of  a  one-sided  alternative.  For  example,  a  manufacturer  produces 
wire  cable  which  must  have  a  breaking  strength  not  less  than  1500 
pounds.  A  new  and  cheaper  process  for  making  the  cable  is  discovered 
and  he  wishes  to  change  to  the  new  process,  provided  that  cable  so  pro 
duced  will  have  a  mean  breaking  strength  greater  than  1500  pounds. 
Thus  he  could  formulate  the  hypothesis  Hi  !*,<}**  =  1500  pounds  as 
opposed  to  the  alternative  ,A:/x>1500  pounds.  The  hypothesis  H 
would  be  rejected  if  a  sample  of  the  new  cable  presented  sufficient  evi 
dence  that  M  actually  exceeds  1500  pounds. 

In  general  terms,  when  testing  the  hypothesis  H:^<^G  versus 
the  procedure  is  to  calculate 


114  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

*  =   (3s  -  MO)  A*  =  Vn(X  -  MO)  A  (7.2) 


and  reject  H  if  2><(i—  «xn—i);  otherwise,  accept  H. 

If  the  hypothesis  J£T:M>MQ  versus  A:M<MO  is  under  investigation, 
the  test  statistic  is  calculated  as  in  Equation  (7.2),  but  now  H  is 
rejected  only  if  t<  —  £(!_«)  CTI-_I). 

Example  7.9 

A  manufacturer  of  television  sots  purchases  tubes  from  one  of  the 
few  large  suppliers  of  such  specialized  material.  He  will  not  purchase 
tubes,  however,  unless  it  can  be  demonstrated  that  the  average  length 
of  life  will  exceed  500  hours.  A  random  sample  of  9  tubes  is  subjected  to 
a  "life  test"  and  the  following  values  are  obtained  :  T  =*  600  and  s2  —  2500. 
It  is  assumed  that  the  "lengths  of  life"  (measured  in  hours)  are  normally 
distributed.  Shall  the  hypothesis  //:^<500  be  accepted?  For  this  ex 
ample,  *=  (GOO  —  500)/16.67  =  6,00  far  exceeds  £0.95(8)  =1.860,  and  the 
null  hypothesis  is  rejected.  As  can  be  seen,  a  5  per  cent  significance  level 
was  used.  This  means  that  the  maximum  risk  of  rejecting  //:/z<500 
when  //  is  really  true  is  5  per  cent.  Therefore,  the  manufacturer  of  tele 
vision  sets  will  undoubtedly  purchase  tubes  from  this  supplier. 

The  reader  is  again  referred  to  Bowker  and  Lieberman  (3)  for  sample 
OC  curves  related  to  these  te«t  procedures, 


7,5      NORMAL   POPULATION;   H:<r2=*=<rg   VERSUS 

Suppose  that  we  have  a  sample  of  sixe  n  drawn  randomly  from  a 
normal  population  and  some  predetermined  value  of  the  variance  is  to 
be  substantiated  or  refuted;  i.e.,  we  wish  to  test  the  hypothesis, 
//:<r2  =  <r§  as  opposed  to  the  alternative  A  :<r2  ^<TQ.  If  a  probability,  <x, 
of  making  a  Type  I  error  has  been  chosen,  i.e.,  a  significance  level  of 
100  <*  per  cent  has  been  selected,  the  hypothesis  //  will  be  accepted  if 

xV/.Xn-l)    <    Z)    (X    -    DVorJ    <    XV^C^'  <7-3) 

Otherwise,  //  will  be  rejected, 

Example  7*10 

Consider  the  data  given  in  Example  7.8*  Do  these  values  mipport  the 
hypothesis  that,  if  repeated  measurements  arc  assumed  to  be  normally 
distributed,  the  true  variance  of  all  such  measurements  IB  equal  to  2? 
Here  the  hypothesis  in  //:<72»2  and  the  alternative  IB  A  :cr2p^2.  It 
IB  determined  that  S(-V  —  3T)*M-20.  Since  xS.onw>  -0.216  and 
Xo.««»>  ""9-#5»  we  Bee.  that  //  is  rejected.  You  will  note  that  once  again  a 
probability  of  Type  I  error  equal  to  0.05  was  ehcmen.  That  is,  we  have 
run  a  maximum  risk  of  5  per  cent  of  rejecting  a  true  hypothesis, 

Jf  wo  winh  to  determine  the  CO  csurve  for  this  teat,  the  required 
values  can  bo  ealcsulated  using 

»     »  i     <  X2  <   (^o/o-^X 


7.7       BINOMIAL    POPULATION;    H:p  =  p0   VERSUS    A:p  ^  p0 


115 


The  reader  is  referred  to  Bowker  and  Lieberman  (3)  for  examples  of 
OC  curves  associated  with  this  test. 


POPULATION; 
VERSUS   A:o 


H:a*<<r0  VERSUS  A:<r*>o%, 


7.6      NORMAL 

or  H*r*>o 

It  is  usually  more  realistic  to  consider  the  hypothesis  that  the  popu 
lation  variance  is  less  than  or  equal  to  some  particular  value  than  to 
consider  the  hypothesis  that  it  equals  some  value.  This  is  so  because, 
in  general,  a  small  variance  is  considered  to  be  desirable.  In  such  a  case, 
the  hypothesis  H:<T*<CTQ  is  formulated  as  opposed  to  A:cr2>cr|.  The 
hypothesis  H  will  be  rejected  only  if  %2=  XX-3T  —  3f)  2/cro>x2(i-<*)(n_i)- 

Should  jfif:cr2>cr§  (as  opposed  to  A:<r2<j7p)  be  under  investigation, 
the  rejection  region  would  be  x2==  T^  (X  —  3T)  VcrS  <  y£  ,«._,  }  . 

Sample  OC  curves  may  be  observed  in  Bowker  and  Lieberman  (3) 
for  a:  =  0.05  and  <x  =  0.01. 

Example  7.11 

Consider  the  data  of  Table  7.3  which  were  obtained  from  a  random 
sample  of  80  bearings.  To  test  (using  a.  =  0.05)  H  :  cr2  <  0.00005  versus 
A  :  or2  >  0.00005,  we  calculate 

X2  =  ]T  *2/CK00005  ^  0.000474/0.00005  =  9.48. 
Since  this  does  not  exceed  xJ96C79)  =  100.7,  we  are  unable  to  reject  H. 


TABLE   7.3-Number  of  Bearings  Observed  With  the  Indicated  Diameters 


Diameter  (Inches) 

Number 

3.573 

4 

3.574 

2 

3.575 

9 

3.576 

12 

3.577 

12 

3.578 

10 

3.579 

9 

3.580 

9 

3.581 

8 

3.582 

4 

3.583 

1 

Total 

80 

7.7      BINOMIAL  POPULATION;  H:p=pQ  VERSUS 

The  coin-tossing  experiment  discussed  in  Example  7.4  illustrates  the 
type  of  problem  to  be  considered  in  this  section.  However,  in  practice, 
one  is  usually  given  the  sample  size  and  asked  to  determine  the  rejec 
tion  region,  rather  than  (as  in  Example  7.4)  being  asked  to  find  the 


116  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

smallest  sample  size  that  is  consistent  with  a  specified  rejection  region. 
For  example,  for  a  fixed  sample  size,  n,  the  acceptance  and  rejection 
regions  are  determined  by  solving 


C(»,  *)po(l  -  Po)"""  =  «/2  (7.4) 

x«0 

and 

C(w,  *)pS(l  -  po)-""  =  a/2  (7.5) 


for  Z/  and  Z7.  The  acceptance  region  defined  by  these  two  equations  is 
the  set  of  positive  integers  between,  but  not  including,  L  and  C/. 

Unfortunately,  it  is  usually  impossible  to  find  integral  values  of  L 
and  U  to  satisfy  Equations  (7.4)  and  (7,5).  Therefore,  it  is  customary 
to  choose  those  values  of  L  and  U  which  make  the  value  of  each  of  the 
summations  as  large  as  possible  without  exceeding  a/2.  Occasionally, 
the  restriction  of  being  less  than  or  equal  to  a/2  will  be  relaxed  if,  by 
so  doing,  the  probability  of  rejecting  a  true  hypothesis  will  be  only 
slightly  larger  than  the  chosen  a. 

Example  7.12 

In  a  certain  cross  of  two  varieties  of  peas,  genetic  theory  led  the  in 
vestigator  to  expect  one-half  of  the  seecLs  produced  to  be  wrinkled  and 
the  remaining  one-half  to  be  smooth.  Taking  of«0.01  and  n««4()l  deter 
mine  L  and  C7,  and  thus  define  the  acceptance  and  rejection  regions, 
Using  Equations  (7,4)  and  (7,5)  with  po»«0.5y  we  obtain  L**l\  and 
f/aa29.  Therefore,  the  acceptance  region  consists  of  those  values  of  x 
for  which  11  <x  <29. 


Without  adequate  tables,  the  procedure  discussed  BO  far  in  this  sec 
tion  i«  not  very  palatable  to  the  researcher.  Consequently,  some  ap 
proximate  procedures  which  lend  themselves*  to  easy  calculation  will 
bo  investigated. 

In  Section  5.11  the  normal  distribution  wan  HUggented  as  a  posnible 
approximation  to  the  binomial  distribution.  If  such  an  approximation 
is  used,  Equations  (7,4)  and  (7.5)  are  replaced  by 


ssss.-^  >    «  a/2  (7.6) 

—  PO)    * 

and 

(  (U  —  0.5)  —  npo) 

P<Z  >   • — •==s^--==r=™.>    =  «/2  C7.7) 

^      ~        '       "        po)  f  /  ^        ; 


where  Z  is  a  standard  normal  variate-  These  equations  may  then  be 
solved  for  L  and  t/* 


7.7      BINOMIAL    POPULATION;    H:p  =  p0  VERSUS    A:p  =7^  p0  117 

Example  7.13 

Using  the  normal  approximation,   find  L  and    17  for  the  situation 
described  in  Example  7.12.  Since  a.  =  0.01,  we  see  that 

{  (L  +  0.5)  -  40(0.5)  }A/40(0.5) (0.5)  -  -  2.575 
and 


{(£/  -  0.5)  —  40  (0.5)}  A/40  (0.5)(0.5)  =  2.575. 
Thus,  L  =  11.  4  ^11  and  E7  =  28.6^29. 

Rather  than  proceed  as  indicated  in  the  preceding  paragraph,  it  is 
common  to  take  the  number  of  events  (x)  occurring  in  the  class  associ 
ated  with  p  and  calculate 

(oc  +  0.5)  —  np0  ..      . 

Z  =         —    J          "    ,          torx<npQ  (7.8) 

V^po(l  —  p)o 

or 

(x  -  0.5)  -  npQ 
Z  =  -  -  «  for  oc  >  npQ.  (7.9) 


Then,  the  hypothesis  H:p  =  pQ  will  be  rejected  if  Z^Zo./z  or  if  Z>2i_a/2 
where  za  and  ^i_«/2  are  found  in  Appendix  3. 

Example  7.14 

Consider  the  situation  described  in  Example  7.12.  A  random  sample 
of  40  seeds  segregated  into  30  wrinkled  and  10  smooth.  Using  Equation 
(7.9),  _ 

Z  =  {(30  -  0.05)  -  40(0.5)  }/V40(0.5)  (0.5)  ^3  >  2.995  =  2.58. 

Therefore,  H:p=*Q.5  is  rejected. 

Another  useful  approximation  is  available  because  the  square  of  a 
standard  normal  variate  is  distributed  as  chi-square  with  one  degree  of 
freedom  (see  Section  5.18).  When  the  chi-square  approximation  is  used, 
the  test  statistic  is 


t—  1 
where 

0i  =  x 

0%  =  n  —  x 


The  hypothesis  H:p  =  p<>  will  be  rejected  if  x2>x?_*a:);  otherwise, 


118  CHAPTER    7,    STATISTICAL    INFERENCE;    TESTING    HYPOTHESES 

will  be  accepted.  It  should  be  clear  that  O  stands  for  observed  and  that 
E  stands  for  expected  in  Equation  (7.10). 

Example  7.15 

It  will  be  instructive  to  rework  Example  7.14  using  this  method. 
Thus, 

xs  «  (|  30  _  20  [    -  0.5)  a/20  +(  |  10  -  20  |    -  0.5)  2/20  =  9.025  >  X299(1)  «  6.63. 

Therefore,  as  in  the  preceding  example,  //:p==0.5  is  rejected.  (NOTE: 
X2  =  9.025  ==J£2^(3)*.) 


7.8      BINOMIAL.    POPULATION;    H~p<pQ    VERSUS 
OR  H:p>pQ  VERSUS  Aip<p* 

In  many  practical  situations,  the  hypothesis  H:p<p^  is  more  ap 
propriate  than  //:  p~po*  An  example  of  this  would  be  any  hypothesis 
concerning  the  per  cent  of  defective  items  in  a  production  lot.  When 
dealing  with  this  type  of  problem,,  the  researcher  may  use  only  the 
exact  procedure  or  the  normal  approximation,  The  chi-squarc  approxi 
mation  may  not  be  used  because  it  effectively  adds  together  the  areas 
under  both  tails  of  the  standard  normal  curve  when  only  a  one-tailed 
test  is  appropriate. 

Only  the  case  If:p^po  versus  yl:p>po  will  be  discussed  in  detail. 
The  discussion  for  the  case  H:p*>pQ  versus  A  :p<.p^  would  proceed  in 
a  similar  fashion,  the  only  change  being  which  tail  of  the  distribution 
is  used  for  the  rejection  region.  The  value  of  U  which  defines  the  ac 
ceptance  and  rejection  regions  is  determined  by  solving 

C(n,  *)po(l  -  po)w—  ==  <*  (7,11) 


for  U,  As  before,  it  will  be  necessary  to  settle  for  that  value  of  U  such 
that  the  value  of  the  summation  closely  approximates  a.  The  rejection 
region,  then,  consists  of  the  positive  integers  greater  than  or  equal  to 
U*  If  the  normal  approximation,  is  used,  calculate 

Z  «  {  O  -  0,5)  -  npoJ/VnpiC^Vo)-  (7-12) 


Example  7*16 

Prom  past  experience  it  has  been  determined  that  a  qualified  operator 
on  a  certain  machine  turning  out  400  items  per  day  produces  20  or 
fewer  defective  items  per  day,  A  new  operator  in  hired  to  run  the  same 
machine  and  the  hypothesis  is  made  that  he  IB  a  qualified  operator, 
Taking  ^»M).03,  determine  f/,  and  thus  define  the  acceptance  and 
rejection  regions.  Here,  the  hypothesis  is  //  :p  <0.05  and  n**400,  Using 
Kquatkm  (7*11),  wo  find  that  t/«29.  Thus,  if  the  new  operator  pro 
duced  more  than  28  defective  items  in  a  run  of  400,  we  would  reject  the 
hypothesis  that  ho  is  a  qualified  operator* 

Example  7.17 

Using    the    normal   approximation,    test   the    hypothesis    //:p;<0*05 


T.9       TWO    NORMAL    POPULATIONS;    H:M*  =  J-t*    VERSUS    At^  ^  M*  119 

versus   A:p>Q.Q5,   given   that   £=  32/400  =  0.08.    Let   <x  =  0.03.    Using 
Equation  (7.12), 

Z  =  {(32  -  0.5)  -  400(0.05)  }  A/400(0.05)  (0.95)  s*  2.6  >  Zt97  =  1.89. 
Thus,  #:?  <0.05  is  rejected. 

7.9     TWO  NORMAL  POPULATIONS;  H:^^^  VERSUS 


The  methods  to  be  described  here  are  closely  allied  with  those  dis 
cussed  when  obtaining  confidence  limits  for  MI  —  ^2.  Consequently,  it  is 
recommended  that  the  reader  review  the  earlier  material.  Without 
further  preamble,  we  shall  present  and  illustrate  the  appropriate  pro 
cedures. 

Case  I:  <rf  =  of 

In  this  case  the  procedure  is  to  calculate 

/  -  (3*1  -  X2)A^_^2  (7.13) 

where 


and 

x  +  w2  ~  2) 
-  2),  (7.15) 


and  to  reject  H:^^^  if  *<  —  «a«a/2)cn1H-n2~2)   or  if  _a 

Clearly,  some  simplification  of  the  formulas  will  occur  if  ni  =  n2. 

Example  7.18 

Wire  cable  is  being  manufactured  by  two  processes.  We  wish  to 
determine  if  the  processes  are  having  different  effects  on  the  mean 
breaking  strength  of  the  cable.  Laboratory  tests  were  performed  by 
putting  samples  of  cable  under  tension  and  recording  the  load  required 
to  break  the  cable.  Using  the  data  given  in  Table  7.4,  and  letting 
CK  =  0,05,  test  the  hypothesis  H:/jLi=fj,%  versus  A  :^ix^2.  Calculations 
yield 

!Fi  «  8.17,         ^2  «  11.29,          s2  =  5.29,     and     t  =  —  2.44. 
Since  <.  975(11)  =  2.201,  the  hypothesis  H  is  rejected. 

Example  7.19 

Two  rations  (feeds)  are  to  be  compared  with  respect  to  their  effect 
on  the  weight  gains  of  hogs.  Ten  animals  are  available,  and  five  are  fed 
feed  No.  1  while  the  other  five  are  fed  feed  No.  2.  Using  <2  =  0.10;  test 
the  hypothesis  that  the  two  feeds  are  equally  effective  in  causing  hogs 
to  gain  weight.  The  data  obtained  are  given  in  Table  7.5.  Calculations 
yield  t=  —1.58.  Since  £.95(8)  =  1.860,  we  are  unable  to  reject  -ff:/zj=£rj. 


120 


CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

TABLE  7.4-Critical  Values  of  the  Load  (Coded  Data) 


Process  No.  1 

Process  No.  2 

9 

14 

4 

9 

10 

13 

7 

12 

9 

13 

10 

8 

10 

TABLE  7.5-Gains  in  Weight  (in  Lbs.) 

Feed  No.  1 

Feed  No.  2 

1 

4 

2 

3 

4 

9 

5 

10 

8 

9 

Case  II:  <rf 

When  this  situation  prevails,  that  is,  when  we  are  unwilling  to  as 
sume  that  erf  equals  a\,  a  reasonably  good  approximate  procedure  is 
as  follows.  Compute 


*'  -  (x,  - 


+ 


and  reject  if 
or  if 
where 


tf 


(7.17) 
(7.18) 


Example  7.20 

As  an  illustration,  Example  7  AS  will  be  reworked  on  the  assumption 
that  of  does  not  equal  cr|.  Thus,  Xi^S.17,  ^«11,29,  $f«=5.4, 
sl*«6.2,  w>lS«0.9,  toa«0.74,  *'-—  2.4,  <J«2.571,  ^a««  2.447,  and  the 
weighted  average  of  ti  and  ts  is  2.52.  Conclusion:  accept  H* 


7.9       TWO    NORMAL    POPULATIONS; 

Case  III:  Paired  Observations 

The  procedure  in  this  case  is  to  calculate 

t  = 


=  p*   VERSUS    A:JJ* 


121 


j  (7-19) 

and  to  reject  H:fj.i  =  v>2  (or  Hfifj,D  =  AH  — M2  =  0)  if  t< — £<!-«/ 2)  c«~i>  or  if 
£>£(!-«/ 2)<n-i).  Here,  of  course,  n  is  the  number  of  pairs  of  observa 
tions.  Or,  in  other  words,  n  is  the  number  of  differences,  D~X—Y. 
Also,  it  should  be  clear  that  £>  =  X—  F. 

Example  7.21 

In  a  Brinell  hardness  test,  a  hardened  steel  ball  is  pressed  into  the 
material  being  tested  under  a  standard  load.  The  diameter  of  the  spher 
ical  indentation  is  then  measured.  Two  steel  balls  are  available  (one 
from  each  of  two  manufacturers)  and  their  performance  will  be  com 
pared  on  15  pieces  of  material.  Each  piece  of  material  will  be  tested 
twice,  once  with_each  ball.  The  data  obtained  are  given  in  Table  7.6.  Cal 
culations  yield  Z>  =  8,  s2^121.6,  and  £^2.81.  Using  c*  =  0.05,  it  is  seen 
that  2.81  >t. 975CU)  =* 2. 145,  and  thus  we  reject  the  hypothesis  that  the 
two  steel  balls  give  the  same  average  hardness  indication. 

TABLE   7.6-Data  Obtained  in  a  Brinell  Hardness  Test 


Sample  No. 

Diameters 

D  =  X—  Y 

X 

F 

1  

73 

43 
47 
53 
58 
47 
52 
38 
61 
56 
56 
34 
55 
65 
75 

51 
41 
43 
41 
47 
32 
24 
43 
53 
52 
57 
44 
57 
40 
68 

22 
2 
4 
12 
11 
15 
28 
—  5 
8 
4 
—  1 

—  10 
2 

25 

7 

2  

3  

4  

5  

6  

7  

8  

9  

10  

11  

12  

13  

14  

15  

As  has  been  stated  before,  the  OC  curves  associated  with  tests  of 
significance  must  be  examined  if  one  is  to  be  certain  that  the  suggested 
test  procedure  is  discriminating  enough.  Once  again  we  shall  beg  the 
question,  and  refer  the  reader  to  Bowker  and  Lieberman  (3)  for  samples 
of  such  curves. 


172 


CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


7.10     TWO 


NORMAL 

OR  jHT:i 


POPULATIONS; 
VERSUS  A:/J 


<MZ     VERSUS 


By  now,  the  technique  for  one-tailed  test  procedures  should  be  clear. 
Consequently,  only  a  brief  discussion  will  be  given.  With  reference  to 
the  three  cases  discussed  in  the  preceding  section,  the  same  test  sta 
tistic  will  be  calculated  here  as  was  calculated  there.  The  only  differ 
ence  will  be  in  the  selection  of  the  critical  values  of  t  from  the  table  in 
Appendix  5.  As  in  other  examples  of  one-tailed  tests,  the  values  will 
be  chosen  so  that  all  of  a.  (rather  than  a/2)  will  be  at  one  end  of  the 
distribution.  Some  OC  curves  are  again  available  in  Bowkcr  and 
Liebcrman  (3).  The  tests  are  summarized  in  Table  7.7. 

TABLE  7.7-One-Sided  Test  Procedures  for  Comparing  the  Means  of  Two 

Normal  Populations 


Hy- 

po  thesis 

Assumption 

Statis 
tic 

Equa 
tion 

Rejection  Region 

Mi  ^Ma 

Cl-al 

/ 

7  A3 

t  >*<>-<*)  (m-Hn^-a) 

Mi  <M2 

CT?  7*01 

t' 

7.16 

t'  >  weighted  average  using 

100(1  —a)  per  cent  points 

M/><0 

paired  observations 

t 

7,19 

£>2<l.~-  <*)<n~  1) 

Mi  >M'2 

a?~c-8 

I 

7.13 

/<  —  J<l—orX»H  nj-.*) 

Mi  >Ma 

a*  *d 

t' 

7.16 

£'<the     negative     of     the 

weighted    average    using 

100(1  —a)  per  cent  points 

Ml>>0 

paired  observations 

t 

7.19 

£<  —~£(l~~«)(n^  l> 

Example  7.22 

Two  pieces  of  moat,  one  a  control  and  the  other  treated  to  tenderize 
the.  fibers,  are  to  be  tested.  Tenderness  will  be  measured  by  the  force 
needed  to  shear  samples  of  meat.  (Lower  shear  force  values  indicate 
more  tender  moat.)  trivon  the  data  in  Table  7,8,  and  letting  o:^ 0.025, 
test  the  hypothesis  7/:jur>^  versus  A : ^Tj<fjLa,  ^O^l^nltitlonB  yield 


TABLK  7.8-Shear  Force  Values  for  Tenderness  Test 


Control 
50 


44 
24 
50 
41 
43 


Treated 

46 
40 
32 
23 
54 
51 


7.12       TWO    NORMAL    POPULATIONS;    H:0f  ^  a*    VERSUS    A-.o\>ol  123 

Since  t. 975(12)  =2.179,  we  reject  H  and  conclude  that  the  treatment  does 
improve  the  tenderness  of  the  meat. 

Further  examples  could  be  given.  However,  rather  than  take  up 
space  for  such  a  purpose,  we  will  rely  on  problems  to  illustrate  the 
other  cases. 

7.11      TWO      NORMAL      POPULATIONS;     flr:of=of     VERSUS 


As  in  Section  6.10,  the  F-ratio  will  be  the  appropriate  statistic.  That 
is,  the  procedure  will  be  to  calculate 

F  =  s\/  si  (7.20) 

)n^^  or  if  F>Fa 

larger  sample  variance 


and  reject  H  if  F<F^m(ni^)n^^  or  if  F>Fa_a/2><nr-i,nj^L>.  Alterna 
tively,  we  can  calculate 


(7,21) 
smaller  sample  variance 

and  reject  only  if  F  >  F  &—*/%)  ^iy  v$  where  z>i  and  z>2  represent,  respec 
tively,  the  degrees  of  freedom  associated  with  the  numerator  and  de 
nominator.  OC  curves  may  be  obtained  by  calculating 


/3    =    P{  (a-2/CTj),F(a/2)(ni—  I,n2—  1)    <    -P    <    (<^2/Vi)F  (1—  a/2)  (Wi—  I,n2—  1)  }  - 

Sample  OC  curves  are  given  in  Bowker  and  Lieberman  (3). 

Example  7.23 

Using  the  data  of  Example  6.4  and  letting  a:  =  0.05,  test  the  hypothesis 
J/:oi  =<r|  versus  A  10^7*0$.  It  is  seen  that  F  =  109.63/65.99  =  1.66 
with  jfi  —  40  and  v2  =  30  degrees  of  freedom.  Since  F.  975(40,30)  =2.01,  we 
are  unable  to  reject  H. 

7.12     TWO     NORMAL     POPULATIONS;     JHTiofrSof     VERSUS 
A:oi><ri,  OR  H:a%>02  VERSUS  A:oi<oi 

As  was  done  in  Section  7.10,  only  a  summary  of  the  test  procedures 
will  be  given.  This  appears  in  Table  7.9.  No  examples  will  be  given, 
but  some  of  the  problems  at  the  end  of  the  chapter  will  provide  an 
opportunity  to  apply  the  indicated  method.  As  in  other  sections,  solv 
ing  of  problems  is  strongly  recommended  as  an  aid  in  increasing  an 
understanding  of  the  various  methods. 

TABLE  7.9-One-Sided  Test  Procedures  for  Comparing 
The  Variances  of  Two  Normal  Populations 


Hypothesis 

Statistic 

Equation 

Rejection  Region 

3i3 

F 
F 

7.20 
7.20 

£l?r^ir" 

124  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

Reference  is  again  made  to  Bowker  and  Lieberman  (3)  for  those  who 
wish  to  examine  OC  curves  associated  with  the  tests  of  this  section. 

7.13      MULTINOMIAL   DATA 

Many  times,  our  sample  elements  may  be  assigned  to  any  one  of  sev 
eral  different  classes,  or  categories,  rather  than  simply  to  one  or  the 
other  of  two  classes  as  in  Section  7.7.  In  such  a  situation  we  must  work 
with  the  multinomial  distribution  rather  than  the  binomial  distri 
bution. 

A  common  problem  is  to  test  the  hypothesis 

H:pi  =  pi0  (i  =  1>  2,  -  -  -  ,  K) 
where  there  are  k  classes.  Of  course, 

k  k 

^  Pi  =  52  P™  =  i- 

t~l  i«~l 

A  simple  test  procedure  is  available  by  means  of  the  ehi-square  approxi 
mation.  In  this  case,  the  degrees  of  freedom  equal  h—1,  that  is,  one 
less  than  the  number  of  classes  (or  parameters).  The  procedure  is  to 
calculate 

x2  -  Z  (0<  -  RWRi  (7.22) 

*•—  i 

where  O*  represents  the  number  observed  in  the  ith  cla^s  and  JKt^npiQ 
reresents  the  number  expected  in  the  ith  class  if  //  Ls  true.  Clearly, 
i  —  n*  Then,  if  X*>XU~~«)<A-~I»  ^e  hypothesis  //  is  rejected. 


Example  7.24 

In  a  particular  genetic  experiment,  the  observations  were  classified 
as  follows: 

Class  A—99 
Class  B<—  33 
Class  C—  24 
Class  D—  4 

but  genetic  theory  called  for  a  9:3:3:1  ratio.  Using  a  5  per  cent  mg~ 
nificanee   level,    do   the   data  support   the   theory?   Calculation   yields 

xi  „.  (99  -  90)  a/^>  +  (33  -  30)V3D  -f-  (24  -  3Q}*/30  +  (4  -  I0)»/10  -  6,0. 


This  is  less  than  xtascs)  "«7.81,  and  thus  we  are  unable  to  reject  the  hy 
pothesized  theory. 

7.14      PO1SSON    DATA 

There  are  several  processes  which  give  rise  to  observations  distri 
buted  according  to  the  Poisson  probability  function 

/(#)  «  <r*X*/#l;          a;  -  0,  1,  2,  •  -  -  .  (7,23) 


7.14       P01SSON    DATA  125 

Some  examples  are:  (1)  radioactive  disintegrations,  (2)  bomb  hits  on  a 
given  area,  (3)  chromosome  interchanges  in  cells,  and  (4)  flaws  in  ma 
terials. 

Obviously,  many  hypotheses  and  alternatives  could  be  considered 
and  discussed.  However,  for  purposes  of  illustrating  the  methods  of 
analysis,  only  two  will  be  examined. 

To  test  the  hypothesis  H:\<\Q  versus  A  : X > X0,  it  would  be  appropri 
ate  to  obtain,  for  a  sample  of  one, 

P  =  i  ~  p(p  _  i)  (7.24) 

where  F(x)  is  read  from  Appendix  2  under  the  assumption  X  =  X0.  If 
P<&,  the  hypothesis  H  would  be  rejected. 

Example  7.25 

A  random  sample  of  two  phonograph  records  shows  1  and  4  de 
fects  per  record,  respectively.  Assuming  a:  =  0.01,  test  the  hypothesis 
H:\  <0.5  versus  A  :X  >0.5.  (NOTE:  This  is  testing  the  hypothesis  that 
the  average  number  of  defects  per  record  is  less  than  or  equal  to  J.) 
Since  we  have  a  total  of  5  defects  from  2  records,  we  make  use  of  the 
fact  that  w~ xi+x$  also  follows  a  Poisson  distribution  with  parameter 
X'  =  nX  =  2A.  Consulting  Appendix  2  for  X' =  2A0  =  2(§)  =  1,  we  see  that 
F(w  — 1)=^(5  —  1)=JP(4:)  =0.996  and  thus  P  =  1  —F(w  —  l) i  =0.004. 
Since  this  is  less  than  oc  =  0.01,  the  hypothesis  H:\<0.5  is  rejected  in 
favor  of  the  alternative  A  :X  >0.5. 

The  second  situation  to  be  examined  is  of  interest  from  a  methodo 
logical  point  of  view  since  it  combines  the  assumption  of  a  Poisson  dis 
tribution  with  the  chi-square  method  of  analysis.  Essentially,  it  is  a 
comparison  of  several  Poisson  distributions  to  see  if  the  parameters 
(that  is,  the  X's)  differ  significantly.  The  procedure  is  best  illustrated 
by  an  example. 

Example  7.26 

Suppose  a  phonograph  record  manufacturing  company  is  investigat 
ing  5  different  production  processes.  Four  records  are  selected  at  ran 
dom  from  those  produced  by  each  process  and  the  number  of  defects 
per  record  is  counted.  The  data  are  given,  in  Table  7.10.  Chi-square  is 
then  computed  for  each  process,  using  the  observed  process  average  as 
the  expected  number  of  defects  per  record  for  that  process.  Each  of  these 
chi-squares  has  3  degrees  of  freedom.  Using  the  additive  property  of 
chi-square,  it  is  noted  that  the  total,  9.70,  has  15  degrees  of  freedom. 
It  can  be  verified  that  none  of  these  6  values  of  chi-square  is  significant 
at  the  1  per  cent  level.  Thus,  there  is  little  question  about  the  uni 
formity  of  records  produced  by  the  same  process.  However,  if  the  chi- 
square  representing  the  variation  among  processes  is  calculated,  that  is 

X2=,  [(64-35,2)2-4-(28-35.2)2-l-(32— 35.2)2+  (32-35.2)2+  (20— 35.2)2]/35.2  =  32.18, 

we  see  that  x2==::32.18  >x%o(4)  =  13.3.  Therefore,  the  hypothesis  of  no 
differences  among  processes  is  rejected.  It  might  be  concluded  that  some 
processes  (probably  numbers  II,  III,  IV,  and  V)  will  allow  production 
of  product  containing  fewer  defects. 


126 


CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


TABLE  7.10-Number  of  Defects  per  Record  From  XYZ 
Manufacturing  Company 


Number  of  Defects 

Process 

Process 

^      (0     _  R.y 

X       fn         7                   -  •-"'  •-  -      -I,- 

Process 

per  Record 

Totals 

Means 

i              ^-^                   77 
j-1                  J&» 

I  

11,  16,  17  20 

64 

16 

42/16=2.62 

II 

5     7     5    11 

28 

7 

24/7    —3.43 

III  

11,    9,    7,    5 

32 

8 

20/8   =2.50 

IV  

8   10,    7     7 

32 

8 

6/8   =    .75 

V.  .  . 

5654 

20 

5 

2/5    «=    .40 

Total 

176 

9.70 

7.15      CHI-SQUARE  TEST  OF  GOODNESS  OF    FIT 

One  thing  that  is  often  done,  with  no  justification  other  than  saying 
it  appears  reasonable,  Is  to  assume  that  the  variate  under  discussion 
follows  a  particular  distribution.  For  example,  data  are  frequently 
assumed  to  be  samples  from  a  normal  population,  and  you  may  well 
question  this  assumption.  At  this  time,  one  procedure  useful  in  check 
ing  on  the  validity  of  such  assumptions  will  be  presented. 

The  procedure  is  to  make  a  comparison  between  the  actual  number 
of  observations  and  the  expected  number  of  observations  (expected 
under  the  "uwwiunption")  for  various  values  of  the  variate.  The  ex 
pected  numbers  are  usually  calculated  by  iivsing  the  assumed  distribxi- 
tion  with  the  parameters  set  equal  to  their  sample  estimates.  The  chi- 
sqxiare  statistic  will  be  calculated  according  to  Kquatiou  (7.22)  and  the 
degrees  of  freedom  will  be  /c— -p  — 1,  where  p  represents  the  number  of 
parameters  estimated  by  sample  statistics.  For  example,  if  a  normality 
assumption  were  xnuler  test,  &  and  a-2  would  be  estimated  by  5T  and  s*2, 
and  the  degrees  of  freedom  would  be  Ai  — 3,  where  k  represents  the  num 
ber  of  class  intervals  used  in  fitting  the  distribution.  If  the  assumption 
of  a  Poisson  distribution  were  being  tested,  X  — /x  would  be  estimated  by 
X)  and  the  degrees  of  freedom  would  be  A?  —  2. 

Rather  than  continue  the  discussion  in  general  terms,  an  example 
involving  the  Poisson  distribution  will  be  studied. 


Example  7.27 

The*  data  given  in  Table  7.11  show  the  number  of  "senders**  (a  type 
of  automatic  equipment  used  In  telephone,  exchanges)  that  were  in  xise 
at  a  given  instant.  Observations  were  made  on  3754  different  occasions. 
The  expected  numbers  were  calculated  from/(.r)  s*e,  xA*/x!  where  A  wan 
set  eqxial  to  3T«I()-44.  Hince  x2a«43.43  >X*gg(<w)  ••37.6,  the  hypothesis 
of  a  PoiBHon  diHtrihution  with  ju«*  10.44  is  rejected. 

One  point  to  bo  noted  in  Kxamplc  7.27  wan  the  combination  of  the 
entries  of  the  top  two  linen  of  the  table  to  form  a  ningle  elana.  Thin  was 


7.15       CHI-SQUARE    TEST    OF    GOODNESS    OF    FIT  127 

TABLE   7.11-Number  of  Busy  Senders  in  a  Telephone  Exchange* 


Number 

Observed 
Frequency 

Expected 
Frequency 

Deviation 

(O-JS)2 

Busy 

(0) 

OB) 

(O-E) 

JE 

0  

o\ 

0   11\ 

-t-    3   74 

11    01 

1  

> 

Si 

1    15} 

2  

**  ) 

14 

5  98 

+   8  02 

10   76 

3.  .  . 

24 

20  82 

-1-   3    18 

4-Q 

4  

57 

54  33 

+   2   67 

i  ^ 

5  

111 

113  44 

—   2  44 

05 

6  

197 

197  38 

—   0  38 

00 

7.  .. 

278 

294  38 

—  16  38 

01 

8  

378 

384.16 

—    6  16 

10 

9  

418 

445  63 

—  27  63 

1    71 

10  

461 

465  24 

—   4  24 

03 

11  

433 

441  .  56 

—    8  56 

17 

12  

413 

384.  15 

+  28.85 

2   17 

13  

358 

308.50 

+49.50 

7  94 

14  

219 

230.05 

—  11.05 

.53 

15  

145 

160.11 

—  15.11 

1  43 

16  

109 

104.47 

+   4.53 

.20 

17  

57 

64.  16 

—    7.16 

.80 

18  

43 

37.21 

+   5.79 

.90 

19...           .     . 

16 

20.45 

—   4.45 

97 

20  

7 

10.67 

—    3.67 

1.26 

21  

8 

5.31 

+   5.69 

1.36 

22  

3 

4.51 

—    1.51 

.51 

Total 

3754 

3753.77 

+   0.23 

X2  =  43.43 

*  Source:  Thornton  C.  Fry,  Probability  and  Its  Engineering  Uses  (New  York:  D.  Van 
Nostrand  Company,  Inc.,  1928),  p.  295, 


done  because  the  expected  number  on  the  first  line  was  too  small.  The 
reason  for  avoiding  such  expected  numbers  is  that  they  lead  to  large 
chi-square  values  (perhaps  even  significant  values  of  chi-square)  which 
do  not  reflect  a  departure  of  "observed  from  expected"  but  only  the 
smallness  of  the  "expect ed."  In  other  words,  if  some  expected  numbers 
are  too  small,  the  chi-square  statistic  will  be  a  poor  indicator  of  the 
validity  of  the  hypothesis  under  test.  Some  authors  say  that  "too 
small"  means  less  than  3;  others  say  less  than  5.  Since  not  everyone  is 
agreed  on  the  interpretation  of  what  is  too  small,  you  should  feel  free 
to  use  any  reasonable  definition.  Personally,  I  favor  the  value  "3." 


128  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

7.16      BINOMIAL  POPULATION;  MORE  THAN  ONE  SAMPLE 

A  situation  which  occurs  frequently  in  experimental  work  is  the  fol 
lowing:  A  hypothesis  is  to  be  tested  and  several  experiments  are  con 
ducted  to  produce  data  which  bear  on  the  problem.  When  this  situa 
tion  prevails,  it  is  natural  to  think  of  combining  the  experimental 
results.  For  example,  if  the  hypothesis  H :  p  =  po  is  being  tested  relative 
to  the  alternative  Aip^ps,  it  is  quite  common  to  have  available  k 
samples  (perhaps  of  different  sizes)  as  a  result  of  k  replications,  or 
repetitions,  of  the  basic  experiment. 

How  should  the  data  from  the  several  samples  be  combined?  There 
are  two  ways  this  can  be  done,  and  each  will  be  discussed  and  then  il 
lustrated  in  Example  7.28.  It  will  be  noted  that  the  analysis  performed 
involves  the  chi-square  distribution  and  depends  on  the  previously 
mentioned  additive  property  of  chi-square.  Actually,  several  chi-square 
values  are  calculated,  and  each  of  these  contributes  a  different  item  of 
information  relative  to  the  hypothesis  under  test. 

You  will  note  that  a  chi-square  value  (with  1  degree  of  freedom)  is 
found  for  each  sample.  Each  of  these  values  can  be  interpreted  as 
in  Section  7.7.  As  the  next  step  in  the  analysis,  wo  may  calculate 
x2===:Xi+X2  +  "  •  •  +  X&  with  k  degrees  of  freedom*  This  value  will  be 
referred  to  as  the  pooled  chi-square,  and  it  is  clearly  a  pooling  or  accum 
ulation  of  the  bits  of  evidence  provided  by  the  k  independent  samples. 
This  value  may  now  bo  used  to  assess  the  validity  of  the  hypothesis 
under  test.  An  alternative  way  of  pooling  the  information  from  several 
samples  is  to  lump  the  original  data  into  one  large  sample  arid  compute 
the  total  chi-square  (with  1  degree  of  freedom)  associated  with  this 
super  sample.  One  other  statistic  should  also  be  obtained,  namely,  the 
heterogeneity  chi-square.  This  quantity,  which  has  fc  —  1  degrees  of 
freedom,  is  found  by  subtracting  the  total  chi-square  from  the  pooled 
chi-square.  It  is  used  to  measure  the  lack  of  consistency  among  the 
several  samples. 

Example  7.28 

Consider  again,  the  hypothesis  tested  in  Example  7.12.  Now,  instead 
of  I  sample,  8  separate  experiments  give  rise  to  8  samples  as  shown  in 
Table  7.12.  Assuming  c*««0.01,  it  is  seen  that:  (1)  no  1  of  the  8  samples 
leads  to  rejection,  (2)  the  super  sample  of  1600  observations  yields 
X**»2.66  which  is  not  significant,  and  (3)  the  pooled  chi-square  is  sig 
nificant,  Why  do  we  get  these  seemingly  contradictory  results?  The 
pooled  chi-square  is  significant  because  we  have  accumulated  enough 
evidence  from  each  sample  to  indicate  that  the  hypothesis  //:p»«0,5 
should  be  rejected.  The  reason  the  total  chi-square  did  not  give  the 
same  answer  is  that  in  3  samples  smooth  seeds  predominated  while  in 
5  samples  wrinkled  seeds  predominated*  This  effect  was  hidden  (i.e., 
the  majorities  in  opposite  directions  tended  to  cancel  out)  when  the 
data  were  lumped  into  one  large  sample.  Attention  is  called  to  the 
previously  mentioned  lack  of  consistency  among  the  8  samples  by  the 
significant  heterogeneity  chi-square. 


CONTINGENCY    TABLES 


129 


TABLE   7.12-Chi-Square  Analysis  Combining  Data  From  Several  Samples 

of  Smooth  and  Wrinkled  Peas 


Sample 
Number 

Sample 
Size 

Number 
Wrinkled 

Number 
Smooth. 

On-Square  (x*) 

d.f. 

1  

100 

60 

40 

4  00 

i 

2  

200 

108 

92 

1    28 

i 

3  

180 

80 

100 

2   22 

i 

4  

208 

118 

90 

3    77 

i 

5  

300 

165 

135 

3  00 

1 

6  

182 

106 

76 

4  Q4 

1 

7  

230 

105 

125 

1    7S 

i 

8  

200 

90 

110 

2  00 

1 

Pooled  x2-  -  - 

8 

22.94  =    y^  x2 

8 

Total 

1600 

832 

768 

•"•*"  •  7  •*              ^—-/    A.  . 
<-l         * 

2.56 

1 

Difference 

20  38 

7 

7.17      CONTINGENCY  TABLES 

Suppose  n  randomly  selected  items  are  classified  according  to  two 
different  criteria.  The  tabulation  of  the  results  could  be  presented  as 
in  Table  7.13,  where  O^  represents  the  number  of  items  belonging  to 

TABLE  7.13-An  rXc  Table 


Rows 


1. 

2. 


Columns 


02c 


Or 


the  «?")th  cell  of  the  rXc  table.  Such  data  can  be  used  to  test  the  hy 
pothesis  that  the  two  classifications,  represented  by  rows  and  columns, 
are  statistically  independent.  If  this  hypothesis  is  rejected,  the  two 
classifications  are  not  independent  and  we  say  there  is  some  interaction 
between  the  two  criteria  of  classification. 

The  exact  test  for  independence  is  difficult  to  apply.  However,  if  n, 
the  sample  size,  is  sufficiently  large,  a  reasonably  good  approximate 
procedure  is  to  calculate 


X2  = 


(7.25) 


130  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

where 

Oij  =  observed  number  in  the  (i/)th  cell, 

=  expected  number  in  the  (f/)th  cell, 

^j  =  observed  number  in  the  ith  row,  and 


y— i 


=  observed  number  in  the/th  column. 


The  value  of  chi-square  given  by  Equation  (7.25)  has  *>=(?  —  l)(c  —  1) 
degrees  of  freedom.  If  xa^x*i--«>i:<r--i)Ce-i>]>  the  hypothesis  of  inde 
pendence  should  be  rejected. 

Example  7.29 

A  company  has  to  choose  among  throe  proposed  pension  plans.  One 
hypothesis  that  the  company  wislics  to  investigate  is:  Preference  for 
plans  is  independent  of  job  classification.  It  asks  the  opinion  of  a 
sample  of  the  employees  and  obtains  the  information  presented  in 
Table  7.14.  The  expected  numbers  for  each  ceil  arc  calculated  and 
appear  in  Table  7.15.  Calculation  then  yields  x2==ll  <X*99(6)  ~  10*8  so 
the  hypothesis  cannot  be  rejected.  Thus,  it  is  concluded  that  the  em 
ployees'  choices  of  pension  plans  are  quite  probably  independent  of  their 
job  classifications. 

TABLE  7,14-Classification  of  Employees  by  Job  and 
Pension  Plan  Preference 


Number  of 

Employees 

Favoring 

Classification 

Plan  A 

Plan  B 

Plan  C 

Total 

Factory  employees    ...*.... 

160 

30 

10 

200 

Clerical  employees  .  .  *  .,,.*. 

140 

40 

20 

200 

Foremen  and  supervisors  .... 

Executives      

80 
70 

10 
20 

10 
10 

100 
100 

Total 

450 

100 

SO 

600 

TABLE  7JS~~Expected  Number  of  Observations 


Classification 

Wan  A 

Plan  B 

Plan  C 

Factory  employees  *     ,  

150 

100/3 

50/3 

Clerical  employees  ,  

ISO 

100/3 

50/3 

Foremen  sine!  supervisors  

75 

100/6 

50/6 

Executives  ..*<,,  

75 

100/6 

50/6 

7.18       SPECIAL    APPROXIMATE    METHODS    FOR    2X2    TABLES 


131 


If  we  are  presented  with  an  A7'- way  contingency  table,  that  is,  one  in 
which  the  individual  elements  are  assigned  to  the  cells  of  the  table  by 
N  different  criteria,  the  hypothesis  of  mutual  independence  of  the  N 
criteria  may  be  tested  by  a  simple  extension  of  the  rules  formulated  for 
the  rXc  table.  As  usual,  we  shall  compute  the  sum  (over  all  cells)  of 
"(observed  —  expected) ^/expected,"  where  the  expected  value  in  any 
cell  is  given  by  the  product  of  the  marginal  (border)  totals  associated 
with  the  row,  column,  etc.,  in  which  the  cell  is  located  divided  by  n^"1. 
The  resulting  statistic  is  approximately  distributed  as  chi-square  with 
"(r  —  l)(c—l)  •  •  .  "  degrees  of  freedom,  where  there  are  r  rows,  c  col 
umns,  etc.,  in  the  YV-way  table.  Other  hypotheses  may  also  be  tested 
in  such  tables,  for  example,  see  Mood  (12),  but  we  shall  not  discuss 
these  at  this  time. 


7.18 


SPECIAL 
TABLES 


APPROXIMATE       METHODS       FOR       2X2 


If  the  contingency  table  consists  of  two  rows  and  two  columns,  as  in 
Table  7.16,  a  short-cut  method  of  computing  chi-square  is  available. 
The  appropriate  formula  is 


/&  (7.26) 

where  n  =  a+b  +  c+d  and  k=  (a+&)  (c+d)  (a+c)  (6  +  d).  This  will  give 
the  same  numerical  value  of  chi-square  that  would  be  obtained  if  the 
procedure  of  Section  7.17  were  followed.  It  should  be  clear  that  the 
chi-square  statistic  thus  obtained  will  have  only  1  degree  of  freedom. 

TABLE    7.16-A  2X2  Table 


Ai 

A* 

Total 

JBi  

d 

I 

a+b 

Bi  

c 

d 

c+d 

Total 

a+c 

b  +  d 

n 

As  in  Section  7.7,  a  correction  for  continuity  may  be  used  to  sharpen 
the  approximation.  This  is  accomplished  by  calculating 

=  «(  |  ad  —  be  |    —  n/2y/k 


X 


~  0.5) 


(7.27) 


It  must  be  remembered  that  this  correction  should  not  be  applied  to 
rXc  tables  in  which  r>2  and 


Example  7.30 

A  random  sample  of  250  men  and  250  women  were  polled  as  fo  their 
desires  concerning  the  ownership  of  television  sets.  The  data  in  Table 
7.17  resulted.  Calculation  by  either  method  yielded 


<?  =  13*  > 


L.99(l> 


6.63. 


132 


CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


Thus,  the  hypothesis  that  desire  to  own  a  television  set  is  independent 
of  sex  is  rejected. 

TABLE   7.17-Results  of  Sample  Poll  on  Television  Ownership 


Classification 

Men 

Women 

Total 

Want  television  

80 

170 

250 

Don't  want  television 

120 

130 

250 

Total 

200 

300 

500 

7.19  THE  EXACT  METHOD  FOR  2X2  TABLES 

It  should  be  noted  that  an  alternative  way  of  looking  at  a  2X2  table 
is  to  consider  the  two  fractions,  pi  =  a/(a+6)  and  P2=c/(c-|-rf)7  as  esti 
mates  of  pi  and  p<2>  the  parameters  of  two  binomial  populations.  In  this 
frame  of  reference,  a  comparison  of  pi  and  p%  should  yield  evidence  rela 
tive  to  the  hypothesis  H:pi~p^  If  we  wish  to  test  fl:pi~pz  versus 
Aipi^pz,  we  may  use  the  approximate  method  of  the  preceding  sec 
tion  or,  if  we  choose,  an  equivalent  test  based  on  the  normal  approxi 
mation, 

In  this  case,  however,  the  exact  test  procedure  is  not  too  difficult  to 
apply,  especially  if  a  digital  computer  is  available.  Thus,  it  seems  ap 
propriate  to  indicate  the  nature  of  the  exact  method. 

It  can  be  shown  that  the  exact  probability  of  observing  pi  =  a/(a+b) 
and  p^^c/(c+d)  when  pi~pz  is 


+ 


I 


alblcldlnl 


(7.28) 


To  obtain  the  final  probability  to  be  used  in  assessing  the  validity  of 
//:y>i  =  p2»  it  is  necessary  to  add  to  Pi  the  probabilities  of  more  diver 
gent  fractions  than  those  observed.  Assuming;  Pi<.p%  (and  the  table 
can  always  be  arranged  to  make  this  so),  the  next  more  divergent  situa 
tion  would  be  the  one  in  which  a  and  d  are  each  decreased  by  xmity,  and 
6  and  c  are  each  increased  by  unity.  For  this  array,  we  calculate 


(7.29) 


The  cell  entries  are  again  changed,  following  the  same  rule  as  before, 
and  PS  is  calculated.  Continue  in  this  manner  until  Pa+\  Is  calculated. 
Then,  if 


(7.30) 


*'— I 


is  less  than  or  equal  to  ix,  the  hypothesis  H:pi  =  pz  should  be  rejected. 


7.20       SEVERAL    NORMAL    POPULATIONS;    H:}!*  =  M«  .  .  .  =  I**  133 

Example  7.31 

Robertson  (13)  reported  on  the  analysis  of  an  experiment  involving 
the  evaluation  of  a  silicon  dip  as  a  protection  for  vacuum  tubes.  The 
data  shown  in  Table  7.18  were  obtained.  Using  the  procedure  outlined 
above,  he  found  P  =  PX+P2+P3  =  .0957.  He  then  concluded  that  "the 
failure  rate  for  protected  tubes  is  fust  barely  significantly  less  than  that 
for  unprotected  tubes.77  Apparently  a  10  per  cent  significance  level  had 
been  decided  upon  prior  to  the  analysis. 

TABLE  7.18-Success-Failure  Results  From  an  Experiment 
on  690  Vacuum  Tubes* 


Failures 

Nonfailures 

Protected  

*2 

338 

Unprotected    .  .  . 

7. 

343 

/ 

*  Source:  W.  H.  Robertson,  "Programming  Fisher's  exact  method  of  comparing  two  per 
centages,"  Technometrics,  Vol.  2,  No.  1,  pp.  103-7,  Feb.,  1960. 

7.20     SEVERAL  NORMAL  POPULATIONS;  H:^=^2  =^ 

In  Section  7.9  a  &-test  was  proposed  for  testing  H:JJL^  =  ^  versus 
Aipir^fjiz  under  the  assumption  that  cri  =  cr|.  Now  we  wish  to  propose 
a  procedure  for  handling  the  situation  in  which  we  have  k  normal  pop 
ulations,  fc>2. 

Intuitively,  it  seems  reasonable  that  the  validity  of  the  hypothesis 
H :  MI  =  M2  =  -  -  •  =  &k  should  be  assessed  by  comparing  the  sample  esti 
mates  of  pii,  Hz,  •  •  •  ,  Mfc.  That  is,  it  is  to  be  expected  that  any  suggested 
test  procedure  will  involve  a  comparison  of  Fa,  F2,  •  •  •  ,  7*;.  (NOTE: 
The  choice  of  Y  rather  than  X  as  the  symbol  denoting  the  character 
istic  was  prompted  solely  by  the  desire  to  agree  with  symbolism  to  be 
used  in  certain  techniques  that  will  be  presented  later  in  the  book.) 

If  the  assumption  is  made  that  o-f  ==  of  =  -  -  -  =  a>,  that  is,  if 
homogeneous  variances  are  assumed,  the  appropriate  test  procedure  is 
to  calculate 


_  *— 1 

*•  =  — - (7.31) 

k          n<  y        fc 

y^  v  f  F--  —  T^2  /  ^T  (*  —  n 

--C-^  ^—r  v  ^  u  *  *J      /      Z-j  \™i         *•} 

where 

F*7  —  jth  observation  in  the  ith  group  (sample)  ;  i  =  1,  -  -  •  ,  k      (7.32) 

J  =  1,  -  -  -  ,  n* 
n*  =  number  of  observations  in  the  ith  group  (7 . 33) 


134 


CHAPTER    r,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


ii/ni  =  mean  of  the  observations  in  the  ith  group  (7  . 34) 


Y  = 


Then,  if  F>Fci-. 


of  all  observations. 


ya) where  *>i  =  k  —  1  and 

k 

X-  -  i), 


z— 1 


(7.35) 


the  hypothesis  H  :  /*i  ==  pc2  =  -  -  •  =M^  would  be  rejected.  (NOTE:  The 
reader  may  easily  verify  that,  if  k  =  2y  the  procedure  outlined  in  this 
section  is  algebraically  equivalent  to  that  of  Section  7.9.) 

Before  presenting  an  example,  a  convenient  tabular  form  for  carry 
ing  out  the  specified  test  procedure  will  be  indicated.  Invoking  the 
identity 


km 

^  £ 


or,  in  abbreviated  form, 


G 


vv 


K/ 


(7.36) 


(7.37) 


the  necessary  calculations  are  conveniently  presented  as  in  Table  7.19. 
(NOTE :  Such  a  table  is  usually  referred  to  as  an  analysis  of  variance 
table.)  Actually,  the  labor  involved  in  calculating  the  various  sums  of 
squares  may  be  materially  reduced  if  we  use  the  following  algebraically 
equivalent  forms : 

TABLE  7,19-Tabular  Presentation  of  the  /Mest  for  the  Equality  of 

Means  of  k  Normal  Populations  Under  the  Assumption 

of  Homogeneous  Variances 


Source  of 
Variation 


Among  groxips .  ,  . 
Within  groups.  . .  . 

Total 


Degrees  of 
Freedom 

Sum  of 
Squares 

Mean  S^ 

1 

ib  (««  -  1) 

Gyv 

G™.  /***    /  /  1 
v^i/j//  t" 

V-fTw,/ 

k 

y*  y» 

*-d 

F-  Ratio 


G/W 


7.20       SEVERAL    NORMAL    POPULATIONS;    H:JA,i  =  \*>*  .  .  .  =  M.A 


135 


Myy      = 


=  sum  of  the  squares  of  all  the  observations,  (7.38) 

/     & 

it  (7.39) 


and 


In  the  above  equations, 


(7.40) 
(7.41) 


"^-  =  total  of  the  observations  in  the  ^th  group,      (7.42) 


i  =  total  of  all  observations, 


(7.43) 


and 

k 


=  total  number  of  observations  in  all  the  groups  combined.    (7.44) 


Example  7.32 

Consider  the  data   of  Table  7.20.   Using  Equations   (7.38)    through 
(7.41),  the  results  shown  in  Table  7.21  were  obtained.  Since 

F  =  72  >  ^.99(3,16)  =  5.29, 


the  hypothesis  H  :/xi=/jt2  = 
cance  level. 


is  rejected  at  the  1  per  cent  signifi 


TABLE  7.20-Sample  Data  From  Four  Normal  Populations  To  Be  Used 

in  Example  7.32 


Groups 

1 

2           3 

4 

45 

35          34 

41 

46 

33          34 

41 

49 

35 

44 

44 

34 

43 

Observations 

33 

41 

42 

44 

41 

41 

Totals 

184 

68         170 

378 

Means 

46 

34          34 

42 

136 


CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


TABLE   7.21-Analysis  of  Variance  Using  the  Data  of  Table  7.20  To  Test 
the  Hypothesis  Hi^i  =M2  =  M3  =  ^4 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

Meatx      .    .         . 

1 

32   000 

32   000 

Among  groups  

3 

432 

144 

72 

Within  groups  

16 

32 

2 

Total 

20 

32,464 

^ 

7.21      SEVERAL 

H  :oi  =  erf  = 


NORMAL  POPULATIONS; 

•    •    •    =  erf 


In  Section  7.11  an  .P-test  was  proposed  for  testing  the  hypothesis 
Hi <r?  =  <T2  versus  A:<rf=^cr|.  At  this  time,  we  wish  to  consider  the 
situation  in  which  we  have  k  normal  populations,  &>2.  Several  test 
procedures  have  been  proposed  for  handling  this  type  of  problem,  but 
only  the  method  due  to  Bartlett  (2)  will  be  presented  in  this  book. 

As  in  Section  7.207  the  sample  observations  will  be  denoted  by 
Y a  0&=  1,  -  •  -  ,  7c;j  =  l,  •  -  -  3  nt).  Other  symbols  will  also  be  defined  as 
in  the  preceding  section  and,  in  addition,  we  will  denote  K^  — "F»-  by 
2/ij.  Thus,  in  agreement  with  an  earlier  definition, 

nt 

Using  this  notation,  the  mechanics  of  Bartlett's  procedure  are  as 
shown  in  Table  7.22.  If,  in  this  table,  x2^x*i-.«o<*--i:>>  the  hypothesis 
£r:a-f  =  cr|==  -  .  -  =  cr|  would  be  rejected.  (NOTE:  The  researcher 
will  find  it  necessary  to  compute  the  corrected  value  of  chi-square  only 
if  the  uncorrected  chi-square  falls  close  to  and  above  the  tabulated 
value,  and  then  only  if  he  wishes  to  obtain  a  very  accurate  evaluation 
of  the  exact  probability  of  Type  I  error.) 


Example  7.33 

Consider  the  data  of  Table  7.23.  Following  the  procedure  indicated 
in  Table  7.22,  the  results  presented  in  Table  7,24  are  obtained.  It  is 
seen  that  x2sa=2.81  <x^95(s)  SBB7,81J  and  thus  the  hypothesis  of  homoge 
neous  variances  may  not  be  rejected  at  the  5  per  cent  significance  level. 
(WOTE:  There  was  no  need  to  calculate  the  corrected  value  of  chi- 
square  in  this  example;  the  computations  were  carried  out  only  to  illus 
trate  the  method.) 


7.22      SAMPLE  SIZE 

A  question  frequently  asked  of  statisticians  is, 


is  needed  for  this  experiment? 


How  large  a  sample 
The  question  is  deceptively  simple, 


7.22      SAMPLE   SIZE  137 

TABLE  7.22-Computations  for  Bartlett's  Test  for  Homogeneity  of  Variance 


Sample 

Z2 
%' 

Degrees  of 
Freedom 

!/<*/. 

2 

^i 

logiQSf 

(d./.)  logios? 

1  

A        2 
7       'Vi,' 

'Tl\  —  —    1 

l/(n*  —  1) 

2 

c. 

loeTin^i 

(**.  —  .  1^  lnefin?i 

2  

J~l 

712 
EAf«,- 

^2   —    1 

i/cn_  _  i\ 

2 

Crt 

Ififfi  r»9<> 

f'yt.n    —  —     1  )    lOffi  a.V«» 

k      . 

J-l 

n&         „ 

Zyi 

TZjfc   —    1 

l/(nk  —  11 

2 

c  » 

IrtOt,  n  C» 

(  <vt  i     ^«.    1  )    1  /-)  nr-  n  c  t 

j-i 

Sum 

PP", 

fc 
Efo-            11 

^     i 

^         1*7*       —     I"1)   lrt<TinC* 

KK  i/i/ 

t-1 

t-l    Wt   ~   1 

i-1 

/     fc 

Pooled  estimate  of  variance  =  s*  =  TFV2/    /   E  (w*  ™" 

/      i-i 


log 


.  10  [B  -  i  <*,  -  i)  iogl05<  "] 

I—  i—  I  -J 


Correction  factor  -  C 
Corrected 


1  +  [l/3(*  -  l)]f  Z)  -  -  --  1    /  S  («<  -  1)1 

L   i-i     Wi    —    1  /         i«.i  J 

Note:  log«  10  -  2.3026 


but  the  answer  is  hard  to  find.  Before  the  statistician  can  provide  any 
thing  better  than  an  "educated  guess/'  he  must  retaliate  with  several 
questions,  the  answers  to  which  should  enable  him  to  attack  the  prob 
lem  with  some  hope  of  reaching  a  valid  answer.  Frustrating  as  this  may 
be  to  the  researcher,  it  frequently  serves  a  very  good  purpose,  for  it 
forces  the  researcher  to  give  serious  thought  to  several  aspects  of  his 

TABLE  7.23-Four  Samples  From  Normal  Populations 

JS  •<?!  ===Cr2  :==0"3  ^^  04 


1 

2 

3 

4 

48 

42 

33 

78 

49 

39 

42 

69 

67 

51 

46 

60 

75 

57 

47 

52 

53 

75 

50 

63 

33 

45 

50 

35 

138  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

TABLE   7.24-Computatlons  for  Bartlett's  Test:  Data  from  Table  7.23 


Sample 

Zrf 

Degrees 
of  Free 
dom 

1/d.f. 

4 

logics? 

(dj.)  Iog105? 

1  

1113.0 

5 

.2000 

222.6 

2.34753 

11.73765 

2  

820.8 

4 

.2500 

205.2 

2.31218 

9.24872 

3  

173.2 

4 

.2500 

43.3 

1  .  63649 

6.54596 

4.  ,  .  . 

1330.0 

7 

.1428 

190.0 

2.27875 

15,95125 

Sum 

3437.0 

20 

.8428 

43.48358 

Pooled  estimate  of  variance  =  sz  =  3437/20  =  171.85 


B 


<X-  -  1)  =  (2.23515)  (20)  =  44,7030 


(2.3026)  (44.7030  -  43.48358)  =  2.80784 


Correction  factor  =  C  «  1  4-  [1/3(3)  ]  (.8428  —  1/20)  «  1.0881 


Corrected 


2.80784/1.0881  «  2.5805 


problem.  To  illustrate,  some  of  the  questions  that  might  be  asked  by 
the  statistician  are: 

(1)  What  is  your  hypothesis?  What  are  the  alternatives? 

(2)  What  are  you  trying  to  estimate? 

(3)  What  significance  level  are  you  planning  to  use?  What  confi 
dence  level? 

(4)  How  large  a  difference  do  you  wish  to  be  reasonably  certain  of 
detecting?  With  what  probability? 

(5)  What  width  confidence  interval  can  you  tolerate? 

(6)  What  do  you  expect  the  variability  of  your  data  to  be? 

When  answers  to  these  and  other  questions  are  provided  by  the  re 
searcher,  the  statistician  can  be  of  help  in  determining  the  needed 
sample  size. 

Before  you  get  the  impression  that  all  is  lost,  let  me  hasten  to  assure 
you  that  the  picture  is  not  all  black.  In  some  cases,  fairly  simple 
formulas  arc  available  for  estimating  the  required  sample  size.  Also,  if 
OC  curves  are  available  for  the  test  procedure  to  be  used,  the  reqxiired 
sample  size  may  be  determined  upon  examination  of  these  curves. 
Tables  have  also  been  provided  for  certain  procedures  and  four  of  these 
are  reproduced  in  Appendices  9  through  12  for  your  use.  If  all  of  these 
three  approaches  (that  is,  formulas,  OC  curves,  or  tables)  fail  to  meet 
your  demands,  a  professional  statistician  should  be  consulted. 

Example  7.34 

Consider  testing  the  hypothesis  //r^^Mo  versus  A  :^F^Q  at  the  5  per 
cent  significance  level.  If  <r  is  estimated  to  be  0.8  and  a  difference 
5»  (M— MO|  =1.2  is  to  be  detected  with  probability  0.9  (this  is  equiva- 


7.22       SAMPLE    SIZE  139 

lent  to  setting  /3  =  0.1  at  ^==^0~  1.2  and  at  M=Aio+1.2),  how  large  a 
sample  is  needed?  Setting  D  =  1.2/0.8  =  1.5  and  consulting  Appendix  9, 
it  is  found  that  n  =  7. 

Example  7.35 

Consider  testing  H  :JJL  <MO  versus  A  :ju  >/zo  at  the  1  per  cent  significance 
level.  If  or  is  estimated  to  be  1.2  and  5=ju  —  ju0—  0.9  is  to  be  detected 
with  probability  0.95,  how  large  a  sample  is  needed?  Setting 

D  =  0,9/1.2  =  0.75 
and  consulting  Appendix  9,  it  is  found  that  n  =  3l. 

Example  7.36 

Consider  testing  H  ://i  <pc2  versus  A  :^i  >Ma  at  the  2§  per  cent  signifi 
cance  level.  If  <r  is  estimated  to  be  1.0  and  5=/z3  —  MS  —  1.6  is  to  be  de 
tected  with  probability  0.99,  how  large  should  the  two  samples  be?  Set 
ting  D  =  1.6/1.0  =  1.6  and  consulting  Appendix  10,  it  is  found  that 
n1==n2  =  16. 

Example  7.37 

Consider  testing  Hip.  3=/x  2  versus  A  I/JLI  7*1*2  at  the  1  per  cent  significance 
level.  If  cr  is  estimated  to  be  1.5  and  <5  =  [MI—  Ma|  =1.8  is  to  be  detected 
with  probability  0.95,,  how  large  should  the  two  samples  be?  Set 
ting  Z>  =  1.8/1.5  —  1.2  and  consulting  Appendix  10,  it  is  found  that 
ni  =  ™2  =  27. 

Example  7.38 

Consider  testing  H:CTZ<O-Q  versus  A'cr*>o$  at  the  5  per  cent  sig 
nificance  level.  If  a  value  of  cr2  =  4oo  is  to  be  detected  with  probability 
0.99,  how  large  a  sample  is  needed?  Using  R=4  and  consulting  Ap 
pendix  11,  it  is  seen  that  15  <v  <20.  Crude  interpolation  suggests 
*>  =  19  or  n  = 


Example  7.39 

Consider  testing  H:O-*>CTQ  versus  A:o*<o%  at  the  5  per  cent  signifi 
cance  level.  If  a  value  of  cr2  —  0.33  o%  is  to  be  detected  with  probability 
0.99,  how  large  a  sample  is  needed?  Since  Appendix  11  is  constructed 
for  values  of  R>1,  a  slight  change  in  procedure  (from  Example  7.38)  is 
required.  The  table  in  Appendix  11  is  entered  with  of  ==/?==  0.01, 
^'=*  o:  =  0.  05,  and  R'  =  1/R  =  3.  Thus,  it  is  noted  that  24  o  <30.  Crude 
interpolation  suggests  ^  =  26  or  n==^+l=27.  (NOTE:  Although  the 
roles  of  a.  and  /?  were  interchanged  when  Appendix  11  was  consulted, 
the  actual  test  would  be  carried  out  at  the  original  value  of  ex.  which,  in 
this  example,  was  0.05.) 

Example  7.40 

Consider  testing  H:<ri>a%  versus  ^L:<jf<cr|  at  the  .5  per  cent  sig 
nificance  level.  If  a  value  of  (r!  =  4crf  is  to  be  detected  'with  probability 
0.99,  how  large  should  the  two  samples  be?  Using  J?  =  4  and  consulting 


i4O  CHAPTER  7,  STATISTICAL  INFERENCE:  TESTING  HYPOTHESES 

Appendix  12,  it  is  noted  that  30  Oj,=*>2  <40.  Crude  interpolation  sug 
gests  that  i>i  =  if  2  ==  34  or  n\  =  n2  =  35. 

7-23      SEQUENTIAL  TESTS 

In  all  the  test  procedures  described  thus  far,  the  sample  size  has  been 
decided  upon  in  advance.  As  has  been  inferred,  the  determination  of 
the  proper  sample  size  is  often  difficult .  However,  given  the  necessary 
information  (e.g.,  an  estimate  of  the  variability  to  be  encountered  and 
statements  concerning  the  allowable  risks  associated  with  incorrect 
decisions) ,  the  required  sample  size  may  be  specified  (see  Section  7.22). 
The  reader  should  realize,  though,  that  there  is  a  certain  "cost77  at 
tached  to  such  an  approach.  That  is,  there  is  an  implicit  assumption 
in  the  fixed  (predetermined)  sample  size  approach  that  a  sample  of  the 
specified  size  will  be  taken,  and  observations  recorded  for  each  sample 
unit,  regardless  of  whether  all  the  observations  are  needed  to  reach  a 
decision.  In  view  of  this  and  in  the  hope  of  achieving  economies  due  to 
reduced  sample  sizes,  it  seems  desirable  to  seek  a  test  procedure  in 
which  the  sampling  may  be  terminated  as  soon  as  it  is  possible  to  reach 
a  decision  to  either  accept  or  reject  the  hypothesis  under  test.  For  cer 
tain  specific  cases,  namely,  those  which  involve  a  simple  hypothesis 
£T:0  =  0o  and  a  single  alternative  A  :0  =  01?  such  a  test  has  been  devised. 
It  is  known  as  the  sequential  probability  ratio  test.  In  the  remainder  of 
this  section,  the  general  nature  of  this  procedure  will  be  described  and 
certain  specific  applications  illustrated. 

The  sequential  method  of  testing  proceeds  as  follows:  Sample  units 
are  randomly  selected  one  at  a  time  (i.e.,  sequentially)  and,  after  each 
observation  is  obtained,  one  of  the  following  decisions  is  made: 

(1)  Accept  J/:0  =  #o  (i.e.,  reject  A:d^&i). 

(2)  Reject  H:Q  =  9v  (i.e.,  accept  A:0^di). 

(3)  Obtain  an  additional  observation. 

To  determine  which  of  these  three  decisions  is  appropriate,  the  ana 
lyst  should  calculate 

-R.-II—4  (7-46) 

*-i    /o(#*) 

where  /o(#)  is  the  probability  function  (or  probability  density  func 
tion)  under  the  assumption  that  H:Q^QQiB  true  and/i(x)  is  the  proba 
bility  function  (or  probability  density  function)  under  the  assumption 
that  A  :&***&i  is  true.  Then,  depending  on  the  value  of  /?*,  one  of  the 
three  decisions  previously  listed  is  reached  by  proceeding  according  to 
the  following  rule: 

(1)  If  Rn<&(l~-c*)y  accept  H  (i.e.,  reject  A). 

(2)  If  jB*>(l—0)/«,  reject  H  (i.e.,  accept  A). 

(3)  If  £/(!  —  «)  <Bn  <  (1  ~-*j8)/a,  obtain  an  additional  observation. 


7.23      SEQUENTIAL   TESTS  141 

In  the  above,  a.  and  /5  are,  respectively,  the  preassigned  risks  of: 
(1)  rejecting  H  when  H  is  true  and  (2)  accepting  H  when  A  is  true.  If 
an  and  rn  are  used  to  denote  the  acceptance  and  rejection  values,  re 
spectively,  for  a  test  statistic,  the  decision  rule  may  be  restated  in  the 
following  form: 

(1)  If  the  value  of  the  test  statistic  is  less  than  or  equal  to  an, 
accept  H  (i.e,,  reject  ^4). 

(2)  If  the  value  of  the  test  statistic  is  greater  than  or  equal  to  rnt 
reject  H  (i.e.,  accept  A}. 

(3)  If  the  value  of  the  test  statistic  is  greater  than  an  and  less 
than  rn,  continue  sampling. 

It  should  be  clear,  of  course,  that  the  sample  size  is  a  variable  in  a 
sequential  procedure  as  contrasted  to  its  role  as  a  (predetermined) 
constant  in  the  classical  test  procedures.  Thus,  in  addition  to  examin 
ing  the  power  of  a  sequential  test  procedure  by  studying  its  OC  func 
tion,  it  is  appropriate  that  its  "cost"  be  assessed  by  considering  the 
average  size  of  sample  required  to  reach  the  decision  to  accept  or  to 
reject.  This  analysis  is  usually  made  in  terms  of  the  ASN  function, 
where  the  letters  ASN  stand  for  average  sample  number.  Rather  than 
go  into  details  concerning  the  ASN  function  and  the  savings  due  to 
reduced  sample  sizes,  let  us  be  content  with  the  general  statement  that 
the  potential  savings  are  considerable,  in  some  cases  as  much  as  50 
per  cent. 

Considerable  space  could  be  devoted  to  a  detailed  discussion  of  the 
sequential  probability  ratio  test  for  each  of  the  commonly  encoun 
tered  situations.  However,  it  is  doubtful  if  such  discussions  would  serve 
any  useful  purpose.  Accordingly,  the  tests  have  been  specified  in  Table 
7.25. 

Example  7.41 

Consider  a  binomial  popxilation  and  the  hypothesis  JEJ:p  =  0.10 
versus  the  alternative  A:p~0.20.  Let  <x  ==0.01  and  /?==  0,05.  Then  log 
[/3/(l  — «)]  =  —2.986  and  log  [(1  ~/3)/a]  =4.554.  If  we  represent  a 
sample  unit  possessing  the  characteristic  associated  with  p  by  the 
symbol  d  and  a  unit  not  possessing  this  characteristic  by  g  (e.g.,  defec 
tive  and  nondefective  units,  respectively),  then  the  sequence 

gggdgdggdgggddgdgddgd 
would  terminate  at  this  point  with  the  decision  to  reject  H  and  accept  A. 

Example  7.42 

Consider  a  normal  population  with  known  standard  deviation,  cr  =  10. 
Test  the  hypothesis  £f:/x  =  50  versus  the  alternative  A:jjL  =  7Q.  Let 
a  ===0,01  and /?==  0.01.  Then  log  [/?/(!— a)  ]  = —4.595  and  log  [(1— /8)/a] 
—  4.595.  If  sequential  sampling  yielded,  in  the  order  shown,  the  fol 
lowing  values  of  X  (60,  75,  65,  70),  the  sampling  would  terminate  at  this 
stage  with  the  decision  to  reject  H  and  accept  A. 


I 

c3 


| 

w 

! 

o 


s 

I 

2 

PH 


o 


?x 

I 


Is 


/ 

CQ. 


!V 


^ 

I 


. 

ecu 


•S 

o 


CO 

.A 


PP 

25 


£ 

CO 

s 


8     ° 

Cj         <!.* 

n  H' 


§3     "§S 

8    S 


1 

>< 


1X1 


5 


.1 

pq 


A 

- 


A 

to*        =3. 


I      1     A 

r^fl  6ft  '    X 


a   A 


it     ii 

ex*    ex 


b      t> 

II      II 

b        b 


II         If 
b       b 


II        II        II        II        II        II 

fcq    *nj    tej    ^    tej    ^ 


-S 

SJ 


•c 


PROBLEMS  143 

Problems 

7.1  A  company  engaged  in  the  casting  of  pig  iron  must  be  concerned  with 
the  per  cent  of  silicon  in  the  pig  iron.  The  data  given  below  constitute 
a  random  sample  of  the  production  records.  Using  a.  ==0.02  and  assum 
ing  normality,  test  the  hypothesis  that  the  process  average  is  0.85 
grams  of  silicon  per  100  grams  of  pig  iron. 

NUMBER  OF  GRAMS  OF  SILICON  INT 
100-GRAM  SAMPLES  OF  PIG  IRON 


1.13 

0.87 

0.80 

0.92 

0.85 

0.81 

0.60 

0.97 

0.97 

0.48 

0.92 

1.00 

0.94 

0.92 

0.72 

0.61 

1.17 

0.81 

0.87 

0.71 

0.36 

0.97 

0.68 

0.89 

0.73 

1.16 

0.82 

0.68 

0.79 

1.00 

7.2  Consider  these  observations  to  represent  the  average  hourly  earnings 
during  May,  1940,  of  a  random  selection  of  50  male  workers  in  a  speci 
fied  industry. 

EARNINGS 
(in  Cents  per  Hour) 


35 

65 

68 

77 

81 

52 

82 

74 

73 

71 

68 

79 

73 

70 

67 

82 

61 

77 

84 

56 

29 

53 

61 

83 

92 

99 

80 

62 

50 

64 

76 

47 

59 

64 

72 

55 

63 

107 

48 

70 

55 

70 

43 

66 

85 

79 

90 

39 

88 

86 

(a)    What  is  your  best  estimate  of  the  average  hourly  earnings  for  all 

male  workers  in  the  industry? 
(6)    How  good  is  your  estimate  in  (a)   above?  What  is  its  standard 

error? 


144  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

(c)  Establish  confidence  limits  for  your  estimate  in  (a)  above.  Write 
out  your  statement  about  these  confidence  limits  in  words.  State 
your  assumptions  clearly. 

(d)  Is  your  estimate  in  (a)  above  in  agreement  with  the  hypothesized 
true  value   of   68   cents   per  hour  for  average  earnings  in  May, 
1940?  Explain  your  answer. 

(e)  What  additional  data  would  you  need  to  estimate  the  total  earn 
ings  in  the  industry  for  the  month  of  May? 

(/)    Test  the  hypothesis  that  ^  <80. 

7.3  Test  the  hypothesis  that  the  mean  life  (in  years)  of  wooden  telephone 
poles  is  less  than  8  years.  State  any  assumptions  you  make  about  the 
following  data: 


LENGTH  OF  LIFE  OF  1000  WOODEN 
TELEPHONE  POLES 


Life 
(in  years) 

Number  of  Poles 
Replaced 

.5  but  under     1  ,5 

4 

1.5  but  under     2.5 

7 

2.5  but  under     3.5 

15 

3  .5  but  under     4.5 

32 

4.5  but  under     5.5 

30 

5.5  but  under     6.5 

57 

6.5  but  under     7.5 

61 

7.5  but  under     8.5 

73 

8.5  but  under    9.5 

96 

9.5  but  under  10.5 

104 

10  .  5  but  under  11,5 

103 

11.5  but  under  12,5 

95 

12.5  but  under  13.5 

91 

13.5  but  under  14.5 

73 

14.5  but  under  15.5 

64 

15.5  but  under  16.5 

38 

16.5  but  under  17.5 

30 

17.5  but  under  18.5 

18 

18.5  but  under  19.5 

5 

19.5  but  under  20.5 

1 

20.5  but  under  21.5 

1 

21  .5  but  under  22.5 

2 

Total    1000 

7.4  A  consumer  panel  report  on  the  economic  and  geographic  distribution 
of  the  purchases  of  a  particular  product  reveals  among  other  things 
that  the  nation's  families  bought,  on  the  average,  17.5  Vbs.  of  that 
product  in  1949,  This  estimate  was  based  on  returns  from  a  supposed 
random  sample  of  122*5  families,  and  the  standard  deviation  of  indi 
vidual  family  purchases  in  this  sample  was  found  to  be  7.5  lb&»  From 
sales  and  inventory  records,  it  is  determined  that  average  purchases 


PROBLEMS  1 45 

per  family  in  1948  must  have  been  at  least  18.5  lbs.;  or  1  pound  more 
than  the  sample  estimate  for  1949.  Could  this  difference  of  1  pound 
be  due  to  sampling  variation,  or  does  it  indicate  that  average  con 
sumption  of  the  product  by  families  had  decreased  in  1949  from  the 
1948  level  of  consumption?  What  assumptions  did  you  make? 

7.5  Using  the  data  in  Problem  6.1,  test  the  hypothesis  #:/*  =  6.55X10-27 
versus  A:  M  5^  6. 55X1 0-27.  Let  a>  =  0.01. 

7.6  Using  the  data  in  Problem  6.5  and  letting  <x  =  0.025,  test  H:fj,<l.I2 
inches  versus  A  ip,  >1.12  inches. 

7.7  Using  the  data  of  Problem  6.6  and  letting  a  =  0.05,  test  H:jji>1.55 
versus  A  :&  <1.55. 

7.8  Using  the  data  of  Problem  6.7  and  letting  «  =  0.25,  test  H:fj,<29QQ 
yards  versus  A  :/JL  >2900  yards. 

7.9  Using  the  data  of  Problem  6.1  and  letting  a:  =  0.01,  test  H:cr<0.0l 
X10-27  versus  A  :cr>0.01  X10~27. 

7.10  Using  the  data  of  Problem  6.5  and  letting  a:  =  0.10,  test  H  :cr*  <0.0001 
versus  A  :or2  >0.0001. 

7.11  Using  the  data  of  Problem  6.6  and  letting  ot.  =  0.005,  test  #:cr>0.05 
versus  A:cr<0.05. 

7.12  Using  the   data  of  Problem   6,7   and  letting  <x  =  0.01,   test  £T:o-  =  50 
yards  versus  A:cr^50  yards. 

7.13  In  making  a  certain  cross.,  a  geneticist  expected  a  segregation  of  15 
A's  to  1  B.  In  a  random  sample  of  800  he  observed  730  A's  and  70 
B's.  Do  the  data  support  the  expected  ratio?  Why? 

7.14  In  a  random,  sample  of  400  farm  operators,  65  per  cent  were  owners 
and  35  per  cent  were  nonowners.  Test  the  hypothesis  that  in  the  pop 
ulation  of  farm  operators  60  per  cent  are  owners.  Use  a  probability  of 
Type  I  error  equal  to  ,05. 

7.15  A  manufacturer  of  light  bulbs  claims  that  on  the  average  1  per  cent 
or  less  of  all  the  light  bulbs  manufactured  by  his  firm  are  defective. 
A  random  sample  of  400  light  bulbs  contained  12  defectives.  On  the 
evidence  of  this  sample,   do  you  believe  the   manufacturer's  claim? 
Why?  Assume  that  the  maximum  risk  you  wish  to  run  of  falsely  reject 
ing  the  manufacturer's  claim — the  true  fraction  defective  is  .01 — has 
been  set  at  2  per  cent. 

7.16  A  sampler  of  public  opinion  asked  400  randomly  chosen  persons  from 
some  specified  population  whether  they  favored  candidate  A  or  B] 
220  voted  for  A  and  180  for  B.  Using  a  probability  of  Type  I  error 
equal  to  .05,  do  you  think  that  opinion  in  the  population  may  have 
been  equally  divided?  Why? 

7.17  A  supermarket  is  to  be  built  in  a  new  location.  The  question  arose  as 
to  whether  provision  should  be  made  for  individual  customer  service 
at  the  meat  counter,  or  whether  a  self-service  counter  with  all  meats 
ready-cut  and  packaged  would  adequately  serve  customers  in  the  new 
area.  The  management  decision  was  that  individual  customer  service 
would  not  be  supplied  unless  40  per  cent  of  the  prospective  customers 
desired  such  service.  A  random  sample  of  160  prospective  customers 
showed  only  50  respondents  desiring  individual  service.  Does  it  appear 
that  the  proportion  of  preference  in  the  population   of  prospective 
customers  equals  or  exceeds  the  critical  level  set  by  management? 

7.18  In  a  triangular  test  for  selecting  judges  to  compose  a  taste  panel,  a 
prospective  judge  was  successful  in  selecting  the  odd  sample  11  times 


146  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

in  15  trials.  Would  you  select  him  for  the  panel?  How  many  would  he 
have  to  pick  correctly  to  be  chosen?  What  is  the  probability  of  Type  I 
error  if  we  accept  the  above  judge  for  our  panel?  Construct  the  com 
plete  table  of  probabilities,  showing  them  also  in  cumulative  form,  for 
n-^15. 

7.19  Retail  sales  data  indicate  that  |  of  the  families  in  the  WOI-TV  area 
have  television  sets.  A  random  sample  of  900  families  from  the  area  is 
to  be  taken,  (a)  What  is  the  expected  number  of  television  families  for 
the  sample?  (b)  The  sample  yields  360  families  with  television  sets. 
Indicate  at  least  two  methods  by  which  we  may  obtain  approximate 
confidence  limits  for  the  population  proportion  of  families  owning  tele 
vision  sets,  (c)  Is  the  observed  number,  360,  in  "reasonable"  agree 
ment  with  the  expected  number? 

7.20  Eighty  out  of  1000  randomly  chosen  cases  of  diphtheria  resulted  in 
death.    What   methods   or   techniques   are   available   for   using   these 
results  to  tost  the  hypothesis  that  the  true  percentage  of  fatality  is 
10  per  cent?  State  whether  the  tests  are  exact  or  approximate. 

7.21  A  botanist  observed  350  seedlings  for  the  purpose  of  studying  chloro 
phyll  inheritance  in  corn.  The  seed  came  from  self-fertilized  hetero 
zygous  green  plants.  Hence,  green  and  yellow  seedlings  were  expected 
in  proportions  of  3  green  to  1  yellow.  The  sample  showed  120  green 
and  30  yellow  seedlings.  Is  this  sample  in  agreement  with  expectation? 

7.22  A  metropolitan  newspaper  was  considering  a  change  to  tabloid  form. 
A  random  sample  of  900  of  its  daily  readers  was  polled  to  secure 
readership  reaction  to  such  a  change.   Of  this  sample,   541    persons 
opposed  the  change  in  format  for  the  paper,  (a)  Is  it  likely  that  more 
than  50  per  cent  of  the  readers  are  in  favor  of  the  change?  (b)  Describe 
two  or  more  procedures  for  obtaining  confidence  limits  for  the  popu 
lation  proportion  opposed  to  the  change. 

7.23  From  a  keg  containing  1000  bolts,  a  random  sample  of  20  bolts  has 
been  presented  to  you  for  testing.  One  hundred  per  cent  of  the  bolts 
in  the  sample  successfully  pass  the  test.  Of  all  the  bolts  in  the  keg, 
what  is  your  estimate  of  the  percentage  that  will  pass?  What  limits 
would  you  place  on  the  reliability  of  your  estimate;  that  is,  what  con 
fidence*,  statement  would  you  make  about  the  true  percentage  of  all 
the  bolts  that  will  pass  the  test? 

7.24  After  a  survey  of  opinion  is  made,  point  and  interval  estimates  are  cal 
culated.  The  investigator  states  that  the  95  per  cent  confidence  inter 
val  is  from  (>0  per  cent  to  75  per  cent  of  the  population  in  favor  of  a 
law.  Describe  precisely  the  moaning  of  this  statement. 

7.25  IMng  the  data  of  Problem  «.20  ami  letting  a  «=  0.005,  test  ff:^^<f^i 
versus  A  :/ia  >Mx- 

7*2Ci  lining  the  data  of  Problem  6.2)1  and  letting  a  »  0.05,  tost  //r^i— Ms 
versus  ^1  in\  *£&%. 

7.27  UniTig  the  data  of  Problem  0,22  and  letting  a«0.10,  test  //r/xi^Ma 
vorfcms  A  :/xi  y^Ma- 

7.2cS  Wo  are  told  that  the  moan  yields  of  two  corn  hybrids  wore  75  and  cS5 
buaholB  per  acre,  respectively,  and  that  each  had  boon  tritul  in  10 
fields  Holoctod  at  random  from  ftomo  population  of  fiolcln.  Further, 
uHHmning  that  cr'f  —  cri,  wo  are  told  that  the  ntamlard  error  of  each  of 
the*  above  means  wan  3.  Tost  the  hypothesis  that /*i -"/AS- 


PROBLEMS 


147 


7.29  The  diameter  of  a  cylinder  was  measured  by  16  persons.  Each  person 
made  three  determinations  using  a  micrometer  caliper  and  three 
determinations  using  a  vernier  caliper.  Following  are  the  averages  of 
the  three  determinations  (in  inches) ,  for  each  caliper,  made  by  the 
16  persons. 


Micrometer 

Vernier 

Micrometer 

Vernier 

Micrometer 

Vernier 

1.265 

1.265 

1.270 

1.269 

1.264 

1.267 

1.265 

1,267 

1.267 

1.273 

1.266 

1.272 

1.267 

1.267 

1.268 

1.270 

1.266 

1.273 

1.266 

1.266 

1.267 

1.270 

1.268 

1.267 

1.268 

1.267 

1.267 

1.267 

1.265 

1.268 

1.265 

1.267 

7.30 


Is  there  any  difference  between  the  means  of  the  populations  of  meas 
urements  represented  by  the  two  samples?  The  method  to  be  used  is 
determined  by  the  fact  that  each  person  used  both  calipers.  Do  you 
think  the  difference  is  attributable  to  imperfections  of  the  calipers  or 
to  the  difficulty  of  setting  the  vernier  caliper? 

The  following  are  the  lengths  in  millimeters  of  6-year-old  white  crap- 
pies  from  East  Lake,  Lucas  County,  Iowa,  in  1948.  Measurements 
were  made  by  William  Lewis  and  T.  S.  English. 


Males 


Females 


228 

217 

219 

231 

225 

219 

230 

'   217 

222 

214 

224 

220 

225 

220 

221 

225 

221 

228 

222 

233 

239 

225 

234 

222 

227 

223 

223 

222 

223 

234 

241 

223 

225 

253 

220 

233 

213 

224 

235 

281 

224 

212 

218 

235 

231 

231 

220 

224 

264 

251 

321 

223 

246 

247 

214 

241 

272 

Is  there  any  difference  between  the  lengths  of  male  and  female  crap- 
pies  of  this  age  group  in  East  Lake  in  1948? 

7.31  In  order  to  test  two  methods  of  teaching  spelling,  40  pupils  were  ran 
domly  assigned  to  two  classes  and  one  method  was  tried  on  each  class. 
At  the  end  of  the  trial  a  test  was  given.  Following  are  the  scores  on  the 
tests : 


148 


CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


Method 

A 

Method 

B 

10 

48 

20 

57 

20 

50 

27 

60 

25 

51 

35 

63 

30 

52 

40 

64 

33 

54 

41 

65 

37 

56 

50 

67 

41 

57 

50 

67 

43 

65 

54 

73 

46 

73 

56 

83 

46 

86 

57 

95 

Test  the  hypothesis  that  the  two  methods  of  teaching  spelling  are 
equally  effective.  State  all  your  assumptions. 

7.32  Using  the  data  of  Problem  6.23  and  letting  <x  =  0.01,  test  H:fj,D^Q 
versus  A  :&&  ?^0. 

7.33  A  certain  stimulus  administered  to  each  of  9  patients  resulted  in  the 
following  increases  in  blood  pressure:  5,  1,  8,  0,  3,  3,  5,  — 2,  4  mm.  Hg. 
Can  it  be  concluded  that  the  stimulus  will  be  in  general  accompanied 
by  an  increase  in  blood  pressure? 

7.34  Suppose  an  investigator  of  group  differences  in  I.Q.  finds,  for  inde 
pendent  random  groups  A  and  B  of  11  subjects  each  assumed  to  be 
from  normal  populations  of  same  variance,   a  difference  in  sample 
means  of 

TA  -  7*  «  3.9  I.Q.  points 

and  an  estimated  standard  error  of  the  mean  difference  of  2.0,  He 

selects  lOOcx  as  5  per  cent, 

(a)   For  the  data  as  given,  what  hypothesis  might  he  test?  Perform 

the  required  test  and  state  your  conclusions. 
(6)    Suppose  group  A  had  been  given  special  coaching  designed  to 

"increase"  I.Q,,  while  group  B  had  been  maintained  as  a  con- 

troL  What  hypothesis  might  he  test?  Perform  the  required  test 

and  give  the  resulting  inferences. 

7.35  In  examining  the  resistance  to  crushing  offered  by  kernels  of  a  single 
ear  of  corn,  we  choose  at  random  two  lots  of  10  kernels  each  with  the 
following  results: 

CRUSHING  RESISTANCE 
(in  points) 


Lot  I 

Lot  II 

8 

18 

8 

20 

14 

20 

15 

20 

16 

22 

16 

24 

17 

27 

18 

28 

18 

30 

20 

31 

PROBLEMS 


149 


Using  the  method  of  paired  observations,  we  find  the  difference  be 
tween  the  two  means  to  be  significant.  We  draw  four  more  sets  of  two 
samples,  each  time  with  a  significant  difference.  This  seems  surpris 
ing,  since  all  the  samples  were  taken  from  the  kernels  of  the  same  ear. 
Can  you  explain  the  results? 

7.36  (a)  For  the  data  given  below,  test  the  hypothesis  that  the  true  mean 
tensile  strength  of  the  product  of  the  C  and  C  Manufacturing 
Company  is  greater  than  the  corresponding  value  for  its  competi 
tor.  State  all  your  assumptions. 

(6)  Ignoring  any  assumption  about  variances  you  may  have  found  it 
necessary  to  make  in  (a)  above,  test  the  hypothesis  that  the  two 
population  variances  are  equal. 


TENSILE    STRENGTH    or    SCREW 
DRIVER  OP  34  VALVE  CAPS 
PRODUCED  BY  THE  C  AND  C 
MANUFACTURING  COMPANY 


Test 

Tensile 
Strength 
in 
Pounds 
Y 

Test 

Tensile 
Strength 
in 
Pounds 
Y 

1  

130.1 

19  

153.5 

2  

132.3 

20  

154.1 

3  

133.4 

21  

154.7 

4.  ... 

135.5 

22  

155.4 

5  

137.7 

23.  .  .  . 

156.7 

6  

139.3 

24  

157.5 

7  

140.4 

25  

158.4 

8  

144.2 

26  

159.4 

9  

145.0 

27.  .  .  . 

160.7 

10  

146.7 

28  

161.9 

11  

147.4 

29  

163.1 

12  

148.3 

30  

164.8 

13.  .  .  . 

149.7 

31  

169.3 

14.  ... 

150.6 

32  

171.2 

15.  .  .  . 

151.1 

33  

174.0 

16  

151.8 

34  

180.7 

1  *7 

1   CO      1 

1  /  .    .   .    . 

18  

I  3  £  .  X 

152.7 

Total 

5183.7 

TENSILE    STRENGTH    OF    SCREW 
DRIVER  OF  36  VALVE  CAPS 
PRODUCED  BY  A  COMPETITOR 
OF  THE  C  AND  C  MANUFAC 
TURING  COMPANY 


Test 

Tensile 
Strength 
in 
Pounds 
Y 

Test 

Tensile 

Strength 
in 
Pounds 
Y 

1.  ... 

65.7 

20  

149.4 

2.  ... 

101.3 

21  

151.0 

3  

103.0 

22  

153.3 

4  

103.6 

23.  ... 

155.2 

5  

107.2 

24  

157.6 

6  

115.9 

25  

160.7 

7  

117.4 

26  

164.3 

8.  ... 

122.6 

27.  ... 

166.1 

9  

126.5 

28  

168.8 

10  

129.1 

29.  ... 

170.4 

11  

132.3 

30.  ... 

180.6 

12.  ... 

134.6 

31.  ... 

184.6 

13  

135.2 

32  

188.8 

14.  .  .  . 

136.7 

33  

192.9 

15  

138.3 

34  

196.0 

16.  .  .  . 

142.1 

35  

200.4 

17  

143.4 

36  

204.8 

18  

147.2 

19.  .  .  . 

148.2 

Total 

5295.2 

7.37  Using  the  data  which  follow,  test  the  hypothesis  that  the  true  mean 
crushing  strengths  of  air-dried  and  green  Douglas  fir  wood  are  the  same. 
State  all  your  assumptions  and  interpret  your  results. 


ISO 


CHAPTER   7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


CRUSHING  STRENGTHS  OF  248  SAMPLES  OF  AIR-DRIED  DOUGLAS  FIR,  SIZE 
2"  BY  2"  BY  8".  TESTED  BY  FOREST  PRODUCTS  LABORATORY  DOMINION 
GOVERNMENT,  AT  UNIVERSITY  OF  BRITISH  COLUMBIA,  1945 


N  1 

4713 

N  1 

5641 

N  9 

7145 

E  7 

6508 

S  5 

8413 

W  3 

7446 

2 

5516 

2 

5550 

E  3 

6200 

8 

6828 

6 

7690 

5 

7941 

3 

5956 

3 

7433 

4 

7501 

9 

6098 

7 

8484 

6 

8159 

4 

5652 

4 

7097 

5 

8086 

10 

6359 

8 

8139 

7 

9316 

5 

5951 

5 

7865 

6 

8055 

S  1 

5208 

9 

7595 

8 

9515 

6 

7178 

6 

8045 

7 

8042 

2 

4648 

10 

7021 

9 

8171 

7 

6630 

7 

7408 

8 

8678 

3 

7153 

11 

6416 

10 

9001 

8 

6284 

8 

7344 

9 

6710 

4 

6504 

W  3 

6657 

N  3 

8161 

9 

6246 

9 

7518 

10 

7512 

5 

6562 

5 

8264 

4 

7820 

10 

4689 

10 

7280 

S  1 

6438 

6 

7105 

7 

7268 

6 

8560 

11 

4825 

E  4 

7174 

3 

6074 

7 

7114 

8 

8101 

7 

8222 

12 

4697 

5 

7234 

4 

7170 

8 

6263 

9 

7066 

8 

8387 

E  3 

5757 

6 

8452 

5 

7306 

W  3 

5530 

10 

7301 

9 

7500 

4 

6661 

7 

8709 

6 

7760 

4 

6632 

N  1 

5961 

10 

2181 

5 

6098 

8 

7710 

7 

7049 

5 

6429 

2 

6254 

11 

7655 

6 

5867 

9 

7609 

8 

6863 

6 

6912 

3 

7247 

E  3 

7373 

7 

5573 

10 

6731 

9 

6987 

7 

7053 

4 

7480 

4 

7949 

8 

6282 

S  3 

6342 

10 

6511 

8 

6370 

5 

8512 

5 

8199 

9 

5536 

4 

6924 

W  3 

7025 

9 

7413 

6 

8911 

6 

8547 

10 

4941 

5 

7712 

4 

6775 

10 

6335 

7 

8988 

7 

8464 

S  2 

4003 

6 

6805 

5 

7754 

N  1 

6584 

8 

9330 

8 

8594 

3 

4789 

7 

7539 

6 

7495 

3 

7518 

9 

9899 

9 

7092 

4 

4889 

8 

7630 

7 

7990 

4 

7106 

10 

9025 

10 

7433 

5 

5304 

9 

7501 

8 

6149 

5 

7135 

11 

8920 

S  1 

6444 

6 

5350 

10 

7531 

9 

6774 

6 

7596 

E  3 

6419 

2 

6545 

7 

5601 

11 

6096 

10 

7137 

7 

7573 

4 

8403 

3 

7320 

8 

5932 

12 

6983 

N  1 

4858 

8 

7521 

5 

8220 

4 

7886 

9 

5245 

W  3 

6212 

3 

6148 

9 

7261 

6 

9501 

5 

8173 

10 

5585 

4 

6530 

4 

5388 

10 

6364 

7 

9250 

6 

7844 

11 

4313 

5 

7800 

5 

5883 

11 

6905 

8 

9479 

7 

7613 

12 

4924 

6 

7713 

6 

5930 

E  3 

7608 

9 

9985 

8 

8469 

W  3 

5196 

7 

7759 

7 

6252 

4 

6793 

10 

9686 

9 

7675 

4 

4810 

8 

7253 

8 

5920 

5 

7734 

11 

8849 

10 

7371 

5 

6641 

9 

6898 

9 

6260 

6 

6465 

S  3 

6693 

W  3 

7113 

6 

4625 

10 

7403 

10 

6403 

7 

7499 

4 

6338 

4 

7283 

7 

6704 

N  2 

6144 

11 

6644 

8 

7703 

5 

5976 

5 

8337 

8 

5555 

3 

6717 

12 

5841 

9 

7470 

7 

8495 

6 

8509 

9 

6813 

4 

7021 

E  3 

6650 

10 

7178 

8 

9184 

7 

7510 

10 

6061 

5 

8096 

4 

5802 

S  1 

6201 

9 

9485 

8 

8361 

11 

4959 

6 

7608 

S 

7287 

3 

7878 

11 

8507 

9 

7485 

12 

5618 

7 

8025 

6 

6379 

4 

7155 

12 

8270 

10 

8522 

14 

3958 

8 

8115 

PROBLEMS 


151 


CRUSHING  STRENGTHS  OF  248  SAMPLES  OF  GREEN  DOUGLAS  FIR,  SIZE  2"  BY 

2"  BY  8".  TESTED  BY  FOREST  PRODUCTS  LABORATORY  DOMINION 

GOVERNMENT,  AT  UNIVERSITY  OF  BRITISH  COLUMBIA,  1945 


N  1 

2428 

W13 

2343 

N  7 

3639 

E  3 

3446 

S  3 

3412 

S  7 

4088 

2 

2173 

N  2 

2603 

8 

3645 

4 

2892 

4 

3904 

8 

4377 

3 

2896 

3 

2911 

9 

3487 

5 

3629 

5 

4030 

9 

4267 

4 

2980 

4 

3158 

10 

3351 

6 

3442 

6 

4212 

10 

4256 

5 

3378 

5 

3553 

E  3 

3591 

7 

3412 

7 

4423 

11 

4109 

6 

3167 

6 

3659 

4 

2849 

8 

3477 

8 

4575 

12 

3325 

7 

3208 

7 

3800 

5 

3911 

9 

3474 

9 

4318 

W  3 

3297 

8 

3342 

8 

3645 

6 

2591 

10 

3007 

10 

3829 

4 

3606 

9 

2982 

9 

3505 

7 

2769 

S  1 

2493 

11 

3933 

5 

3534 

10 

3301 

10 

3834 

8 

4097 

2 

2505 

12 

4608 

6 

4159 

11 

2330 

E  3 

2506 

9 

3203 

3 

3449 

W  4 

3340 

7 

4393 

12 

2651 

4 

2818 

10 

3179 

4 

3224 

5 

3887 

8 

3992 

E  3 

2478 

5 

3775 

S  1 

2668 

5 

3485 

6 

4097 

9 

4049 

4 

2665 

6 

3318 

2 

2766 

6 

3667 

7 

3440 

N  1 

2813 

5 

3033 

7 

3686 

3 

3280 

7 

3343 

8 

4503 

2 

2574 

6 

3205 

8 

3705 

4 

3295 

8 

3431 

9 

3806 

3 

3286 

7 

3282 

9 

3543 

5 

3844 

W  3 

2643 

10 

3939 

4 

3310 

8 

3229 

10 

3848 

6 

4022 

4 

3039 

N  1 

2902 

5 

3610 

9 

3137 

S  3 

2778 

7 

3575 

5 

3510 

2 

2869 

6 

3637 

10 

2693 

4 

2743 

8 

3784 

6 

3469 

3 

3610 

7 

3871 

S  1 

2128 

5 

3541 

9 

3621 

7 

3635 

4 

3547 

8 

3757 

2 

2200 

6 

3580 

11 

3698 

8 

4016 

5 

4012 

9 

3716 

3 

1977 

7 

3803 

W  3 

3032 

9 

3777 

6 

3919 

E  3 

3105 

4 

2498 

8 

3787 

4 

3132 

10 

3642 

7 

4585 

4 

3172 

5 

2732 

9 

3623 

5 

3781 

N  3 

3257 

8 

4553 

6 

3679 

6 

2920 

10 

3848 

6 

4141 

4 

3426 

9 

4235 

7 

3854 

7 

3102 

11 

3530 

7 

3730 

5 

4001 

10 

4495 

8 

3670 

8 

3050 

12 

3296 

8 

4162 

6 

3993 

11 

3694 

9 

3386 

9 

3230 

W  3 

2845 

9 

3559 

7 

4201 

12 

3492 

10 

3368 

10 

3053 

4 

3015 

10 

3532 

8 

4555 

E  3 

3173 

S  1 

2688 

11 

2993 

5 

3384 

N  1 

2296 

9 

3914 

4 

3879 

3 

3089 

12 

2518 

6 

3671 

2 

2458 

10 

3931 

5 

3751 

4 

3212 

W  3 

2938 

7 

3794 

3 

2794 

E  3 

3769 

6 

4197 

5 

3618 

4 

2272 

8 

3863 

4 

3075 

4 

3622 

7 

4110 

7 

3551 

5 

3144 

9 

3712 

5 

3166 

5 

4168 

8 

4061 

8 

3752 

6 

2904 

10 

3553 

6 

3255 

6 

4246 

9 

4589 

9 

3474 

7 

3314 

N  1 

2607 

7 

3233 

7 

4282 

10 

3762 

10 

3556 

8 

3448 

2 

2591 

8 

3600 

8 

4118 

11 

2733 

W  3 

3181 

9 

3468 

3 

3042 

9 

3471 

9 

3928 

S  4 

3071 

4 

3163 

10 

3289 

4 

2450 

10 

3735 

S  1 

3095 

5 

3886 

9 

3733 

11 

2456 

5 

3444 

11 

3329 

2 

3218 

6 

3873 

10 

3823 

12 

3078 

6 

3593 

7,38  Two  lots  of  steers,  10  head  in  each  lot,  were  used  in  a  90-day  feeding 
trial.  Lot  1  received  standard  ration  A*  Lot  2  received  special  ration 
K.  Steers  on  ration  A  gained  1.84  Ibs.  per  head  per  day,  while  the  ani 
mals  fed  K  gained  at  the  rate  of  2.36  Ibs.  per  head  per  day.  Two 
questions  were  of  interest, 
(a)  Will  daily  gains  on  ration  K  exceed  2  Ibs.  per  day?  The  variance 

of  the  mean  gain,  2.36,  was  found  to  be  .0144. 
(6)    Is  ration  K  better  than  standard  ration  A  in  producing  gains? 

The  variance  of  the  mean  gain,  1.84,  for  lot  1  was  .0256;  thus, 


152 


CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


7.39 


7.40 


7.41 


we  see  that  the  pooled  sum  of  squares  for  daily  gain  of  the  two  lots 

is  3.60, 

Answer  the  two  questions  with  the  information  given  above.  Why  do 
we  use  twice  the  pooled  variance  in  examining  the  difference  in  gains 
between  the  two  lots,  whereas  in  answering  question  (a)  we  use  the 
variance  without  such  modification? 

A  sample  of  rural  families  and  of  urban  families  was  taken  to  study 
differences  in  coffee  purchases  by  the  two  groups.  The  data  obtained 
are  listed  below  in  terms  of  pounds  per  family  purchased  annually. 


Family  No. 

Rural 

Urban 

1  

12.1 

8.3 

2.     ... 

6.8 

9.3 

3 

9.1 

9.2 

4  

11.1 

11.1 

5 

11.4 

10   7 

6.  . 

13.3 

4.6 

7  

9.8 

9.8 

8  

11.3 

7,9 

9       

9.4 

8.5 

10 

10.2 

9.1 

11  

9.7 

12  

6.2 

Would  you  attribute  the  difference  in  coffee  consumption  observed  in 
those  samples  to  normal  sampling  fluctuation,  or  is  there  a  real  dif 
ference  between  rural  and  urban  coffee  consumption?  Select  your  own 
level  for  control  of  the  Type  I  error  and  draw  your  conclusion  accord 
ingly*  What  is  the  specified  population  from  which  these  data  provide 
you  a  sample?  State  your  assumptions. 

It  has  been  suggested  that  the  resistance  of  wire  C  is  greater  than  the 
resistance  of  wire  D.  The  following  data  (in  ohms)  were  obtained  from 
tests  made  on  samples  of  each  wire : 


C 

D 

0.140 

0.135 

0.138 

0.  140 

0.143 

0.142 

0.142 

0.136 

0.144 

0.137 

0.139 

Assuming  that  trc*3*^?  test  (using  a«=0.0l)  the  hypothesis 
H:^c<fjL0  against  the  alternative  Aifj.c>fMI>.  State  your  conclusion 
and  interpret  the  results. 

If  the  estimate  of  the  population  standard  deviation  from  one  sample 
of  45  is  12,  and  a  corresponding  estimate  from  another  sample  of  45 
is  18,  arc  these  samples  consistent  with  the  hypothesis  that  they  are 
from  normal  populations  with  the  same  variance? 


PROBLEMS  153 

7.42  Two  methods  of  determining  moisture  content  of  samples  of  canned 
corn  have  been  proposed  and  both  have  been  used  to  make  determi 
nations  on  portions  taken  from  each  of  21  cans.  Method  I  is  easier  to 
apply  but  appears  to  be  more  variable  than  Method  II.  If  the  varia 
bility  of  Method  I  were  not  more  than  25  per  cent  greater  than  that  of 
Method  II,  we  would  prefer  Method  I,  Based  on  the  following  sample 
results,  which  method  would  you  recommend? 

ni  =  n2  =  21,  Pi  =  50,  F2  =  53,  &?=720,  S  2/1  =  340.  (Hint:  Test 
H :  af  =  1 .25<ri  against  A  :<ri>l. 25o-f .  Under  this  hypothesis 
(sf/1.25)/sl  is  distributed  as  F(VjV),  where  ^1=^2  =  ^1  —  I=n2 
—  1=20.) 

7.43  The  amount  of  surface  wax  on  each  side  of  waxed  paper  bags  is  be 
lieved  to  be  normally  distributed.  However,  there  is  reason  to  believe 
that  there  is  greater  variation  in  the  amount  on  the  inner  side  of  the 
paper  than  on  the  outside.  A  sample  of  25  observations  of  the  amount 
of  wax  on  each  side  of  these  bags  was  obtained  and  the  following  data 
recorded: 


Wax  in  Pounds  per  Unit  Area  of  Sample 


Outside  surface 

Inside  surface 

^=0,948 
]CX2-=91 

7  =  0.652 

2:1^=82 

Conduct  a  test  (using  a:  =  0.05)  of  the  hypothesis  HKTQ^O^  against 
the  alternative  A:o%  <cr|. 

7.44  Using  the  data  of  Problem  6,21   and  letting  a.  =  0.05,   test  H:o-1=cr2 
versus  A  :<ri  7^0-2. 

7.45  Using  the  data  of  Problem  6,22  and  letting  a:  =  0.01,   test  .Z7:crj.  =  cr2 
versus  A  :<ri  r^cr^ 

7.46  Using  the  data  of  Problem.  6.20  and  letting  01  =  0.01,  test  H:cri<cr2 
versus  A  :o-\  >crz. 

7.47  Using  the  data  of  Problem  6.23  and  letting  oi  =  0.05,  test  H:&A  <crB 
versus  A  :cr^  ><rB. 

7.48  A  child  psychologist,  analyzing  personality  differences  in  children  by 
a  protective  technique,  classified  the  responses  of  a  group  of  99  pre 
school  children  into  three  major  types:  static  form  of  response,  23; 
outer  activity,  51;  inner  activity,  25.  Do  these  data  differ  significantly 
from  a  chance  distribution  of  responses?  Use  a;  ==.01. 

7.49  A  random  sample  of  147  women  college  students  were  interviewed 
with  regard  to  their  habits  concerning  the  purchase  of  clothing.  The 
source  of  each  individual's  income   was  also  determined.   Given  the 
data  below  and  letting  a.  =  0.10,  test  the  hypothesis  that  women  pur 
chase  clothing  without  planning  in  the  following  proportion: 

Frequently 10  per  cent 

Seldom 80  per  cent 

Never 10  per  cent 


154 


CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


Source  of  Income 

Numbers  Who  Purchased  Clothing  Items 
Without  Planning 

Frequently 

Seldom 

Never 

Earned  all  of  spending  money.  .  . 
Earned  part  of  spending  money  . 
Had  regular  allowance  , 

2 
8 
4 
15 

14 
17 
12 
25 

27 
5 
7 
11 

Money  given,  as  needed  

7.50 


7.51 


7.52 


7.53 


Referring  to  the  data  of  Problem  7,49  and  letting  oc  =  0.01,  test  the 
hypothesis  that  frequency  of  purchasing  clothing  items  without  plan 
ning  is  independent  of  source  of  income. 

An  experimenter  testing  three  chemical  treatments  applied  each  to 
200  randomly  selected  seeds,  and  then  conducted  germination  tests. 
The  following  results  were  obtained: 


Number 


Chemical 

Germinating 

Not 
Germinating 

A.  

190 

10 

B  

170 

30 

C  

180 

20 

Test  the  hypothesis  that  the  percentage  of  seeds  germinating  is  inde 
pendent  of  the  chemical  used. 

An  experimenter  feel  different  rations  to  three  groups  of  chicks.  As 
sume  that  the  chicks  were  assigned  to  the  rations  (groups)  at  random 
and  that  all  other  management  practices  for  the  three  groups  were  the 
same.  A  record  of  mortality  is  given  below.  Would  you  attribute  the 
differences  among  the  mortality  rates  of  the  three  groups  to  rations? 
Why? 


Ration 

Number 

Lived 

Died 

A  

87 
94 
89 

1$ 
6 
11 

B  

C  

I  selected  a  random  sample  of  students  at  Arizona  State  University 
and  asked  their  opinions;  on  a  proponed  radio  program.  The  results  arc 
given  below.  The  same  number  of  each  sex  waa  included  within  each 
class  group,  that  is,  freshmen  and  sophomores  each  consisted  of  100 
men  and  100  women,  while  juniors  and  seniors  each  consisted  of  50 


PROBLEMS 


155 


men  and  50  women.  Test  the  hypothesis  tnat  opinions  are  independent 
of  the  class  groupings. 


7.54 


7.55 


7.56 


Num 

ber 

Class 

Favoring 
Program 

Opposed  to 
Program 

Freshmen          

120 

80 

Sophomores  

130 

70 

Tuniors                   

70 

30 

Seniors                      .... 

80 

20 

An  agency  engaged  in  market  research  conducted  some  of  its  sampling 
by  mail.  For  one  survey  these  results  in  terms  of  response  to  succes 
sive  mailings  were  obtained: 


Response  No.:          1st 


2nd 


3rd 


4th        Original    Mailing 


Returns: 


150 


60 


40 


20 


1000 


Another  agency  obtained  the  following  results  in  a  mail  sampling  of  a 
similar  population: 


Response  No,:          1st 


2nd 


3rd 


4th        Original    Mailing 


Returns: 


200 


30 


50 


25 


800 


Does  it  appear  that  the   two   mail   samplings  were  homogeneous  in 
eliciting  replies  from  the  two  populations? 

In  a  large  city  the  division  of  the  voting  strength  between  two  candi 
dates  for  mayor  appeared  to  be  about  equal.  The  campaign  manager 
for  candidate  A  polled  a  random  sample  of  2500  voters  two  weeks 
before  the  election.  In  this  sample  1313  of  the  voters  indicated,  they 
would  vote  for  A.  If  the  sample  is  representative  of  the  population  of 
voters  in  this  city,  is  it  likely  that  A  will  be  elected?  Establish  99  per 
cent  confidence  limits  for  the  proportion  of  voters  favoring  A.  ^ 
An  opinion-polling  agency  reported  the  distribution  of  a  sample  in  tins 
manner : 


Republicans 


Democrats 


Independents 


Total 


400 


450 


150 


1000 


A  newspaper  poll  in  the  same  area  yielded  this  distribution  in  terms 
of  declared  political  opinion  of  respondents :  


Republicans 
300 


Democrats  Independents  Total 

325  75  700 


156 


CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 


Are  these  two  samples  homogeneous  with  regard  to  division  of  politi 
cal  opinion? 

7.57      The  following  data  were  obtained  from  a  random  sampling  of  the  rec 
ords  of  a  specific  company. 

NUMBER  OF  BREAKDOWNS 


7.58 


7.59 


7.60 


7.61 


7.62 


Machine 

A 

B 

C 

D 

Total  per  Shift 

Shift  1 

10 

6 

12 

13 

41 

Shift  2 

10 

12 

19 

21 

62 

Shift  3 

13 

10 

13 

18 

54 

Total 
per  machine 

33 

28 

44 

52 

157 

Test  the  hypothesis  that  the  number  of  breakdowns  on  each  machine 
is  independent  of  the  shift.  Use  ct  =  0.05. 

Road  tests  gave  the  data  shown  below  regarding  tire  failures.  Letting 
<x  =  0.05,  test  the  hypothesis  that  left-right  tire  wear  is  independent  of 
front-rear  tire  wear. 

NUMBER  or  FAILURES 


Front 

Rear 

Totals 

Left 
Right 

115 
125 

65 
95 

180 
220 

Totals 

240 

160 

400 

A  car  rental  firm  has  a  particular  car  that  has  experienced  13  break 
downs  in  the  past  year.  Using  a  Poisson  distribution  and  letting 
<x  =  0.01,  test  H:jj,<lQ  versus  ^L:M>10. 

Ignoring  the  correction  for  continuity  in  Equation  (7.10),  that  is, 
dropping  the  adjustment  of  — 0.5,  show  that  x2  =  (a  —  rb)2/r(a  +  b) 
where  a  and  b  are  the  observed  numbers  in  the  two  classes  and  r 
equals  the  hypothesized  ratio  of  type  A  to  type  B. 

Work  the  preceding  problem  using  the  correction  for  continuity  and 
show  thatx2=(|a~-r&j  —  (r  +  l)/2)2/V(>+&). 

Rework  the  problems  noted  below,  using  the  method  described  in 
Section  7.20: 


(a)  7.26 

(6)  7.27 

(c)  7.28 

W)  7.30 


(e)  7.31 
(/)  7.37 
(<7)  7.39 


7.63     Given  the  following  data  (three  random  samples  from  three  normal 


PROBLEMS 


populations)    and    assuming    homogeneous    variances,    test    the    hy 
pothesis  jHr:jui=/z2=M3.  Let  o:  =  0.10. 


Sample  1  Sample  2  Sample  3 


48 

72 

48 

24 

24 

12 

36 

48 

24 

48 

7.64  Assuming  homogeneous  variances,  test  the  hypothesis  that  the  four 
normal  populations,  from  which  the  following  random  samples  were 
obtained,  have  the  same  mean.  Let  c*  =  0.025. 

Sample  1  Sample  2  Sample  3  Sample  4 


95 

45 

95 

20 

50 

40 

130 

55 

105 

95 

15 

50 

10 

65 

135 

80 

60 

45 

125 

7.65  Using  the  data  of  Table  7.23  and  assuming  homogeneous  variances, 
test  the  hypothesis  H:^i=jU2==M3=M4-  Let  a.  =  0.01. 

7.66  Using  the  data  of  Table  7.20,  test  the  hypothesis  H:<r?  =<r|  =o-§  ==oi. 
Let  <*=«0.05. 

7.67  Letting  OL  =  0.10,  test  the  hypothesis  of  homogeneous  variances  for 
each  of  the  following  problems:  (a)  7.63  and  (b)  7.64. 

References  and  Further  Reading 

1.  Anderson,  R.  L.,  and  Bancroft,  T.  A.  Statistical  Theory  in  Research.  McGraw- 
Hill  Book  Company,  Inc.,  New  York,  1952. 

2.  Bartlett,  M.  S.  Some  examples  of  statistical  methods  of  research  in  agricul 
ture  and  applied  biology.  Jour.  Roy.  Stat.  Soc.  (Suppl.),  4:137,  1937. 

3.  Bowker,  A.  H.,  and  Lieberman,  G.  J.  Engineering  Statistics.  Prentice-Hall, 
Inc.,  Englewood  Cliffs,  N.J.,  1959. 

4.  Brownlee,  K.  A.  Statistical  Theory  and  Methodology  in  Science  and  Engineer 
ing.  John  Wiley  and  Sons,  Inc.,  New  York,  1960. 

5.  Davies,  O.  L.   (editor).  The  Design  and  Analysis  of  Industrial  Experiments. 
Second  Ed.  Oliver  and  Boyd,  London,  1956. 

6.  Dixon,  W.  J.,  and  Massey,  F.  J.  Introduction  to  Statistical  Analysis.  Second 
Ed.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1957. 

7.  Freund,  J.  E.  Modern  Elementary  Statistics.  Second  Ed.  Prentice-Hall,  Inc., 
*  Englewood  Cliffs,  N.J.,  1960. 

g    9  Liver  more,  P.  E.,  and  Miller,  I.  Manual  of  Experimental  Statistics. 

Prentice-Hall,  Inc.,  Englewood  Cliffs,  N.J.,  1960. 

9.   Hald,  A.  Statistical  Theory  with  Engineering  Applications.  John  Wiley  and 
Sons,  Inc.,  New  York,  1952. 

10.  .  Statistical  Tables  and  Formulas.  John  Wiley  and  Sons,  Inc., 

York,  1952, 


158  CHAPTER    7,    STATISTICAL    INFERENCE:    TESTING    HYPOTHESES 

11.  Huntsbcrger,  D.  V.  Elements  of  Statistical  Inference.  Allyn  and  Bacon,  Inc., 
Boston,  1961. 

12.  Mood,    A.    M.    Introduction   to   the    Theory    of  Statistics,    McGraw-Hill    Book 
Company,  Inc.,  New  York,  1950. 

13.  Robertson,   W.   H,   Programming  Fisher's   exact   method  of  comparing  two 
percentages.  Technometrics,  2  (No.l):  103-7,  Feb.,  1960. 

14.  Snedecor,   G.   W.  Statistical  Methods.  Fifth  Ed.   The  Iowa  State  University 
Press,  Ames,   1956. 

15.  Statistical    Research    Group,    Columbia    University.    Selected    Techniques    of 
Statistical   Analysis.    (Edited   by   C,    Kisenhart,    M.    W.    Hastay,    and    W,    A. 
Wallis.)  McGraw-Hill  Book  Company,  Inc.,  New  York,   1947. 

16.  Wadsworl.li,  Cr.  P.,  arid  Bryan,  J.  CT.  Introduction  to  Probability  and  Random 
Variables.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1900. 


C  H  APTE  R    8 

REGRESSION  ANALYSIS 

THE  METHODS  OF  ANALYSIS  studied  thus  far  in  this  text  have  been  con 
cerned  with  data  on  only  one  characteristic  associated  with  the  experi 
mental  units.  That  is,  in  any  given  problem  we  have  been  working  with 
only  one  variable.  However,  as  you  will  realize,  many  problems  involve 
more  than  one  variable.  Consequently,  it  is  necessary  that  techniques 
developed  for  analyzing  multivariate  problems  be  studied.  Some  of 
these  techniques  will  be  investigated  in  this  chapter. 

8.1      FUNCTIONAL   RELATIONS  AMONG  VARIABLES 

When  we  possess  information  on  two  or  more  related  (or  concomitant} 
variables,  it  is  natural  to  seek  a  way  of  expressing  the  form  of  the 
functional  relationship.  In  addition,  it  is  desirable  to  know  the  strength 
of  the  relationship.  That  is,  not  only  do  we  seek  a  mathematical  func 
tion  which  tells  us  how  the  variables  are  interrelated,  but  also  we  wish 
to  know  how  precisely  the  value  of  one  variable  can  be  predicted  if  we 
know  the  values  of  the  associated  variables.  The  techniques  used  to 
accomplish  these  two  objectives  are  known  as  regression  methods  and 
correlation  methods.  Regression  methods  are  those  used  to  determine  the 
"best"  functional  relation  among  the  variables,  while  correlation 
methods  are  used  to  measure  the  degree  to  which  the  different  variables 
are  associated. 

More  specific  statements  will  be  forthcoming  in  succeeding  sections. 
For  the  moment,  it  will  suffice  to  say  that  the  functional  relationships 
will,  in  general,  be  represented  mathematically  by 

•  •  •  ,  Xp|0x,  •  •  •  ,O  (8.1) 


where 

77  =  the  response  (or  dependent}  variable 
Jft  =  the  ith  independent  variable  (i=l,  -  -  -  ,  p), 

0j=the  jfch  parameter  in  the  function  (j  =  1,  •••,?), 
and  <t>  stands  for  the  assumed  form  of  the  function.  Equation  (8.1)  is 
sometimes  written  as 

77  =  <£(^i,  •  •  •  ,  X*).  (8.2) 

When  this  abbreviated  form,  is  used,  one  should  always  remember  that 
the  parameters  belong  in  the  expression;  they  have  been  omitted  solely 
to  achieve  brevity.  In  the  language  of  statistics,  a  function  such  as 
specified  by  Equation  (8.1)  is  known  as  a  regression  function.  However, 
in  some  areas  of  application,  a  more  natural  (and  common)  expression 

[159] 


16O  CHAPTER    8r    REGRESSION    ANALYSIS 

is  response  function.  In  this  book  both  expressions  will  be  used,  the 
choice  being  dictated  by  the  context. 

8.2  A  WORD  OF   CAUTION   ABOUT    FUNCTIONAL    RELA 
TIONS 

In  any  analysis  it  is  hoped  that  the  postulated  (assumed)  function 
represents  some  basic,  or  causal,  mechanism  associated  with  the  ex 
perimental  units  and  the  factors  under  investigation.  However,  science 
is  not  always  so  far  advanced  that  the  basic  variables  and  the  basic 
mechanisms  of  a  process  are  known  with  certainty.  In  such  cases,  the 
methods  of  regression,  and  correlation  may  still  prove  useful  as  analytic 
and  predictive  tools. 

Because  of  the  frequent  uncertainty  about  basic  variables  and  basic 
mechanisms,  a  word  of  warning  must  be  sounded  relative  to  the  inter 
pretation  of  analyses  involving  concomitant  variables.  This  warning 
is:  J^lst  because  a  particular  functional  relationship  has  been  assumed 
and  a  specific  computational  procedure  followed,  do  not  assume  that  a 
causal  relationship  exists  among  the  variables.  That  is,  because  a  func 
tion  has  been  found  that  is  a  good  fit  to  a  set  of  observed  data,  we 
are  not  necessarily  in  a  position  to  infer  that  a  change  in  one  variable 
causes  a  change  in  another  variable. 

In  summary,  the  only  person  who  can  safely  say  that  the  basic 
variables  are  those  used  and  that  the  basic  mechanism  operates  in 
accordance  with  the  selected  mathematical  function  is  a  pcrsoxi  well 
trained  in  the  subject  matter  field  in  which  the  experiment  was  per 
formed.  The  statistical  analysis  (in  this  instance,  a  regression  and/or 
correlation  analysis)  is  only  a  tool  to  aid  him  in  the  analysis  and  inter 
pretation  of  data. 

8.3  THE  CHOICE  OF  A   FUNCTIONAL    RELATION 

How  does  an  analyst  go  about  choovsmg  a  particular  functional  rela 
tionship  as  representative  of  the  population  under  investigation?  Two 
methods  arc  employed.  Those  are:  (1)  mi  analytical  consideration  of 
the  phenomenon  concerned,  and  (2)  an  examination  of  scatter  diagrams 
plotted  from  the  observed  data.  While  the  first  method  is  preferred, 
the  second  should  not  be  underrated.  If  little  is  known  about  the  basic 
mechanisms  involved,  the  use  of  scatter  diagrams  can  be  quite  helpful. 

8-4     CURVE   FITTING 

Once  we  have  decided  on  the  type  of  mathematical  function  that 
best  seems  to  fit,  or  represent,  our  concept  of  the  exact  relationship 
existing  among  the  variables,  the  problem  of  choosing  a  particular 
member  of  this  family  of  functions  arises.  That  is,  a  certain  function 
has  been  postulated  as  being  an  expression  of  the  true  state  of  affairs 
in  the  population,  and  it  is  now  necessary  to  estimate  the  parameters 
of  this  function.  The  determination  of  these  estimates  and  thus  the 
specification  of  a  particular  function  is  commonly  referred  to  as  curve 
fitting. 


8.5       THE    METHOD    OF    LEAST   SQUARES  161 

How  do  we  go  about  fitting  a  curve  to  a  set  of  data?  That  is,  how  are 
the  estimates  of  the  parameters  obtained?  Again  we  are  faced  with  the 
problem  of  choosing  among  several  methods  of  estimation.  The  ap 
proach  taken  should,  of  course,  provide  us  with  the  "best77  estimates, 
Since  this  is  simply  another  part  of  the  general  problem  of  estimation, 
the  criteria  by  which  estimators  are  selected  will  be  similar  to  those 
outlined  in  Chapter  6. 

8.5      THE    METHOD   OF    LEAST  SQUARES 

To  proceed  to  the  method  of  estimation,  it  should  be  noted  that 
there  are  several  methods  outlined  in  the  literature,  all  of  which  give 
acceptable  answers.  For  our  purposes  it  will  be  sufficient  to  discuss  the 
method  of  least  squares.  By  the  use  of  this  method  excellent  results  may 
be  obtained.  In  fact,  if  the  usual  assumption  of  normality  is  made,  the 
method  of  least  squares  becomes  equivalent  to  that  of  maximum 
likelihood. 

To  study  the  method  of  least  squares,  assume  that  we  are  consider 
ing  a  certain  characteristic  (77)  which  is  related  to  or  depends  on  certain 
other  characteristics  (X\,  •  •  -  ,  X^)  according  to  the  relationship 


97  =  <^(-X"i,  •  -  -  ,  2£p\  0i,  •  •  •  ,  6g).  (8.3) 

Both  the  form  of  the  function  and  the  values  of  the  parameters  must 
be  determined.  (NOTE:  In  practice,  the  form  is  usually  assumed  to  be 
known  and  thus  the  problem  reduces  to  estimating  the  parameter 
values.) 

The  reason  that  the  parameter  values  cannot  be  determined  without 
error  is  that  the  observed  values  of  the  dependent  variable  will  seldom 
agree  with  the  expected  values.  That  is,  even  if  we  can  control  the  X 
values  (or  measure  them  without  error)  the  observed  value  of  the 
dependent  variable,  denoted  by  Y,  will  not  equal  the  expected  value,  77. 
This  is  expressed  by 

Y  «  n  +  e  =  *(Xi,  ••-,  X,  |  0lf  •••,*,)+  e  (8.4) 

where  6  stands  for  the  error  made  in  attempting  to  observe  77.  Many 
factors  contribute  to  the  value  of  e,  but  it  seems  reasonable  to  as 
sume  that  it  (c)  is  a  random  variable  with  mean  0  and  variance  o-^. 
Under  these  conditions  we  must  be  content  with  estimating  the  un 
known  parameters,  namely,  the  0's  and  cr^. 

To  see  how  the  method  of  least  squares  operates,  consider  the  data 
of  Table  8.1.  Denoting  the  estimator  of  0/  by  §j(j=  1,  -  *  •  ,  g),  form,  the 
n  differences 


F2  -  *(X«,  •  -  •  ,  X,*  I  *!,  -  •  •  ,  0«)  -   F2  -   F2 
,  -  -  -  ,  Xpn  |  $x,  -  -  -  ,  0ff)  =  Yn  -   t 


162  CHAPTER    8,    REGRESSION    ANALYSIS 

TABLE  8.1— Symbolic  Representation  of  n  Observations  on  p  +  1  Variables 


Dependent 

Variable 

Independent  Variables 

F 

Xi                         A  2         *  •  •          Xp 

Fi 

%      s      a 

Yn 

"\^                                   "V                                   V 

-<A,  ITT.                                    -A.  2n.                                    -^yn 

The  values  of  the  §j(j '  =  1,  •  *  •  7  #)  are  then  determined  by  minimising 
the  sum  of  the  squares  of  the  deviations  specified  by  Equation  (8.5). 
That  its,  the  §j  arc  found  by  minimizing 


- 


(8.6) 


This  is  a  familiar  problem  in  calculus:  *S  is  differentiated  with  respect 
to  each  of  the  estimators,  and  each  partial  derivative  is  set  equal  to  0. 
Symbolically, 


dS 


(8.7) 


and  this  system  of  equations  must  be  solved  for  the  estimates,  that  is; 
for  the  $/. 

8-6      GRAPHICAL     INTERPRETATION     OF    THE     METHOD 
OF   LEAST  SQUARES 

In  order  to  portray  graphically  the  concepts  of  the  method  of  least 
squares,  it  will  be  convenient  to  restrict  our«elve«  to  the  case  of  two 
related  variables,  This  is  because  cases  involving  three  or  more  vari 
ables  present  great  difficulties  in  graphic  presentation.  A  two-variable 
scatter  diagram  might  appear  as  in  Figure  8.1 


FfG.  8,1—  Example  of  a  scatter  diagram,  often 
referred    to    as    a    scattergrarru 

Suppose  that  the  functional  relationship  assumed  to  exist  between 


8.6       GRAPHICAL    INTERPRETATION    OF    METHOD    OF    LEAST    SQUARES  163 

77  and  X  is  that  specified  by  the  mathematical  model 

v  =  0!  +  62X.  (8.8) 

The  associated  statistical  model  is 

Y  =  0X  +  <923T  +  e.  (8.9) 

The  method  of  least  squares  would  then  be  used  to  obtain  a  regression 
(or  estimating  or  predicting}  equation.1  Since  Equation  (8.8)  is  the 
equation  of  a  straight  line,  the  regression  equation  would  also  turn 
out  to  be  a  straight  line  such  as  shown  in  Figure  8.2.  The  resulting 

Y 


FIG.   8.2—  Example  of  a  scatter  diagram  with  a  straight  line   inserted 

showing    the    vertical    deviations    whose    sum    of  squares    is    to    be 

minimized    by    the    proper    choice    of    straight    line. 

regression  equation  would  be  denoted  by 

Y  =  0!  +  0*X  (8.10) 

where  8j  estimates  Oj(j=  1,  2)  and  Y  estimates  both  Y  and  77. 

If  a  more  complicated  functional  relationship  had  been  assumed, 
for  example, 

(8.11) 


the  problem  would  be  somewhat  more  complex  to  handle  mathemati 
cally  bxit  the  principle  would  be  unchanged.  The  parameters  would  still 
be  estimated  by  minimizing  the  sum  of  the  squares  of  the  vertical 
deviations  about  the  appropriate  curve.  The  reader's  attention  is 
directed  to  Figure  8.3  for  an  illustration  of  a  situation  associated  with 
the  mathematical  model  specified  in  Equation  (8.11). 

Y 


X 


FIG,   8,3— Example  of  a  scatter  diagram  with   a   second   degree  polynomial 

inserted    showing    the    vertical    deviations    whose    sum    of    squares    is 

to    be    minimized    by    the    proper    choice    of    parabola. 

1  The  terms  regression  equation,   estimating  equation,   and   prediction  equation 
are  used  interchangeably. 


164  CHAPTER    8,    REGRESSION    ANALYSIS 

8.7      SIMPLE    LINEAR    REGRESSION 

Let  us  now  consider  in  detail  the  case  in  which  the  postulated  func 
tional  relationship  is  of  the  form2 

77  -  /50  +  ftiX  (8.12) 

or 

F  =  fto  +  foX  +  e.  (8.13) 

The  problem  is,  of  course,  to  estimate  (3o  and  /?i  from  the  observed 
sample  data.  That  is,  estimates  of  /30  and  ft,  denoted  by  60  and  61, 
must  be  found.  Using  the  method  of  least  squares,  the  estimates  60 
and  61  are  determined  by  solving  the  normal  equations* 


The  solutions  are 


rs  i  ^ 

** 

and 

i0  =  7  —  j^  (8.16) 

where  a;  =  -X"  —  5T  and  y==  F  —  7\  These  estimates  are  then  used  to  give 
\is  the  regression  equation 

f  «  J0  +  i^.  (8.17) 

The  presentation  of  a  worked  example  will  be  deferred  until  Section  8.9 

8.8      PARTITIONING  THE  SUM  OF  SQUARES  OF  THE  DE 
PENDENT  VARIABLE 

Regression  computations  may  also  be  looked  upon  as  a  process  for 
partitioning  the  total  sum  of  squares,  ^,Y*,  into  three  parts,  each  of 

*  The  symbols  $o>  ft,  60,  and  61  arc  \isod  here  rather  than  0o»  #i»  £o,  and  0\.  This 
ia  in  conformanco  with  general  usage.  Some  authors  prefer  ctf  ft,  a,  and  6  rather 
than  #g,  A,  baj  and  61.  Howovor,  the  latter  aro  hotter  suited  to  extensions  to  three 
or  raoro  variables. 

*  The  phraeo  "normal  oquationfl"  IB  \xsed  to  describe  the  equations  roaxilting 
from  the  least  gquaros   differentiation.   It  has  no  connection  with  the   normal 
distribxition, 


8.8       PARTITIONING    THE    SU/W    OF    SQUARES    OF    DEPENDENT    VARIABLE  165 

which  is  meaningful  and  useful.  Prior  to  this  chapter,  the  total  sum  of 
squares  was  shown  to  be  the  sum  of  two  quantities,  the  corrected4 
sum  of  squares  and  the  correction  for  the  mean  : 

Total  S.S.  ==   Correction  for  the  mean  +  Corrected  S.S. 

=  S.S.  due  to  the  mean  +  S.S.  of  deviations  about  the  mean.    (8.18) 
That  is, 

2.  (3-19) 


Using  regression  methods,  the  corrected  sum  of  squares  may  also  be 
subdivided  into  two  parts,  the  sum  of  squares  due  to  (simple  linear) 
regression  and  the  sum  of  squares  of  the  deviations  about  regression: 

Corrected  S.S. 

=  S.S.  due  to  regression  +  S.S.  of  deviations  about  regression.      (8  .  20) 
That  is, 


Substituting  Equation  (8.21)  in  Equation  (8.19)  gives 

]T)  F2  =  (Z  YY/n  +  6x  2:  xy  +  Z  (F  -    *")2-         (8.22) 
Expressing  this  in  words, 

Total  S.S.  =   (S.S.  due  to  the  mean)  +  (S.S.  due  to  regression) 

+  (S.S.  of  deviations  about  regression).  (8.23) 

More  properly,  this  result  should  be  stated  as 

Total  S.S.  =  (S.  S.  due  to  60)  +  (S.S.  due  to  Z>i  1  60) 

+  (Residual  S.S.).  (8.24) 

A  more  extensive  discussion  of  this  type  of  manipulation  and  of  the 
associated  notation  is  given  in  Sections  8.15  and  8.16.  For  the  present 
it  is  recommended  that  the  reader  make  an  effort  to  learn  the  notation 
and  the  manipulative  skills  involved.  The  acquisition  of  such  knowl 
edge  will  prove  most  helpful  in  the  remainder  of  this  book. 

Graphically,  each  of  the  indicated  partitions  of  the  total  sum  of 
squares  can  be  associated  with  the  sums  of  squares  of  segments  of  the 
F-ordinates.  This  is  illustrated  in  Figure  8.4,  where  the  ordinate  YQ, 
associated  with  Xe,  is  partitioned  according  to  the  identity 


Y  -   Y  +  (f  -   Y)  +  (Y  -   ?)•  (8,25) 

In  words,  this  says  that 

4  The  expression  "corrected  sum  of  squares'7  is  used  to  represent  the  total  sum 
of  squares  minus  the  adjustment  (or  correction)  for  the  mean.  That  is,  it  is 
simply  a  synonym  for  the  sum  of  the  squares  of  the  deviations  about  the  mean. 


166 


CHAPTER    8,    REGRESSION    ANALYSIS 


F(G.  8.4— Diagram  to  illustrate  the  partitioning  of 
the    total    sum    of   squares. 

Observed  Y  =  (Contribution  due  to  the  mean) 

+  (An  additional  contribution  due  to  regression) 
+  (Deviation  from  regression). 


(8.26) 


If  we  carry  through  the  algebraic  manipulations  without  error,  Kqua- 
tion  (8.22)  may  be  derived  from  Equation  (8.25)-  The  proof  of  this  is 
left  as  an  exercise  for  the  reader. 

When  each  partition  of  23^2  is  associated  with  a  corresponding  por 
tion  of  the  total  degrees  of  freedom,  the  technique  is  known  as  analysis 
of  variance.  Such  results  are  usually  presented  in  tabular  form, 
referred  to  as  an  analysis  of  variance  table.  This  is  illustrated  in 
Table  8.2.  The  first  line  of  this  table  is  frequently  omitted  and  the 
total  line  expressed  as  ]C//2sssa  S^2 — (5D^)2/n'  with  n— 1  degrees  of 
freedom.  However,  in  this  book  the  results  will  always  be  presented  in 
the  form  unod  in  Table  8.2. 

TABLE  8.2— General  Analysis  of  Variance  for  Simple  Linear  Regression 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of  Squares 

Mean  Square 

Due  to  &o  

Due  to  &i|  60-  .  - 
Residual      .  ,  ,  , 

1 
1 
n  —  2 

(  r  YY/U 

b\  23  xy 

V  (Y  —  J>)2 

(  z;  nv» 

61  H  *y 
y;  fy  _  #•)«/(«  —  2) 

Total 

71 

T\  K2 

8.9       A    PRACTICAL    EXAMPLE 


167 


8.9     A  PRACTICAL   EXAMPLE 

To  illustrate  the  methods  discussed  in  the  preceding  sections,  con 
sider  the  data  in  Table  8.3.  Following  the  methods  outlined  in  the 

TABLE  8.3-Schopper-Riegler  Freeness  Test  of  Paper  Pulp  During  Beating 


Hours  of 
Beating 
X 

Schopper-Riegler 
(in  degrees} 
Y 

Hours  of 
Beating 
X 

Schopper-Riegler 
(in  degrees} 
Y 

1 

17 

8 

64 

2 

21 

9 

80 

3 

22 

10 

86 

4 

27 

11 

88 

5 

36 

12 

92 

6 

49 

13 

94 

7 

56 

Source:  O.  L.  Davies,  Statistical  Methods  in  Research  and  Production,  Oliver  and  Boyd, 
Edinburgh,  1949,  p.  161.  By  permission  of  the  author  and  publishers. 


preceding  sections,  we  obtain 

1360  +     916i  =     732 
916o  +  8196i  =  6485 

which  yields  1?  =  3*962+7A78X  as  the  regression  equation.  It  is  also 
observed  that 

F2  =  51,712 
Y»  =  41,217.23 

>2  =  10,494.77 

y  =  10,177.59 

^ 317.18 

and  these  results  are  presented  in  analysis  of  variance  form  in  Table 
8.4.  (NOTE:  A  convenient  form  to  use  in  performing  the  calculations 
is  given,  in  Table  8.5.) 

The  estimated  function  is  pictured  in  Figure  8.5.  Examination  of 
Figure  8.5  will  suggest  that  a  cubic  equation  (i.e.,  a  third-degree 
polynomial)  would  be  a  better  fit  to  the  observed  data.  However,  a 
discussion  of  the  desirability  or  appropriateness  of  fitting  a  different 
function  and  of  methods  of  fitting  other  than  a  simple  linear  function 
will  be  deferred  until  later. 


168  CHAPTER    8,    REGRESSION    ANALYSIS 

TABLE  8.4-Analysis  of  Variance  of  the  Schopper-Riegler  Data  of  Table  8.3 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of  Squares 

Mean  Square 

Due  to  60  

1 

41,217.23 

41,217.23 

Due  to  b\  |  60  

1 

10,177.59 

10,177.59 

Residual    

11 

317.18 

28.83 

Total 

13 

51,712.00 

TABLE  8.5-Suggested  Form  for  Calculation  and  Presentation  of  Results 

in  Simple  Linear  Regression 


n  = 

x  - 


Z  F  - 
7  «= 


^c;— «        2    


Z  XY 

Z  ry« 

Z  xy 


Z  -^2 


Source  of 
Variation 


Due  to  60- 
Due  to  bi\ 
Residual  . 


Degrees  of 
Freedom 


Sum  of 
Squares 


Mean 
Square 


Total 


8.10  ASSUMPTIONS  NECESSARY  FOR  ESTIMATION  AND 
TESTING!  HYPOTHESES  IN  SIMPLE  LINEAR  RE 
GRESSION 

Before  we  can  construct  confidence  Intervals  or  specify  teat  pro 
cedures,  certain  assumptions  are  generally  made.  Thus,  in  addition  to 
the  assumption  of  "no  error' *  in  the  independent  variable  made  in 
Section  8.5,  the  usual  assumptions  are: 


8. ID       ASSUMPTIONS    IN    SIMPLE    LINEAR    REGRESSION 


169 


(1)  For  a  given  X,  the  Y's  are  normally  and  independently  dis 
tributed    about    a    mean    nY\x  =  r)  =  /3Q+/3iX    with    variance 
aY\x~air-x-  The  assumption  concerning  the  mean  is  equiva 
lent  to   assuming  that  e  is  normally  and  independently  dis 
tributed  with  mean  0,  where  e  is  defined  by  Equation  (8.13). 

(2)  The  variance  o^lx  is  the  same  for  each  X  and  can  therefore 
be  denoted  by  o-^.  The  subscript  E  is  used  because  o^,  is  the 
variance  of  the  errors  denoted  by  e.  It  (cr|.)  is  commonly  re 
ferred  to  as  the  variance  of  the  "errors  of  estimate." 

The  preceding  assumptions  are  summarized  by 


V  =  /304- 


i=  1, 

3=  1, 


,  * 


(8.27) 


where 


ii  =  the  number  of  values  of  Y  associated  with  the  ith  value  of  X, 
it  =  n  =  the  total  number  of  values  of  F  (or  of  X),  and  the  et-y 


*— 1 


are  normally  and  independently  distributed  with  mean  zero  and  stand 
ard  deviation  0-%.  This  last  phrase  is  frequently  abbreviated  to  "the 
are  NID  (0, 


10O 


80 


^.  _ 
ujo:6O 

ceD 


Q-  — 
O 

§       20 


3.962+  7  478  X 


J  X 


O  2  4  6  8  1O  12 

HOURS   OF    BEATING 

FIG.  8.5— Plot  of  data  in  Table  8.3  with  the  least  squares  line  inserted 


170  CHAPTER    8,    REGRESSION    ANALYSIS 

8  11      ESTIMATES  OF  ERROR  ASSOCIATED  WITH  SIMPLE 
LINEAR    REGRESSION   ANALYSES 

Granting  the  assumptions  of  Sections  8.5  and  8.10,  and  further  as 
suming  that  the  failure  of  the  assumed  model  to  fit  the  observations 
exactly  is  solely  a  function  of  the  errors,  the  mean  square  for  devi 
ations  about  regression  (i.e.,  the  residual  mean  square)  can  be  used  as 
an  estimate  of  a-%.  Symbolically, 

residual  mean  square  =  ]£  (F  -   Y}*/(n  -  2)  =  4.         (8.28) 


We  must  always  remember,  though,  that  such  an  estimate  can  be 
badly  inflated  if  the  assumed  mathematical  model  is  inadequate.  More 
will  be  said  on  this  subject  a  little  later. 

Once  we  have  determined  the  variance  estimate  $%,  it  is  a  straight 
forward  matter  to  obtain  estimates  of  the  variances  of  various  statistics 
calculated  in  the  regression  analysis.  Those  of  general  interest  will  be 
given  without  derivation. 

Estimated  variance  of  the  regression  coefficient  &i 

2  2    /  x    ^       o  /'Q     O  Q"\ 

o         —- ^     c      /    s       3C*t  \O  .  J-t  i'y 

Estimated  variance  of  estimated  'mean  Y  for  given  X 

(8.30) 


Estimated  variance  of  predicted  individual  Y  for  given  X 


Estimated  variance  of  the  regression  coefficient 


These  may  then  be  used  to  test  hypotheses  about,  or  to  provide  interval 
estimates*  of  ,  various  unknown  parameters. 

8.12     CONFIDENCE     AND     PREDICTION      INTERVALS      IN 
SIMPLE   LINEAR    REGRESSION 

In  most  linear  regression  problems  the  estimator  of  greatest  impor 
tance  is  the  slope  61.  This  is,  of  course,  an  estimator  of  0X.  To  provide 
a  lOOy  per  cent  confidence  interval  of  /3i,  we  compute 


-  ii  =F 
where  $6l  is  defined  by  Equation  (8.29). 


8.12       CONFIDENCE    AND    PREDICTION    INTERVALS  171 


Using  the  data  in  Table  8.3  and  the  results  of  Section  8.9,    it  may 
be  verified  that 


Example  8.1 

he  data  in 
that 

s^  =  317.18/11  =  28.83 
and 

4x  =  28.83/182  =  0.1584. 

Thus,   a   95   per   cent   confidence   interval  for  /Si  is   determined   to   be 
(6.602,  8.354), 

If  a  100-y  per  cent  confidence  interval  estimate  of  0o  is  needed,  we 
have  only  to  compute 

L  _ 

=    ft       +    *-2»-2*5  (8.34) 


where  s&o  is  defined  by  Equation   (8.32).  No  example  will  be  given; 
however,  some  of  the  problems  will  require  use  of  Equation  (8.34). 

It  is  also  possible  that  we  might  wish  to  determine  a  confidence 
region  for  the  simultaneous  estimation  of  /Jq  and  jSi.  Making  use  of  the 
fact  that 


is  distributed  as   tf*      and:  that  (n  —  2)sJ/<r^  is  distributed  as  x^n_2)  it 
is  seen  that 


is  distributed  as  F  with  ?i  =  2  and  ^2  =  '^  — 2  degrees  of  freedom.  The 
boundary  of  the  lOOy  per  cent  confidence  region  is  then  determined 
by  solving 

[w(6o  —  /3o)2  +  2nX(bQ  —  /30)(&i  —  £1)  +  (&i  -  £i)2  23  x*\/2sl 

===     Fy(2,n — 2)          (8.37} 

for  /So  and  /Si. 

Another  estimation  problem  of  importance  in  simple  linear  regres 
sion  is  associated  with  5^  =  6o  +  &iX\  As  you  will  remember,  3?"  =  &o  +  &i-X" 
is  an  estimator  of  /xrjX==^  =  /30  +  jS1-X".  Further,  by  the  assumptions  of 
Section  8-10,  77  is  the  mean  of  a  normal  population.  Thus,  it  should  not 
be  surprising  that  a  lOOy  per  cent  confidence  interval  estimate  of  77  is 
provided  by 

\   -=  •&  :r  /  *~  (8  38) 

rrl  ^      [(1-H-Y)  /2]  (n—  2)°F  V0-^0/ 

C// 

where  s$>  is  defined  by  Equation  (8.30). 

It  should  be  noted  that  ^  =  &o+&i^  is  also  a  predictor  of  F"  — /30  + 


172 


CHAPTER    8,    REGRESSION    ANALYSIS 


That  is,  Y  can  also  be  used  to  predict  an  individual  Y-value 
associated  with,  a  given  X- value.  (NOTE:  This  is  in  contrast  to  the 
preceding  paragraph  where  ¥  was  used  to  estimate  the  mean  of  a  nor 
mal  population.)  When  !F  is  used  to  predict  an  individual  value  rather 
than  a  mean  value,  a  100-y  per  cent  prediction  interval  is  provided  by 


L>\- 

Uf] 


(8.39) 


where  s$  is  defined  by  Equation  (8.31). 

The  nature  of  the  confidence  and  prediction  intervals  specified  by 
Equations   (8.38)   and   (8.39)   is  illustrated   in   Figure   8.6.    The   most 


FIG.  8.6— Graphical  representation  of  the  confidence  and  prediction 
intervals  specified  by  Equations  (8.38)  and  (8,39). 


noticeable  feature  of  Figuro  8.0  is  the  curvature  of  the  confidence  and 
prediction  limits.  That  is,  our  estimates  arc  most  precise  at  the  average 
value  of  X  and  may  bo  almost  useless  at  values  of  X  far  removed  f rom 
j?.  By  "almost  useless/'  we  mean  that  the  confidence  and  prediction 
intervals  may  turn  out  to  be  so  wide  an  to  render  them  of  little  value. 
To  state  the  preceding  conclusion  in  a  positive  rather  than  a  negative 
fashion,  any  estimate  of  the  mean  value  of  Y  for  a  given  X  or  any 
prediction  about  an  individual  Y  associated  with  a  given  X  will  be 
mont  meaningful  for  those  values  of  X  near  3?. 

As  a  corollary  to  the  preceding  paragraph,  it  is  clear  that  if  the  esti 
mation  of  £o  is  of  prime  importance,  the  values  of  X  should  have  been 
selected  (ptior  to  collecting  data)  BO  that  3?=«0.  The  reason  for  this 
statement  should  be  clear.  It  is,  of  course,  that  by  so  choosing  the  X- 


8.12       CONFIDENCE    AND    PREDICTION    INTERVALS 


173 


values,  the  narrowest  confidence  and/or  prediction  interval  will  occur 
at  X  =  0,  and  it  is  at  this  value  of  X  that  Y  =  bo. 

Following  up  the  line  of  thought  started  in  the  preceding  paragraph, 
one  might  wonder  if  choosing  the  X- values  in  accordance  with  the  ex 
pressed  recommendation  is  best  for  all  purposes.  For  example,  if  the 
estimation  of  pi  rather  than  /50  is  of  prime  importance,  jshould  the 
values  of  the  controlled  variable  still  be  selected  so  that  X  =  0?  The 
answer  is,  "Definitely  not."  If,  then,  our  only  interest  lies  in  £1  (and  it 
frequently  does),  how  should  the  values  of  X  be  chosen?  In  this  case 
the  appropriate  recommendation  is:  select  two  values  of  X  (as  far 
apart  as  is  reasonable)  and  obtain  random  observations  on  the  Y- 
variable  at  only  those  two  X- values.  By  following  this  rule,  the  standard 
error  of  61  will  be  made  as  small  as  possible  subject  to  the  (uncon 
trollable)  magnitude  of  SE.  In  other  words,  if  we  proceed  as  indicated, 
the  confidence  interval  for  0i  should  be  kept  "small."  (NOTE:  The 
reader  can  verify,  heuristically,  the  wisdom  of  this  approach  by 
noting  that  widely  divergent  -XT- values  will  increase  ^C#2,  the  denomi 
nator  of  Equation  (8.29),  and  thus  decrease  the  size  of  s^.) 


FIG.  8.7—  Illustration  of  the  danger  of  extrapolation. 

Another  fact  which  should  not  be  lost  sight  of  is  that  predicting 
values  of  Y  for  a  given  X  value  is  even  more  hazardous  than  already 
indicated  if  we  attempt  such  a  procedure  for  an  X  value  outside  the 
range  of  the  chosen  values  of  X  used  in  obtaining  the  sample  regression 
line.  That  is,  extrapolation  beyond  the  observed  range  of  the  independ 
ent  variable  is  very  risky  unless  we  are  reasonably  certain  that  the 
same  regression  function  does  exist  over  a  wider  range  of  X- values 
than  we  have  in  our  sample.  A  simple  illustration  will  suffice  to  point 
out  the  possible  trouble.  Suppose  we  have  values  of  X  and  Y  which 
plot  (see  dots)  as  in  Figure  8.7.  In  the  given  range  of  -XT,  a  straight  line 
appears  to  be  a  good  fit  to  the  data  and  we  might  be  tempted  to  project 
our  regression  line  farther  in  both  directions.  However,  it  is  entirely 
possible  that  if  we  had  chosen  a  wider  range  of  -XT-values  and  observed 
the  associated  F-values  (see  circled  dots),  a  second  degree  polynomial 
might  have  been  indicated  as  the  true  form  of  the  regression  function 
rather  than  the  straight  line  we  have  drawn.  You  can  readily  see  that 
predicting  values  of  Y  using  an  extrapolation  of  the  straight  line  could 
lead  to  serious  errors.  Therefore,  the  research  worker  is  advised  to 


174  CHAPTER    8,    REGRESSION    ANALYSIS 

act  with  caution  whenever  he  makes  predictions  which  involve  going 
outside  the  observed  range  of  the  independent  variable. 

One  further  remark  and  we  shall  proceed  to  the  subject  of  testing 
hypotheses.  Although  only  two-sided  confidence  and  prediction  limits 
have  been  discussed  in  this  section,  the  reader  will  realize  that  one- 
sided  limits  should  be  used  if  the  problem  calls  for  such  a  procedure. 
If  only  an  upper  or  lower  limit  is  required,  the  researcher  should  make 
the  same  changes  in  procedure  as  outlined  in  Chapter  6  but  continue, 
of  course,  to  use  the  statistics  specified  in  Section  8.11. 

8.13     TESTS    OF    HYPOTHESES    IN    SIMPLE    LINEAR    RE 
GRESSION 

Sometimes  the  researcher  is  interested  in  determining  whether  the 
estimated  slope  61  is  significantly  different  from  some  hypothesized 
value  of  /Si^say  /5(,  That  is,  he  wishes  to  test  the  hypothesis  H:fBi  =  f3{ 
against  the  alternative  A  :/3i^/3£.  The  appropriate  test  statistic  is 


where  sb  is  defined  by  Kquation  (8.29).  The  hypothesis  //  would  then 
be  rejected  if 

t    =>1    £<l-a/2)(n-2>  (8.41) 

or  if 

t  <  —  £(i—  «/2)<tt—  2).  (8.42) 

A  common  value  of  /5{  is  0  since  this  reflects  the  hypothesis  that  Y  is 
independent  of  X  (in  a  linear  sense)  ;  that  is,  that  X  is  of  no  value  in 
predicting  Y  if  a  linear  approximation  is  used. 

Example  8.2 

Referring  to  Example  8.1  and  letting  a  =  0.01,  test  the  hypothesis 
//:#!«().  Calculation  yields  £=  (7.478  —  0)/0.»08  «  18.788  >*0.o9fc(n> 
=  3.106.  Thus,  the  hypothesis  is  rejected. 

It  is  worth  noting  that  the  tent  of  the  hypothesis  //:$i  —  0  can  also  be 
performed  directly  from  the  analysis  of  variance  table.  To  illustrate 
the  nature  of  this  alternative  procedure,  consider  Table  8.G.  Because 
of  the  assumptions  made  in  Section  8.10,  It  is  possible  to  demonstrate 
that  the  expected  values  of  two  of  the  mean  squares  are  as  shown  in 
Table  8.6.  Thus,  if  //:/?a  =  0  is  true,  both  the  "mean  square  due  to 
&IJ.&D"  and  the  "residual  moan  square77  are  estimates  of  the  same  quan 
tity,  namely,  <r^/It  seems  logical,  therefore,  to  examine  the  ratio 

i  mean  square  due  to  bi    &o 

77  _  „_  ._  _rw  ™  ____  '  (& 

f.        i  A  f  "  ""'""       "  9  ^ 

,/  residual  mean  square 

and  if  this  ratio  is  significantly  larger  than  1,  some  doubt  would  be 
cast  upon  the  validity  of  the  hypothesis  //:#L*=0.  Since  it  may  be 


8.13       TESTS    OF    HYPOTHESES 


175 


TABLE  8.6— General  Analysis  of  Variance  for  Simple  Linear  Regression 
Showing  the  Expected  Mean  Squares 


Source  of 
Variation 

Degrees 
of  Free 
dom 

Sum  of 
Squares 

Mean  Square 

Expected  Mean 
Square 

Due  to  60  

1 

(  2Z  Y)2/n 

(  51  V)2/n 

Due  to  bi\b0.  . 
Residual  

1 

n  —  2 

t>i  2Z  xy 

52  (F  -  F)2 

bi  2Z  xy 

™*»iv_        ^       sr 

^  =  V   (F  —   Y)2/(n  —  2) 

«jr  +  J&?  S  *» 
ffl 

Total 

n 

5Z  Y* 

demonstrated  that  the  F-ratio  specified  by  Equation  (8.43)  follows  an 
^-distribution  with  v\  =  1  and  v2^=n  —  2  degrees  of  f  reedora,  the  hy 
pothesis  H:/3i  =  Q  will  be  rejected  if  F^ 


n_2). 
Example  8.3  ^u        /- 

Referring  to  Table  8.4,  it  is  seen  that  F  =  10,177.59/28.83  =  353.02. 
Since  this  exceeds  Fo.  99(1,11)  =  9.65,  the  hypothesis  .fiT:/3i  =  0  is  re 
jected.  This  is  the  same  conclusion  reached  in  Example  8.2.  (NOTE:  the 
fact  that  t%—F(i,v)  is  the  connecting  link  between  the  equivalent  test 
procedures.) 

Other  test  procedures  in  simple  linear  regression  are  concerned  with 
such  hypotheses  as: 


(1)  H:/3Q  =  /3o; 

(2)  H:vYlx~x0 

(3)  H:/3Q  =  j3o 


and     fa  = 


Rather  than  discuss  these  in  detail,  we  shall  only  indicate  the  appro 
priate  test  procedures.  For  the  three  cases  just  mentioned,  the  respec 
tive  test  statistics  are: 


where  sbo  is  defined  by  Equation  (8.32), 


(8.44) 

(8.45) 
Q,  and 

(8.46) 

The  test  procedures  are  summarized  in  Table  8.7. 

Again  we  shall  do  no  more  than  remind  the  reader  of  the  possibility 
of  one-sided  test  procedures.  By  this  time  the  method  of  procedure  in 


where 
p  = 


is  defined  by  Equation  (8.30)  and  evaluated  at  X  = 
0  -  pfo*  +  2nX(b0  -  / 


176  CHAPTER    8,    REGRESSION    ANALYSIS 

TABLE   8.7— Summary  of  Test  Procedures  in  Simple  Linear  Regression 


Hypothesis 

Statistic 

Equation 

Rejection  Region 

00    =   00 

t 

8.44 

/,             ->           f,f\  Q.  /*>\    fn              ft\ 

or 

1    ^    ""    ^(L^/2)Cn-2) 

01    =    0{  

t 

8.40 

t    >    t  i 

or 

^    5=    —    ^U—  a/2)(n—  2) 

=  u 

t 

8.45 

t    >    * 

J  |  «X"«"«<X'o             0 

—         kL       "/  •*/  ^T*       */ 

or 

^    5:    —    ^(1—  «/2)  (n—  2) 

0o  =  0o  and  0!  =  0(  

F 

8.46 

jC              ^"          £*     (lu     .    ft)   (^      71             *») 

such  cases  should   be  obvious  once  the   two-tailed  tests  have  been 
specified, 

8.14      INVERSE  PREDICTION  IN  SIMPLE  LINEAR  REGRES 
SION 


The  equation  f^  =  &o+&i-XT  may  sometimes  be  used  to  estimate  the 
unknown  value  of  X  associated  with  an.  observed  Y  value.  For  example, 
suppose  that  in  addition  to  the  data  of  Table  8,3,  we  have  a  Schopper- 
Riegler  reading  of  K  =  60  but  the  hours  of  beating,  X,  are  unknown. 
How  shall  this  unknown  value  be  estimated?  The  procedure  is  as  fol 
lows.  Compute 


«=  (Fo  -  60)  /6  1 


(8.47) 


where  F0  is  the  observed  value  of  Y  for  which  we  desire  to  estimate  the 
associated  X  value.  A  lOOy  per  cent  confidence  interval  for  the  true 
but  unknown  X  value  is  defined  by 


L] 

U) 

where 


and 


+ 


_ 
.  -  T).  + 


"n?«±lY 

^V     n     ) 


(8 . 48) 


(8.49) 
(8.50) 


_«>•  (8.51) 

If,  as  in  frequently  tho  case,  one  has  several  (say  m)  values  of  Y 


8.15       THE    ABBREVIATED    DOOLITTLE    METHOD  177 

associated  with  the  unknown  X,  Equations  (8.47)  and  (8.48)  are  modi 
fied  to  read 

X  =  (To  -  5o)/6i  (8.52) 

and 


U)  D  D  'V  \     nrn 

where 


{8.M) 


(8.54) 


(»-  2)4+  ZICFo,-   F0) 


(*',)  2  =  -  -  ,  (8.55) 

n  +-  m  —  3 

t    =    ^(H-T)/2](n+7n-3),  (8.56) 

and  B  and  D  are  the  same  as  before.  Since,  in  practice,  m  is  usually 
quite  small  relative  to  n,  the  computational  labor  may  be  reduced 
materially  by  using  s%  rather  than  (s^)2.  This  leads,  of  course,  to  an 
approximate  solution  but  one  which  is  sufficiently  accurate  for  most 
situations. 

Example  8.4 

Consider  the  problem  posed  at  the  beginning  of  this  section.  Using 
Equation  (8.47)  and  the  results  of  Section  8.9,  we  obtain  j£  =  7.494, 
Using  Equations  (8.48)  through  (8.51),  a  95  per  cent  confidence  interval 
estimate  is  determined  to  be  (5.85,  9.15). 

8.15     THE  ABBREVIATED   DOOLITTLE   METHOD 

Before  proceeding  to  the  consideration  of  regression  problems  of 
greater  complexity  than  simple  linear  regression,  let  us  digress  long 
enough  to  study  the  mechanics  of  a  method  for  solving  a  set  of  simul 
taneous  linear  equations  whose  coefficients  form  a  symmetric  matrix. 
This  digression  will  be  well  worth  while  for  several  reasons,  namely  : 

(1)  In  most  regression  problems,   the  postulated  mathematical 
model  is  linear  in  the  unknown  coefficients. 

(2)  The  method  is  well  suited  to   programming  for  high-speed 
computers  as  well  as  being  useful  when  only  desk  calculators 
are  available. 

(3)  It  incorporates  self-checking  features  which  permit  verifica 
tion  of  the  accuracy  of  the  arithmetic  calculations  at  each 
stage. 

Several  methods  of  solving  sets  of  simultaneous  linear  equations  (or 


178 


CHAPTER    8,    REGRESSION    ANALYSIS 


of  inverting  matrices)  appear  in  the  literature.  Each  of  these  is  a  vari 
ation  of  a  basic  procedure,  the  variations  being  introduced  to  accom 
plish  a  particular  aim  of  the  person  proposing  the  special  technique. 
In  this  book,  only  one  method  will  be  discussed.  It  is  known  as  the 
Abbreviated  Doolittle  Method. 

To  illustrate  the  Abbreviated  Doolittle  Method,  let  us  consider  the 
problem  of  estimating 

77  =  0oXQ  +  faXt  +  fcXz  +  /3*Xz  +  /?4^4  (8.57) 

where  XQ  is  a  dummy  variable  which  always  takes  the  value  1  (i.e,, 
JXTo^l).  The  method  of  least  squares  would  lead  to 

F  =  6 

and  the  coefficients 
equations 

(Z  A"o-Y0)fto  +  (Z 


(8.58) 
=  0,  1,  2,  3,  4)  would  be  found  by  solving  the 


(23  -^^0)60  +  (23 

(Z  -X-aA'0)6o  +  C23 
(23  -Y8AT0)6o  +  (23 
(23  Ar4Jr0)6o  +  (23 


i  +  (23 


i  +  (23 


0^4)  *4  ==  23  ^ 

+  (23  ^i.v8)&n 
1-^064  =  23  -v 

+  (23  ^aA^)J3 

r^064  =-  23 

+  (23  ^a-Yg)^ 

varoj4  =  23  Ar 


(8,59) 


t  4-  (Z  -^4X2)62  +  (Z  Xt 


=     Z 


If  the  data,  consisting  of  n  observations  on  each  of  the  variables,  are 
written  in  matrix  form,  namely, 


Y  = 


LFJ 


and   X 


-Y 


l2 


then  Equations  (8,59)  may  be  written  as 

=  X'Y 


(8.60) 

where  S'  =  [60,  &i?  62,  b%,  b*].  To  simplify  the  writing,  we  shall  denote 
X* X  by  A  and  XfY  by  G.  In  this  notation  Equation  (8.60)  appears  as 


AB  =: 


(8.61) 


8.15       THE    ABBREVIATED    DOOLITTLE    METHOD  179 

If  we  denote  A~~l  by  C,  then 

C  =  A~i  =  (X'X)-^  (8.62) 

and 

B  =  A-1*?  -  (JTJr)-1*?  =  CG.  (8.63) 

The  Abbreviated  Doolittle  Method  will  now  be  used  to  obtain: 

(1)  the  b's, 

(2)  the  sums  of  squares  associated  with  the  sequential  fitting  of 
the  &?s, 

(3)  the  estimated  variances  of  the  &'s, 

(4)  the  estimated  co variances  between  pairs  of  &'s,  and 

(5)  the  elements  of  the  inverse  of  the    X' X  matrix,  that  is,  the 
elements  of  C. 

The  mechanics  of  the  forward  solution  using  the  Abbreviated  Doo 
little  Method  are  summarized  in  Table  8.8.  A  discussion  of  the  steps 
involved  is  best  given  in  two  parts,  one  associated  with  the  first  section 
of  the  table  and  one  associated  with  all  the  succeeding  sections: 

First  Section  [Rows  (0)  Through  (4)1 

(1)  In  the  front  half  of  the  table  are  entered  the  elements  of  the 
matrix  of  coefficients  defined  by  Xr  X}  omitting  those  obvious 
from  symmetry.  That  is,  we  have  entered  a#  — ^JEVXTy  recog 
nizing  that  an  =  a,y . 

(2)  In  the  column  headed  "constant  terms"  are  entered  the  ele 
ments  of  the  vector  Xf  Y.  That  is,  we  have  entered  gt=^L,XiY. 

(3)  In  the  back  half  of  the  table  are  entered  the  elements  of 
the    identity    matrix,    again    omitting    those    obvious    from 
symmetry. 

(4)  In  the  check  column  are  entered  the  sums  of  all  entries  in  the 
corresponding  rows,   including  those  elements  omitted  because 
of  symmetry. 

Succeeding  Sections  [Rows  (5)  Through  (14)] 

(1)  Each  entry  in  a  given  row  is  generated  according  to  the  in 
struction  specified  for  that  row.  (See  the  first  column  of  the 
table.)  This  applies  to  the  front  half,  the  constant  terms,  the 
back  half,  and  the  check  column, 

(2)  The  sum  of  all  the  entries  in  a  row  (with  the  exception  of  the 
entry  in  the  check  column)   should  equal   (within  rounding 
error)  the  entry  in  the  check  column.  The  advantage  of  the 
checking  procedure  should  be  obvious:  If  an  arithmetic  error 
has  been  made,  it  will  be  corrected  before  calculations  are 
started  on  the  next  row. 

(3)  Steps  such  as  those  described  are  continued  until  a  row  is 
reached  in  which  only  a  single  Bpa  appears.  With  the  calcula- 


0 

^s 

,jcj 

4—> 

OJ 

CJ 

^-^- 

** 

t—          CO 

tf.5 

^= 

CO           -*t» 

/latrix  Associated  y 

M-H 

w 

PQ 

CD    CD    CD    CD    --H 
CD    CD    CD    '^H 
CD    CD    ^~~* 

CD    CD 
(""a    CD 

CD    CD 
CD    CD 
CD    CD 

CD    CD 
CD    CD 

^  pq* 

CD    CD 

eo 
-^•co 

^pq9 

:I1 

1^—  < 

10 

X 

to 

_o 

~d 

CD      -r-^ 

T-H 

CD    CD 

--T—  4 

^  pq4 

^J  pq 

"^*"pq 

^  PQ 

o     o 

^  PQ 

IR 

|    00 

E  ^ 

c^-B 
<D    ce 

r~*         r»  < 

4. 
£ 

C 
-1- 
C 

£ 

C 

_J 

1  i 

n     i—t 

***** 

4«s 

^^ 

^=j 

^p<f 

^  pq" 

ittie  Method  Illustrated  on  tl 
tr  Equations  Specified  by  Eqi 

^ 

o 

o 

S     S3     S3     S3 
S     53     8 

HI     g 

3     S 
S     3 

"nj<     pY"\ 

53      £9 
§ 

^<    -^i 
^  pq* 

S5     SS 

^  pq 

^5     ?5 
"^   PQ 

5 

*o    S 

o  « 

Q*  ^ 



t-3 

"S  "8 



l« 

I      I      !      ;      I 

rxj 

>     <^ 

3J 

£2 

*^"* 

^>< 

H 

. 

1 

<< 

'.      '.      ',      !      i 

CD 

s 

35 

ES 

35 

S3 

"*T 

•XT' 

*  '  i 

'**^ 

^^N 

2 

^ 



1 

I 

P^ 

oc? 

o? 

OC? 

*T* 

^S** 

•0 

•*>* 

O 

. 

^ 

r5 

I 

1 

I 

E—  < 

*r~"*^ 

**""*"*» 

X  *s 

^•""•N 

ot, 

^ 

!£* 

^ 

a 

O 

s    31 

CO 



0 

^3  ^ 

"^   J 

"X  X 

x^  ^q 

S 



•^^ 

1    ^  — 

t     ^--v 

t       £^s 

1     ^ 

HH 
c"Y"t 

CD    H-O 

•*—  4     t*— 

CS    CT^ 

f*O    v* 

-^       T—  t 

^ 

,      .      .      .      . 

tl       li 

11          It 

11        11 

11       11 

11              11 

H 

&^'^iz?<3r 

SS^ 

di- 

8.15       THE    ABBREVIATED    DOOL1TTLE    METHOD  181 

tion  of  all   entries  in  this   row  and  the   satisfaction   of  the 
"check,"  the  forward  solution  is  complete. 

The  next  step  in  the  analysis  is  the  completion  of  the  backward 
soUition.  This  will  be  performed  in  two  parts,  one  to  determine  the  b's 
and  the  other  to  determine  the  c^  values. 

Determination  of  the  b's 

The  forward  solution  of  the  Abbreviated  Doolittle  Method  has  pro 
vided  us  with  the  following  set  of  equations  : 

(1)  60  +  (-BoOftx  +  0802)62  +  (£03)63  +  (£04)64  =  BQy 
(1)  &!  +  (£12)62  +  (^13)63  +  (£14)64  =  Siy 

(1)  62  +  (£23)63  +  (£24)64  =  £2,,          (8.64) 
(1)  63  +  (£34)64  =  £3*, 
(1)  64  = 


Solving  these  in  reverse  order  (hence  the  name  "backward  solution"), 
we  obtain: 

64    == 

63  = 

62  =  Biy  —  64£24  —  63£23  (8.65) 

b    =  Bl    —  J4J5i4  —  63£i3  —  62£is 


Determination  of  the  c,/ 

(1)  Since  C=  A^  =  (Xf  X)~~l  is  the  inverse  of  a  symmetric  matrix, 
it  will  be  symmetric.  This  reduces  the  number  of  calculations 
to  be  performed  since  Cji  =  ci:f. 

(2)  All  Cij  values  may  be  calculated  using  the  equation 

4L-£L  (8.66) 


Ar— 0 


in  which  some  of  the  A '  values  may  be  0  or  1  and  some  of  the 
Bf  values  may  be  0.  It  should  be  noted  that  some  of  the  c^ 
values  may  be  read  directly  from  the  forward  solution,  namely, 


£40  ^  £40 


41 


£41  =  £ 

C42  ^  £42 
£43  =  £43 
£44  =  £44 


182 


CHAPTER    8,    REGRESSION    ANALYSIS 


(3)  If  we  choose  to  ignore  the  symmetry  mentioned  in  (1)   and 
calculate  all  the  c^  independently,  a  check  can  be  made  on  the 
arithmetic  by  comparing  c^  and  c/»-. 

(4)  A  final  check  could  be  made  by  seeing  if    CA  equals   /.   It 
should.  However,  rounding  errors  may  cause  minor  discrepan 
cies. 

Having  completed  both  the  forward  and  backward  solutions  using 
the  Abbreviated  Doolittle  Method,  the  next  step  is  to  indicate  how  to 
obtain  the  analysis  of  variance  table  and  the  standard  errors  associ 
ated  with  the  various  statistics.  Using  the  formula 

&!,---,      &1-1      =       AiyBiy,  (8.67) 


S.S.  due  to  bi\ 


the  various  sums  of  squares  may  be  evaluated  easily  once  the  forward 
solution  of  the  Abbreviated  Doolittle  Method  has  been  completed. 
The  analysis  of  variance  may  then  be  recorded  as  in  Table  8.9.  If  we 
do  not  wish  to  record  the  reduction  in  the  residual  sum  of  squares  as 
sociated  with  the  sequential  fitting  of  each  additional  b,  it  is  proper  to 
note  that  the 


S.S,  due  to  regression  = 


(8.68) 


i—l 


and  this  pooled  sum  of  squares  possesses  4  degrees  of  freedom,  (NOTE: 
The  sum  of  squares  due  to  60  is  still  recorded  separately  since  it 
is  actually  CEltY^^/n,  that  is,  it  is  the  correction  for  the  moan.)  The 
estimated  variances  and  covariances  associated  with  the  regression 


TABLE  8.9-Analysis  of  Variance  Associated  With  the  Multiple  Regression 

Problem  Discussed  in  Table  8.8 


Source 

of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean  Square 

Due  to  &o 

t 

./I  {>»/  lJ  (Vy 

A.  Qi/^Ot/ 

Due  to  hi 

60  

1 

,/l  \y  t^ly 

A  ly^ii/ 

Due  to  b% 

60,  b\  ..».,,.. 

1 

"**  SV"^2l,/ 

A  ty&ty 

Due  to  b% 

1  &0>    &ly    62  

1 

***  3i/jCf  3^» 

A^  3t/*^3j/ 

Due  to  64 
Resklual  . 

&0,   &1,    6'2,    &3-  -  - 

1 
w  —  5 

AtyBtu 

Y\  (F—  £V 

A  4j/^4|/ 

^  ^   y-  (K—  l^Wte  —  5) 

Total 

n 

V  F2 

8.15       THE    ABBREVIATED    DOOLITTLE    METHOD  183 

coefficients  are  given  by 

sl*  =  c^s  (8.69) 

and 


From  these,  we  may  obtain 

4  =  (JTCJr)4  (8.71) 

which  must  be  evaluated  at  the  particular  set  of  .XT- values  for  which  an 
estimate  is  desired, 

Example  8.5 

Consider  the  data  given  in  Table  8.10.  The  Abbreviated  Doolittle 
Method  is  applied  to  these  data  in  Table  8.11,  where  the  X's  and  Y 
have  been  coded  so  that  the  elements  of  XfX  and  X'Y  are  approxi 
mately  of  the  same  order  of  magnitude.  (This  is  done  to  facilitate  the 
computations.)  In  this  case,  the  coding  is  as  follows:  divide  Xo  by  10, 
X^  by  100,  X2  by  10,  Xz  by  1000,  X4  by  1000,  and  Y  by  100.  Then 
Equation  (8.65)  is  used  to  obtain  JVlOO  =  —  0.681468(X0/10) 
+  0.227227(Jf  i/100)  +  0.055349(X2/10)  —  1 ,495563(X3/1000)  + 1 .546520 
(AV1000)  and  Equation  (8.66)  yields 

2052.64  —135.410     —53.6729     —538.760       2.25605  ~ 
135.410       20.0032          0.273033        22.3301        0.549940 

53.6729       0.273033        2.73843          18.3161  -0.928040 

538.760       22.3301          18.3161          171,128  -11.6657 
2.25605       0.549940   -0.928040   -11.6657        8.32196  . 

where  the  elements  of  C  reflect  the  coding  explained  above.  The  analysis 
of  variance  of  the  coded  data  is  presented  in  Table  8.12. 

Example  8.6 

The  example  of  Section  8.9  is  reworked  in  Table  8.13  using  the  Ab 
breviated  Doolittle  Method.  Again  we  get  !F  =  3.962  +  7.47S-XV  Also, 
coo  ^  0.3460,  coi  =  —0.0384,  and  cn  =  0.0055.  It  can  be  verified  that  the 
sums  of  squares  agree  (within  rounding  error)  with  the  values  reported 
in  Table  8.4. 

Although  the  Abbreviated  Doolittle  Method  was  explained  with 
reference  to  Equation  (8.57),  it  should  be  noted  that  it  applies  equally 
well  to  any  situation  where  the  equation  is  linear  in  the  unknown  coeffi 
cients.  For  example,  the  following  are  typical  of  cases  frequently  en 
countered  for  which  the  technique  will  prove  useful : 

(1)  n  -  /So  +  PiXi  +  •  •  •  +  faXk, 

(2)  77  =  /30  +  /3i^i  +  PnXl  +  /3mXr,   and 

(3)  77  =  /?o  +  /SxXi  +  faX2  +  faiXl  +  (3^x1  +  ffuXiX*. 

Some  of  these  will  be  considered  in  the  sections  and  chapters  to  follow. 


TABLE  8.10~Crude  Oil  Properties  and  Actual  Gasoline  Yields 


Crude  Oil 
Gravity, 
°API 

Crude  Oil 
Vapor 
Pressure, 
PSIA 

Crude  Oil 
ASTM 
10%  Point, 
°F. 

Gasoline 
End 
Point, 
°F. 

Gasoline 
Yield 
Per  cent  of 
Crude  Oil 

Xi 

X* 

X3 

A"* 

Y 

38.4 

6.1 

220 

235 

6.9 

40,3 

4.8 

231 

307 

14.4 

40.0 

6.1 

217 

212 

7.4 

31,8 

0.2 

316 

365 

8.5 

40.8 

3.5 

210 

218 

8.0 

41.3 

1.8 

267 

235 

2.8 

38.1 

1.2 

274 

285 

5.0 

50.8 

8.6 

190 

205 

12.2 

32.2 

5.2 

236 

267 

10.0 

38.4 

6.1 

220 

300 

15.2 

40.3 

4.8 

231 

367 

26.8 

32.2 

2,4 

284 

351 

14.0 

31.8 

0.2 

316 

379 

14.7 

41.3 

1.8 

267 

275 

6.4 

38.1 

1.2 

274 

365 

17.6 

50.8 

8.6 

190 

275 

22.3 

32.2 

5,2 

236 

360 

24.8 

38.4 

6.1 

220 

365 

26.0 

40.3 

4,8 

231 

395 

34  .  9 

40.0 

6.1 

217 

272 

18.2 

32.2 

2,4 

284 

424 

23.2 

31.8 

0.2 

316 

428 

18,0 

40.8 

3.5 

210 

273 

13.1 

41.3 

1.8 

267 

358 

16.1 

38.1 

1.2 

274 

444 

32.1 

50.8 

8.6 

190 

345 

34.7 

32.2 

5.2 

236 

402 

31.7 

38.4 

6.1 

220 

410 

33  .  6 

40.0 

6.1 

217 

340 

30.4 

40.8 

3.5 

210 

347 

26.6 

41.3 

1,8 

267 

416 

27.  B 

50.8 

8.6 

190 

407 

45.7 

Source:  Nilon  H.  Prater,  "Estimate  Gasoline  Yields  from  Crudes/'  Petroleum  Refiner 
(now  Hydrocarbon  Processing  and  Petroleum  Refiner) ,  Vol.  35,  No.  5,  pp.  236—38,  May, 
1956.  Copyright  1956,  Gulf  Publishing  Co.,  Houston,  Texas.  By  permission  of  the  author 
and  publishers. 


[1841 


o 

1 — I 
00 


oS 

O 


1 


^ 

NO  lO  fO  Os  CM 

r--  \o  ON  o  o 

.378600 
.933125 

OO  CM 
>0  CM 

.567474 
.020022 

11 

t-~  »O 

is 

^HO 

( 

J 

CM  CM  —  i  *-l 

CM 
1      1 

1  7 

S 

ON 

CM 

OOO  0  —1 

oo 

00 

00 

00 

^« 

CM 

lO  CM 

ONOO 
r-»NO 

OOO  TH 

00 

oo 

00 

^2 

** 

rH 

1     I 

±4 

NO 

10  OO 

co  NO 

ON  1-4 

SI 

U 

NO 

§s 

So! 

25 

OO  "i 

oo 

oo 

^0 

°^ 

oo 

1  1 

o 

OM 

1/N  IO 

ooo 

00  CM 

CM  ON 

ro  O 
00  •«* 

NO  ON 

o  10 

O    -H 

00 

1HO 

CM  CM 

1     I 

CM 

00 

§ 

83 

^JS 

OCM 

ON  tO 

CM 

to  r* 
CM  NO 

O  00 

CM  O 

NOO 

CM  CM 

^ 

^co 

***  CO 

JC»-  »O 

"°£ 

OCM 

I      I 

1  1 

onstant 

£j 

o  10  oo  NO  oo 

i-t  ro  CO  ON  VH 
vo  iO  ON-*  CM 

629100 
965938 

NO  OO 

fj  ON 
CO  OO 

CM  TH 

ro  ro 

NOO 
CO  CM 
00  »O 

CJ 

O  CM  CM  i-«  CM 

O-rH 

oo 

OO 

OO 

O  ~H 

ISlii 

88 

iH  CM 

2jg 

to  ON 

vo 

^ 

illll 

CM  O 

ON  00 

O  <*> 

»O  CM 

OO 

§o 

o 

CM 

-<rfxHC4rO 

*~<rO 

00 

1  1 

00 

!    1 

O-i 

o- 

§tHO  CM 
CM  r-  Q 

CM  »O 

ONO 

*H  CM 

to  NO 

00  ON 

1 

^' 

t*-  ON  ON  ON 

f^5 

0^ 

^^ 

8 

3 

OCMCMTH 

OCM 

00 

I  1 

00 

1  I 

O-H 

1 

4, 

II 

go 

O  00 
CM  00 

NO 

CM 

VH!OI> 

~"* 

O  CM 

*-(1H 

- 

O  CM 

vOOO 

CMO 

NOIO 

IOCM 
CM  ON 

OO 

S 

= 

S 

8 

CM 

d 

O-H 

1 

S'CscM'^?^' 

Co  to 

PS? 

c?0^ 

iH  CM 

tt 

[1851 


1  86 


CHAPTER    8,    REGRESSION    ANALYSIS 


TABLE  8.12~Analysis  of  Variance  Associated  With 
the  Regression  Analysis  of  the  Data  in  Table  8.10 


Source 

of  Variation 

Degrees  of 
Freedom 

Sum  of  Squares 

Mean  Square 

Due  to  60 

1 

1.236772 

1.236772 

Due  to  61 

60        

1 

0.021625 

0.021625 

Due  to  Z>u 

60,  &i 

1 

0.030985 

0.030985 

Due  to  63 

&GJ  b\j  bz    

1 

0.002921 

0.002921 

Due  to  64 

&o,  &i7  &•>,  63.  .  •  . 

1 

0.287399 

0.287399 

Residual  . 

27 

0.013477 

0  .  000499 

Total 

32 

1  .593179 

TABLE  8.13-Solution  of  the  Example  of  Section  8,9 
by  the  Abbreviated  Doolittle  Method 


Front 

Half 

Row 

60 

61 

Terms 

Back 

Half 

Check 

(0)  

13 

91 

732 

1 

0 

837 

(1)    

819 

6485 

1 

7396 

(2)  

13 

91 

732 

1 

0 

837 

(3)  

1 

7 

56.3077 

0.0769 

0 

64.3846 

(4)  

182 

1360.9993 

—  6.99790 

1 

1537,0014 

(5)  

1 

7.4780 

—  0.03845 

0  .  0054945 

8.44506 

8.16      SOME  ADDITIONAL    REMARKS  WITH    REGARD   TO 
GENERALIZED    REGRESSION   ANALYSES 

There  are  a  few  additional  items  which  merit  discxission  at  this 
time,  for  they  have  a  great  deal  to  do  with  the  analysis  of  any  sot  of 
data  when  the  regression  technique  is  employed. 

The  first  item  to  be  discussed  is  notation.  By  now  the  reader  may 
have  been  woiidering  what  the  significance  is  of  such  symbolism  as 
61(60  ttnd  62(60,  61.  This  notation  is  closely  allied  with  the  "conditional" 
concept  in  probability.  In  fact,  the  notation  is  \ised  in  exactly  the 
same  manner,  that  is,  to  indicate  a  condition  or  restriction.  In  the 
present  context,  the  notation  calls  attention  to  the  fact  that  sxims  of 
squares  associated  with  various  coefficients  are  obtained  in  a  definite 
(seqxieutial)  order.  Jn  particular,  the  sum  of  squares  due  to  60  is  found 
first  and  is  the  same  as  finding  the  correction  for  the  mean.  (NOTE:  60 
itself  is  not  equal  to  the  mean;  it  also  depends  on  the  nature  of  the 
mathematical  model  used  to  represent  the  data.)  After  finding  the  sum 


8.16       GENERALIZED    REGRESSION    ANALYSES  187 

of  squares  due  to  &0,  we  find  the  sum  of  squares  due  to  61 ;  hence  the 
symbolism  &i|  &0  which  is  read,  "&i  given  that  60  has  been  determined. " 
Referring  to  Table  8.9,  the  next  sum  of  squares  recorded  was  that 
"due  to  62 1  &o,  &i"  which  implies  that  62  was  found  after  &0  and  61  had 
been  determined.  The  remaining  symbols  in  Table  8.9  may  be  explained 
in  a  similar  manner. 

The  next  item  to  be  discussed  is  that  of  testing  various  hypotheses 
associated  with  the  regression  function.  Each  hi  may  be  used  to  test 
the  hypothesis  H:(3i  =  Q.  This  is  accomplished  by  computing 

t  =  bi/ssVc^  (8.72) 

with  v  =  n  —  q  degrees  of  freedom.  In  Table  8.9,  g  =  5,  and  thus 
n  —  q  =  n  —  5.  An  equivalent  test  procedure  is  provided  by 

2 

F  —  (&s/£^-)/(residual  mean  square).  (8.73) 

It  should  be  noted  that  the  various  J^-tests  defined  by  Equation  (8.73) 
are  not  all  independent  since  the  X  variables  are  correlated.  However, 
the  tests  provide  useful  information  if  interpreted  wisely.  Each  of  the 
mean  squares  reported  in  Table  8.9  may  also  be  used  to  form  J^-ratios 
which  provide  additional  important  information.  These  /^-ratios, 
defined  by 

mean  square  due  to  b*  \  bo,  &i,  -  -  -  ,  bi-i 

p  —  ,  ^g  m  74) 

residual  mean  square 

will  assess  the  significance  of  the  additional  reductions  in  the  residual 
sum  of  squares  achieved  by  fitting  the  b's  in  the  particular  order  adopted 
by  the  analyst.  The  italicized  words  emphasize  an  important  point :  The 
order  of  fitting  the  coefficients  has  a  decided  effect  on  the  analysis.  As 
we  shall  see  later,  if  the  variables  -X"i,  X2,  etc.,  represent  successive 
powers  of  a  single  variable  X  (i.e.,  the  mathematical  model  is  a  poly 
nomial  in  -XT),  then  a  natural  order  is  available.  In  other  cases,  the 
order  is  a  result  of  a  decision  on  the  part  of  the  analyst  to  write  the 
terms  of  the  model  in  a  specific  order  when  setting  up  the  Doolittle 
solution.  One  may,  of  course,  make  use  of 


S.S.  due  to  regression  =    s,  AivB^  (8.75) 

and  calculate 

mean  square  due  to  regression 

p  —  , — . ~ 

residual  mean  square 

^  ,  _  ,  (8-76) 


188  CHAPTER    8,    REGRESSION    ANALYSIS 

in  order  to  assess  the  over-all  significance  of  fitting  the  regression 
equation.  Other  tests  which  aid  in  determining  the  order  of  fitting  and 
the  choice  of  variables  to  be  retained  in  the  regression  equation  are 
available.  However,  discussion  of  these  is  not  warranted  in  this  book. 
Instead,  the  reader  is  referred  to  other  sources,  such  as  Hader  and 
Grandage  (10),  for  the  pertinent  details. 

While  not  discussed  at  this  time,  it  should  be  clear  to  the  reader  that 
confidence  interval  estimates  may  be  obtained  through  xise  of  tech 
niques  discussed  in  Chapter  6,  The  appropriate  standard  errors  are 
defined  in  Section  8.15.  For  further  details  the  reader  is  again  referred 
to  Hader  and  Grandage  (10). 

8-17      TESTS    FOR    LACK  OF   FIT 

In  Section  8.11  the  assumption  was  made  that  the  failure  of  the 
model  to  fit  the  observations  exactly  was  solely  a  function  of  the 
errors.  This  assumption  is  seldom  true,  although  it  may  be  nearly  so 
in  many  cases.  To  check  on  its  validity,  one  must  have  some  measure 
of  error  other  than  that  provided  by  the  residual  sum  of  squares*  The 
only  way  to  obtain  such  a  measure  is  to  insist  that  the  experiment  be 
repeated  some  number  of  times  at  at  least  one  value  of  the  independent 
(or  controlled)  variable.  In  addition,  it  is  also  wise  to  insist  on  running 
the  experiment  at  as  many  different  values  of  the  controlled  variable 
as  is  feasible.  In  the  example  considered  in  Section  8.9,  the  latter 
recommendation  was  followed  but  no  repetition  of  the  experiment  at 
any  value  of  the  controlled  variable  was  undertaken.  This  enabled  us 
to  make  a  visual  judgment  about  lack  of  fit  but  no  statistical  analysis 
was  possible. 

To  examine  the  statistical  test  for  lack  of  fit,  consider  the  data 
reported  by  Hunter  (11).  These  data  are  reproduced  in  Table  8.14, 
Introducing  the  dummy  variable,  -ST0,  which  is  identically  equal  to  1, 
and  using  the  Abbreviated  Doolittle  Method,  the  simple  linear  re 
gression  equation  is  determined  to  be  ^=1.76+2.86^3.  The  associ 
ated  analysis  of  variance  is  given  in  Table  8.15. 

In  this  example,  however,  it  is  possible  to  subdivide  the  residual 
sum  of  squares  into  two  parts:  one  part  being  an  estimate  of  experi 
mental  error  and  the  other  a  measure  of  the  lack  of  fit  of  the  linear 
model.  The  reason  such  a  subdivision  is  possible  should  be  clear:  The 
researcher  was  careful  to  provide  some  replication  in  the  performance 
of  the  experiment.  To  actually  perform  this  subdivision  of  the  residual 
sum  of  squares,  it  is  easier  to  calculate  the  experimental  error  sum  of 
squares  and  then  obtain  the  lack  of  fit  sum  of  squares  by  s\ibtraction. 
The  experimental  error  sum  of  squares  is  foxand  by  pooling  the  sums  of 
squares  of  deviations  about  the  mean  for  each  value  of  the  independent 
variable,  that  is, 

Z    (Z  Y*~  (Z  YY/n]  (8.77) 

all  X 


8.17      TESTS    FOR    LACK    OF    FIT  189 

TABLE  8. 14-Percentage  of  Impurities  at  Different  Temperatures 


Coded  Temperature 

Per  Cent  of  Impurities 

Temperature  (°C.) 

-STi 

Y 

200  

1 

6.4 

200  

1 

5.6 

200  

1 

6.0 

210  

2 

7.5 

210  

2 

6.5 

220  

3 

8,3 

220  

3 

7.7 

230  

4 

11.7 

230  

4 

10.3 

240  

5 

17.6 

240  

5 

18.0 

240  

5 

18.4 

Reprinted  from:  J.  S.  Hunter,  "Determination  of  Optimum  Operating  Conditions  by 
Experimental  Methods,  Part  II- 1,  Models  and  Methods/'  Industrial  Quality  Control,  Vol. 
15,  No.  6,  pp.  16-24,  Dec.,  1958.  By  permission  of  the  author  and  editor. 

TABLE  8.15-First  Analysis  of  Variance  for  the  Data  of  Table  8.14 


Source  of  Variation 

Degrees 
of  Freedom 

Sum  of  Squares 

Mean  Square 

Due  to  60  

1 

1281.3333 

1281.3333 

Due  to  61  1  60  

1 

228.5715 

228.5715 

Residual  

10 

40  .  3952 

4.0395 

Total 

12 

1550.3000 

where  the  expression  within  the  braces  is  calculated  separately  for  each 
value  of  -X".  In  the  example, 
Experimental  Error  Sum  of  Squares  = 

{(6.4)2  +  (5.6)2  +  (6.0)*  -  (6.4  +  5.6  +  6.0)2/3} 
+  |(7,5)2  +  (6.5)2  ~  (7.5  + 6.5)  2/2} 
+  {  (8.3)*  +  (7.7)2  _  (8.3  +  7.7)2/2} 
+  {  (11. 7)2  +  (10.3)2  —  (11.7  +  10.3)  V2} 
H-  {  (17.6)2  H-  (18.0)2  +  (18.4)2  -  (17.6  +  18.0  +  18.4)2/3}  =  2.3000. 

Thus,  it  is  now  possible  to  record  the  results  as  in  Table  8.16.  In  this 
table,  the  experimental  error  mean  square  is  a  pooled  estimate  of  o*B 
that  is  uncontaminated  by  any  inadequacy  of  the  assumed  linear 


190  CHAPTER    8,    REGRESSION    ANALYSIS 

TABLE  8.16-Second  Analysis  of  Variance  for  the  Data  of  Table  8.14 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

Due  to  &o  

1 

1281.3333 

1281.3333 

Due  to  61  1  &o  

1 

228.5715 

228.5715 

Lack  of  fit  . 

3 

38.0952 

12.6984 

38.64** 

Experimental  error.  .  .  . 

7 

2.3000 

0.3286 

Total 

12 

1550.3000 

**  Significant  at  the  1  per  cent  level. 

model.  To  test  the  hypothesis  that  the  linear  model  is  appropriate  (i.e., 
no  real  lack  of  fit  exists),  it  is  legitimate  to  obtain  the  ratio  F=*  12.6984 
/0.3286  =  38.64  with  ^i==3  and  *>2  =  7  degrees  of  freedom.  Since  this 
exceeds  ^.99(3,7)  =8.45,  the  decision  is  reached  that  a  serious  lack  of  fit 
exists.  That  is,  the  hypothesis  of  no  lack  of  fit  is  rejected.  Another  way 
of  stating  this  conclusion  is  as  follows:  The  assumed  linear  model  in 
adequately  describes  the  data. 

8.18      NONLINEAR   MODELS 

If  the  /''-ratio  in  Table  8.16  had  turned  out  to  be  nonsignificant,  it 
would  have  been  concluded  that  the  linear  fit  was  probably  adequate. 
However,  since  the  linear  model  was  judged  to  be  inadequate,  the  re 
searcher  is  obliged  to  consider  fitting  some  nonlinear  model.  That  is,  he 
should  attempt  to  discover  a  different  mathematical  model  which  bet 
ter  describes  (or  represents)  the  observations. 

There  are  many  alternatives  to  be  considered  at  this  stage.  For 
example,  should  a  higher  degree  polynomial  be  investigated  or  should 
some  other  functional  form  be  considered?  Pox-haps  some  exponential 
function  might  be  the  appropriate  model  for  the  problem  under  investi 
gation.  A  few  mathematical  models,  other  than  polynomials,  which  are 
frequently  encountered  in  applied  work  are 


77  =«  o>{3 ; 

In  97  »  In-y  +  (lnc*)/3-T; 

1/77  -  T  +  <*/3x; 

77  - 


oj  >  0,  ft  >  0  (8,78) 

a  >  0,  ft  >  0,   y  >  0            (8.79) 

a  >  0,  £  >  0,  r  >  0            (8 . 80) 

0  >  0,  y  >  0.  (8.81) 

These  are  xisually  referred  to  as:  (1)  the  simple  exponential  or  com 
pound  interest  function,  (2)  the  Gompertz  function,  (3)  the  logistic 
function,  and  (4)  the  Mitscherlich  function,  respectively. 

The  selection  of  an  alternative  to  the  linear  model,  i.e.  to  the  first 
degree  polynomial,  is  not 'easy.  The  choice  should  be  made  only  after 
careful  consideration  of  the  basic  mechanism  of  the  system.  If  this  is 
not  feasible,  scatter  plots  should  be  examined.  When  it  is  evident  that 


8.19      SECOND    ORDER    MODELS 


191 


some  degree  of  curvature  is  present  in  the  data  but  no  clear-cut  choice 
of  mathematical  model  is  possible,  a  reasonable  approach  is  to  syste 
matically  examine  polynomials  of  increasing  order  (i.e.,  of  higher 
degree) . 

8.19     SECOND  ORDER   MODELS 

Since  the  first  order  model  (a  straight  line)  was  an  inadequate  repre 
sentation  of  the  data  in  Table  8.14,  it  is  now  proposed  that  a  second 
order  (quadratic)  model  be  investigated.  That  is,  the  model 


f) /^  ~y~    \  /•?  ~y"  —i-  /•?    Y~  f  Q   fto^ 

in  which  X 0  is  identically  1,  will  be  considered.  Writing  the  data  in 
matrix  form,  namely, 


F  = 


6.4' 
5.6 
6.0 
7.5 
6.5 
8.3 
7.7 
11.7 
10.3 
17.6 
18.0 
L18.4 


1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 


1 

1 

1 

1 

1 

1 

2 

4 

2 

4 

3 

9 

3 

9 

4 

16 

4 

16 

5 

25 

5 

25 

5 

25  _ 

and  using  the  Abbreviated   Doolittle  Method   (see  Table  8.17),   the 
regression  equation  is  found  to  be 

1t  ==  8.43  —  3.14-XTi  +  l.OOXi.  (8.83) 

The  accompanying  analysis  of  variance  is  given  in  Table  8.18. 

A  comparison  of  Tables  8.16  and  8.18  indicates:  (1)  Fitting  a  quad 
ratic  term  led  to  a  significant  reduction  in  the  lack  of  fit  sum  of  squares, 
and  (2)  there  is  still  a  significant  lack  of  fit.  In  other  words,  although 
the  quadratic  equation  was  a  better  fit  than  the  linear  equation,  the 
second  degree  polynomial  does  not  adequately  describe  the  data. 
What,  then,  should  be  the  next  step?  Should  higher  degree  polynomials 
be  investigated  or  should  attention  be  directed  toward  some  other 
functional  form?  As  indicated  in  Section  8.18,  the  answer  to  such  a 
question  is  not  easily  obtained.  In  fact,  a  "best'7  answer  may  not  exist. 


192 


CHAPTER    8,    REGRESSION    ANALYSIS 


TABLE   8.17-The  Fitting  of  F  =  60+&i^i+Z>u^i  to  the  Data  of  Table  8.14 
by  the  Abbreviated  Doolittle  Method 


Front 

Half 

Row 

bo          bi 

611 

Terms 

Back  Half 

Check 

(0)  

12          36 

136 

124 

1 

0 

0 

309 

(1)  .  .     . 

136 

576 

452 

1 

o 

1201 

(2>  :  .  :  :  : 

2584 

1920 

1 

5217 

(3)  

12          36 

136 

124 

1 

0 

0 

309 

(4)  

1            3 

11    333333 

10.333333 

0.083333 

o 

o 

25.75 

(5)  

28 

168   000012 

80.000012 

—2.999988 

1 

o 

274 

(6)     . 

1 

6 

2   857143 

—0.  107142 

0.035714 

o 

9.785714 

(7)  .... 

34  666640 

34  666654 

6,666569 

—  5.999952 

1 

70.999931 

(8) 

1 

1 

0   192305 

—0   173076 

0  028846 

2.048077 

TABLE   8.18-Third  Analysis  of  Variance  for  the  Data  of  Table  8.14 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

jF-Ratio 

Due  to  &o 

1 

1281   3333 

1281   3333 

Due  to  bi  5o  

1 

228.5715 

228.5715 

695.59** 

Due  to  bn   60,  b\  

1 

34.6667 

34.6667 

105.50** 

Lack  of  fit  

2 

3.4285 

1.7142 

5.22* 

Experimental  error  .... 

7 

2  .  3000 

0.3286 

Total 

12 

1550.3000 

Significant  at  the  5  per  cent  level. 
'*  Significant  at  the  1  per  cent  level. 


Thus,  rather  than  say  what  should  be  done,  it  seems  politic  to  suggest 
a  procedure  which  can  be  modified  at  the  discretion  of  the  analyst.  The 
suggested  procedure  is:  Rather  than  seek  a  better  fit  in  terms  of  a 
higher  degree  polynomial  (i.e.,  a  degree  greater  than  2),  it  is  probably 
wiser  to  cast  about  for  some  other  functional  form  to  represent  the 
data.  In  the  case  under  consideration  (i.e.,  the  data  of  Table  8.14),  an 
examination  of  a  scattergram  would  suggest  that  an  exponential  func 
tion  be  given  serious  consideration.  The  suggested  procedure  does  not, 
of  course,  preclude  the  fitting  of  higher  degree  polynomials  if  that  ap 
pears  to  be  the  proper  approach.  For  example,  if  we  consider  the  data 
analysed  in  Section  8.9,  a  third-degree  polynomial  will  give  an  excel 
lent  fit*  Of  course,  some  form  of  exponential,  perhaps  of  the  logistic 
variety,  might  also  be  appropriate. 

8-20     ORTHOGONAL   POLYNOMIALS 

When  the  values  of  X  (the  independent  variable)  are  equally  spaced, 
there  is  another  method  of  fitting  polynomial  regression  equations 
which  has  much  to  recommend  it.  This  is  the  method  of  orthogonal 
polynomials.  You  will  have  noticed,  when  fitting  polynomials  by  the 


8.20       ORTHOGONAL    POLYNOMIALS  193 

method  described  earlier,  that  each  time  it  was  necessary  to  start  the 
solution  from  the  beginning  and  solve  a  new  set  of  normal  equations. 
For  example,  when  fitting  a  second  order  model  we  were  unable  to  use 
the  results  of  fitting  the  first  order  model.  However,  when  the  data 
permit  the  use  of  orthogonal  polynomial  techniques,  we  can  salvage 
the  previous  results  and  simply  perform  the  calculations  required  to 
add  a  new  term  to  the  polynomial  (of  one  less  degree)  determined  at  the 
preceding  stage. 

The  method  of  orthogonal  polynomials  will  be  illustrated  only  for 
the  case  where  the  values  of  X  are  equally  spaced  at  unit  intervals  and 
where  each  X  has  but  one  Y  value  associated  with  it.  If  the  X  values 
are  equally  spaced  at  intervals  not  equal  to  unity,  we  may  code  the  X 
values  by  dividing  through  by  the  length  of  the  common  interval  and 
then  proceed  in  the  manner  to  be  developed  below.  If  there  is  more  than 
one  Y  value  associated  with  each  X,  the  method  is  not  applicable  un 
less  we  have  an  equal  number  of  Y  values  associated  with  each  X  value* 
In  the  latter  case,  the  complete  solution  may  be  obtained  by  introduc 
ing  the  proper  divisor  into  the  calculations.  If  the  X  values  are  un 
equally  spaced,  a  solution  may  be  otained  [see  Kendall  (13)],  but  the 
operation  is  so  cumbersome  that  it  will  not  be  presented  in  this  text. 
Thus,  in  all  but  the  simpler  cases,  it  is  usually  better  for  the  research 
worker  to  use  the  method  described  earlier  in  this  chapter.  However,  if 
the  experimental  data  are  amenable  to  simple  treatment  by  the  method 
of  orthogonal  polynomials,  the  research  worker  is  advised  to  use  that 
method,  for  it  saves  time  and  also  allows  him  to  calculate  and  evaluate 
readily,  step  by  step,  the  contribution  made  by  fitting  each  additional 
term  in  the  regression  function. 

What,  then,  is  the  ^method  of  orthogonal  polynomials?  It  may  be 
shown  that  any  polynomial,  for  example, 

Y  =  60  +  biX  +  -  -  -  +  bkXk  (8.84) 

may  be  rewritten  as 

?  -  A,  +  AI&+  •  -  •  +  A*&  (8.85) 

in  which  the  f£  (z  =  l,  -•-,&)  are  orthogonal  polynomials5  and  the 
A.i  (i  =  0,  •  •  -  ,  k)  are  constants  defined  by 

Ao  =  $2  Y/n  -    Y  (8.86) 

and 

£' 

t'\*  '  *«  1,  ••-,*•  (8-87) 

siv 

For  the  case  we  are  considering,  that  is,  where  X  takes  on  the  values 

*  Two  polynomials  are  said  to  be  orthogonal  if,  when  X  takes  on  a  specified  set 
of  values,  S&-£jb«aO  for  i^k,  where  the  summation  means  that  we  first  compute 
the  product  k'ikL  f°r  each  value  of  X  and  then  obtain  the  sum  of  these  products. 


194 

1,2, 


CHAPTER    8,    REGRESSION    ANALYSIS 

,  n,  the  first  three  orthogonal  polynomials  may  be  expressed  as 
~  Z),  (8.88) 


-  (x  - 


7n 

J, 


(8.89) 


(8.90) 


where  the  X*  are  constants  (depending  on  ri)  chosen  so  that  the  values 
of  the  £"s  are  integers  reduced  to  their  lowest  terms.  An  abbreviated 
table  of  £  values  is  given  in  Table  8.19;  a  more  complete  table  may  be 
found  in  Anderson  and  Houseman  (1), 

TABLE  8.19-Partial  Table  of  %  Values 


Degree  of  Polynomial 


A-l 

&  =5  2 

£  =  3 

&«4 

n 

« 

€!      & 

«        3        fc' 

tf 

£2        la 

« 

i  

_! 

—  i     +1 

—  3      +1      —1 

—  2 

+  2      —1 

+  1 

2  

+  1 

0      —  2 

—1      —  1      +3 

—  1      +2 

—  4 

3  

-fl      +1 

0 

—  2          0 

4   .  . 

+3      +1      +1 

1           L  _  Q 

"^^   JL                 ~  ^ 

^ 

5 

+  2 

+  2     4-1 

+  1 

Before  considering  a  numerical  example,  the  equations  necessary  for 
calculation  of  the  various  sums  of  squares  will  be  given.  They  are  as 
follows  : 

(8.91) 

r~ 

(8.92) 


S.S,  due  to  bQ 

S.S,  due  to  fitting  the  ith  degree  term  =====  .4t-(5D 


Example  8.7 

Consider  the  data  of  Table  8.3  and  rewrite  the.  values  in  the  form 
shown  in  Table  8.20.  Equations  (8. 80)  and  (8.87)  then  yield  f^  — 56.3 
+7.478£i  —  .0365&  —  .6801&.  If  Equations  (8.88)  through  (8.90)  are 
evaluated  as  far  as  possible  by  using  the  known  values  of  n,  3Tand  the  X's, 
and  then  substituted  in  the  regression  equation  just  found,  we  obtain 
f  =  21.7277  —  5.8458^+2.3449X^  —  0. 1134AX  The  reduction  in  sum  of 
squares  due  to  fitting  the  various  terms  in  the  regression  function  may, 
of  course,  be  calculated  using  Equations  (8.91)  and  (8.92). 

8.21      SIMPLE   EXPONENTIAL   REGRESSION 

The  regression  function  specified  by  Equation  (8.78)  is  frequently 
encountered  in  experimental  work*  Thus,  it  is  appropriate  that  some 


8.21        SIMPLE    EXPONENTIAL    REGRESSION 

Table  8.20-Table  for  Calculating  the  A's  and  Corresponding 

Suras  of  Squares 


195 


F 

5i 

8 

« 

ni 

F« 

F« 

17  

—  6 

22 

—  11 

—  102 

374 

—  187 

21  

—  5 

11 

o 

—  105 

231 

o 

22  

—  4 

2 

6 

—  88 

44 

132 

27  

—  3 

8 

—  81 

—  135 

216 

36.  .. 

—  2 

—  10 

7 

—  72 

—  360 

252 

49  . 

H 

—  13 

4 

—  49 

—  637 

196 

56  

o 

—  14 

o 

0 

—  784 

0 

64  

1 

—  13 

—  4 

64 

—  832 

—  256 

80.  .. 

2 

—  10 

—  7 

160 

—  800 

—  560 

86  

3 

—  5 

—  8 

258 

—  430 

—  688 

88  

4 

2 

—  6 

352 

176 

—  528 

92  

5 

11 

0 

460 

1012 

0 

94  

6 

22 

11 

564 

2068 

1034 

732.  .  . 

0 

0 

o 

1361 

—  73 

—  389 

X  

1 

1 

1/6 

S  (£02-- 

182 

2002 

572 

discussion  of  the  associated  methods  of  analysis  be  given.  If  the  method 
of  least  squares  is  used,  the  resulting  normal  equations  are  not  amen 
able  to  easy  solution.  Consequently,  some  other  (approximate)  ap 
proach  is  necessary.  The  usual  approximate  procedure  adopted  is  to 
take  logarithms  (logarithms  to  the  base  10  are  most  convenient)  which 
results  in 


log  17  =  log  a  +  (log  ff)X. 

Redesignating  the  quantities  as  f olio ws :  Z  =  log 
and  W  —  X,  Equation  (8.93)  appears  as 

Z  =  A  +  BW 


j,  A  ==log  <x, 


(8.93) 
'  =  log/?, 

(8.94) 


and  we  immediately  recognize  this  as  being  of  the  sameJForm  as  Equa 
tion  (8.12).  Estimates  of  A  and  B,  denoted  by  A  and  B,  may  then  be 
found  following  the  methods  described  earlier.  This  solution,  which  is 
equivalent  to  fitting  a  straight  line  by  least  squares  to  the  data  when 
plotted  on  semi-logarithmic  paper,  is  not  identical  with  a  least  squares 
solution  of  the  original  problem  using  Equation  (8.78)  and  ordinary 
graph  paper.  However,  the  approximation  is  sufficiently  accurate  for 
most  problems. 


Example  8.8 

Consider  the  data  of  Table  8.21.  Using  either  the  method  of  Section 


196  CHAPTER    8,    REGRESSION    ANALYSIS 

8.7    or    the    Abbreviated    Doolittle    Method,    it    is    determined    that 
A  =0.9469  and  J3  =  0.002576. 

TABLE  8.21-Protein.  Content  and  Proportion  of  Vitreous  Kernels  in 

Samples  of  Wheat 


Sample  Number 

Proportion  of 
Vitreous  Kernels 
X(-WO 

Protein  Content  F 

Z  =  log  F 

1    

6 

10.3 

1.013 

2        

75 

12.2 

1,086 

3                 

87 

14.5 

1.161 

4  

55 

11.1 

1.045 

5                     ... 

34 

10.9 

1.037 

6  

98 

18.1 

1.258 

7                 

91 

14.0 

1.146 

8            

45 

10.8 

1.033 

9  

51 

11.4 

1.057 

10   

17 

11.0 

1.041 

11            

36 

10.2 

1.009 

12                 

97 

17.0 

1.230 

13            

74 

13.8 

1.140 

14  

24 

10.1 

1.004 

15        

85 

14.4 

1,158 

16          

96 

15.8 

1.199 

17                     .... 

92 

15.6 

1  .  193 

18  

94 

15,0 

1.176 

19   

84 

13.3 

1.124 

20   

99 

19.0 

1.279 

Reproduced  from  M.  TCzektel,  Methods  of  Correlation  Analysis  (New  York:  John  Wiley 
and  Sons,  Inc.,  1941),  p.  82.  By  permission  of  the  author  and  publishers. 

8-22     THE   SPECIAL  CASE:77=/3X 

In  some  instances,  it  is  reasonable  to  assume  that  the  true  regression 
line  passes  through  the  origin.  That  IB,  if  simple  linear  regression  is  be 
ing  considered,  /5o  in 

&    __L_  o  v  f8  OS"i 

^     =     PQ    •+•    plJ\  V.O.^.*V 

is  assumed  to  be  0  and  Equation  (8.95)  is  rewritten  as 

(8.96) 


where,  of  course,  |S«=£i.  It  is  clear  that  such  an  assumption,  if  justified, 
will  simplify  the  calculations!  procedures.  It  can  be  verified  that  for 
this  special  ease 

j^^ga-  ^  XY/  23  X*-  (8.97) 

Please  do  not  make  the  mistake  of  adopting  this  simpler  form  just 
because  it  is  easier  to  handle.  Further,  even  if  the  assumption  is  justi- 


8.23       WEIGHTED    REGRESSIONS  197 

fied  (such  as  when  X  =  height  and  3^  — weight  of  men),  it  may  be  better 
to  forego  the  simplifying  assumption  and  consider  *7  =  /3o+/3i.-X"  as  being 
more  appropriate  for  the  range  of  X  values  being  studied  in  the  experi 
ment. 

In  summary,  the  mathematical  model  should  only  be  chosen  after 
proper  consideration  has  been  given  to  all  the  factors  involved. 

8,23      WEIGHTED    REGRESSIONS 

Suppose  the  data  to  be  considered  are  of  such  a  nature  that  the  as 
sumption  of  homoscedasticity  (i.e.,,  homogeneous  variances)  is  no  longer 
justified.  That  is,  suppose  we  can  not  assume  that  o^-\^  is  the  same  for 
all  X,  but  that  we  must  assume 


where  the  w^  are  known  constants.  If  -we  restrict  ourselves  to  the  case 
where  77  =  /3v+-@iX,  it  may  be  shown  that  the  resulting  normal  equa 
tions  are 


/    k  \  /    k  \  k     ni 

I    T^  n*Wi  J  £0  -f-  (    23   niW+Xi  J  &!  =   ]>3  23  Wi  Y^ 

\  i=*\  /  \  i^\  /  i««i  J--.1 

/       k  X  /       k  2\  k         n-t 

(    53  ntWiXt  )  60  +  (    23  n^WiXi    )  61  =  23  23  WiXiY+j 
\  *— i  /  \  i*=i  /  t=i  j— i 


(8.99) 


where 

F<,  =  /30  +  ^Si^i^  H-  6,v;          i  «  1,  •  —  ,  fe  (8,100) 

j  =  1,  -  •  -  ,  »f. 

It  is  of  particular  interest  to  consider  the  case  where  erf  is  propor 
tional  to  -XT;,  that  is,  where  we  may  write 

v\  =  <r2A0*  =  o-2-X"i,  (8.101) 

since  this  is  a  fairly  common  occurrence  in  certain  areas  of  experimen 
tation.  Under  these  conditions,  the  normal  equations  reduce  to 


z^l 

(»)«o+(i:  «<**)&!-  2:2: 

where 

fl    :===i     ^  ^    fl<i. 


198  CHAPTER    8,    REGRESSION    ANALYSIS 

In  this  special  case,  a2  is  estimated  by 


-  2).  (8.  103) 

t—i  y-i 

8.24     SAMPLING  FROM  A  B1VAR1ATE  NORMAL  POPULA 
TION 

Let  us  consider,  as  far  as  practicable,  the  consequences  of  obtaining 
a  random  sample  of  values  of  both  X  and  Y  from  some  bivariate  popu 
lation  rather  than  first  choosing  values  of  X  and  then  observing  ran 
dom  F  values  associated  with  these  chosen  X  values.  What  effect  will 
such  a  procedure  have  on  our  estimates?  As  we  have  stated  the  prob 
lem,  it  is  much  too  general  to  permit  a  satisfactory  answer  in  this  book. 
However,  if  we  make  the  assumption  that  our  bivariate  population  is  a 
bivariate  normal  population,  then  we  may  examine  the  effect  of  obtain 
ing  random  pairs  of  X  and  Y  rather  than  choosing  X  values  and  then 
observing  the  random  values  of  Y  associated  with  the  selected  values 
of  X. 

In  this  case,  two  approaches  are  possible:  (1)  Obtain  the  best  regres 
sion  equation  for  estimating  a  value  of  Y  associated  with  a  specified 
value  of  -XT;  (2)  obtain  the  best  regression  equation  for  estimating  a 
value  of  X  associated  with  a  specified  value  of  Y.  That  is,  we  can  obtain 

$  =  60  +  &1-3T  (8.104) 

as  in  Section  8.7,  or  we  can  obtain 

£  =  co  +  CiF  (8.105) 

where 

c*=*°X  —  tf,T  (8.106) 

and 

ci  -  Z>3>/  52  y*.  (8.107) 

It  is  to  be  noted  that  the  above  relations  assume  no  "errors  of  meas 
urement"  in  X  and  Y.  If,  however,  our  variables  arc  subject  to  errors 
in  measurement,  so  that  we  really  observe  Z  —  X  +  *  and  W—  F+S, 
where  €  and  5  arc  independently  and  normally  distributed  with  moan  0 
and  variances  <r«  and  cr*>  respectively,  what  estimation  procedure  may 
we  use?  If,  in  the  future,  we  measure  %  and  wish  to  estimate  Y,  the 
regression  of  W  on  %  should  be  used  ;  if  we  measure  W  and  wish  to  esti 
mate  .XT,  the  regression  of  Z  on  W  should  be  calculated  and  used. 

A  reasonable  question  to  ask  at  this  point  is:  What  effects  do  the 
above-mentioned  errors  of  measurement  have  on  the  accuracy  and  pre 
cision  of  our  estimates?  Some  answers  arc: 

(1)   Tf  the  random,  errors  of  measurement  are  associated  only  with 
the  dependent  variable,  and  are  not  related  to  the  true  values, 


8.25       ADJUSTED    Y   VALUES  199 

they  will  not  affect  our  estimate  of  the  true  slope  but  will 
cause  s%  to  overestimate  <TE. 

(2)  If  the  random  errors  of  measurement  are  associated  only  with 
the  independent   variable,   and  are  not   related  to  the  true 
values,  they  not  only  cause  SE  to  overestimate  o-E  but  also 
tend  to  produce  underestimates  of  the  true  slope. 

(3)  If  both  variables  are  subject  to  error,  the  consequences  are  not 
so  easily  determined,  and  much  care  should  be  taken  when 
making  predictions  based  on  such  data. 

Suppose,  however,  that  we  want  to  estimate  the  true  relationship 
between  X  and  Y.  To  accomplish  this  we  need  further  information 
about  v\  and  cr28.  Such  information  (i.e.,  estimates  s«  and  sf)  can 
sometimes  be  obtained  by  making  duplicate  measurements.  However, 
such  a  procedure  is  not  always  possible.  When  duplicate  measure 
ments  are  not  available,  other  approaches  must  be  explored.  Many 
scholars  have  considered  the  problems  associated  with  regression  anal 
yses  in  which  both  variables  are  subject  to  error,  and  several  solutions 
have  been  proposed.  However,  because  no  general  optimal  solution  has 
yet  been  obtained  and  because  the  subject  may  rightly  be  considered 
to  be  beyond  the  scope  of  this  text,  no  attempt  will  be  made  to  illus 
trate  any  of  the  proposed  methods  of  analysis. 

8.25     ADJUSTED    Y  VALUES 

Closely  related  to  the  reduction  in  sum  of  squares,  mentioned  sev 
eral  times  in  this  chapter,  is  the  technique  of  adjusting  values  of  the 
dependent  variable  to  take  account  of  differences  among  the  associ 
ated  values  of  the  independent  variable.  For  example,  if  we  are  con 
cerned  with  measurements  on  the  gains  in  weights  of  certain  animals,  a 
valid  comparison  among  the  gains  does  not  seem  possible  unless  we 
adjust  for  such  a  value  as  the  initial  weight  of  the  animals  or  the  feed 
consumed.  That  is,  if  one  animal  gains  60  pounds  while  consuming  300 
pounds  of  feed,  and  another  animal  gains  40  pounds  while  consuming 
200  pounds  of  feed,  we  do  not  feel  justified  in  making  a  direct  compari 
son  between  60  pounds  and  40  pounds.  We  should  first  attempt  to 
make  some  adjustment,  or  correction,  for  the  different  amounts  of  feed 
consumed.  One  way  to  make  such  an  adjustment  is  through  regression. 
If,  from  the  present  or  other  data,  we  have  an  estimated  regression 
function  Y  =  bo-{-biX,  where  F  =  gain  in  weight  and  J£T  =  feed  con 
sumed,  we  can  adjust  the  observed  gains  in  weight  to  some  common 
value  for  feed  consumed.  The  value  of  X  most  commonly  selected  is 
the  sample  mean  (X)  but  any  value  will  do.  The  reason  why  the  mean 
is  usually  adopted  as  the  point  of  comparison  is  that,  in  general,  it  is 
near  the  center  of  the  range  of  values  of  the  independent  variable. 

What,  then,  is  the  procedure  for  determining  adjusted  Y  values? 
The  formula  defining  adjusted  Y  values  (adjusted  to  X,  that  is)  is  as 
follows : 

adj.   Y  =    Y  -  bi(X  -  X)  (8.108) 


200 


CHAPTER    8,    REGRESSION    ANALYSIS 


and  the  nature  of  the  adjustment  is  illustrated  in  Figure  8.8.  Here  only 
three  sample  points  have  been  plotted  since  these  are  sufficient  to  illus 
trate  the  technique.  It  is  seen  that  all  the  adjusted  Y  values  (repre 
sented  by  circles)  appear  on  the  line  erected  vertically  at  X  because 
we  adjusted  to  X  =  "X.  Note  that  it  is  possible  to  have  adjusted  Y\ 
>  ad  justed  F2  even  when  Fx<  F2.  Thus,  it  is  readily  apparent  that  the 
adjustment  of  a  set  of  measurements  based  on  a  concomitant  variable 
may  completely  change  the  entire  picture  of  an  experiment.  As  a  con 
sequence,  we  might  reach  much  different  conclusions  based  on  an  anal 
ysis  of  adjusted  values  than  would  have  been  reached  if  no  account 
were  taken  of  the  functional  relation  existing  between  the  dependent 
and  independent  variables. 


X 


FIG.   8.8— Illustration  of  adjusted  Y  values. 

It  should  be  evident  that  adjusted  Y  values  may  also  be  determined 
when  dealing  with  other  than  simple  linear  regression.  For  example,  if  a 
multiple  linear  regression  analysis  has  been  performed  and  the  regres 
sion  cqiiation  determined  to  be 

f>  ~  fto  +  b,Xl  +  .  .  .  +  bkxk,  (8. 109) 

then  adjusted  Y  values  are  defined  by 

adj.  F  =*   F  -  bl(Xl  -  Z*)  -  b^X*  -  Z2)  -  — 

-  b*(X*  -  3"*).  (8.110) 

Equation  (8.110)  would,  of  course,  be  evaluated  using  the  appropriate 
sample  values  (Y^  Xu,  -  -  *  ,  XM)  and  the  calculated  mean  values. 

Rather  than  dwell  on  the  topic  of  adjxasted  values  at  this  time,  we 
shall  defer  further  discussion  until  later  in  the  book  where  a  more  effi 
cient  method  of  analysis,  namely,  covariance  analysis,  will  be  intro 
duced. 


8.26       THE    PROBLEM    OF    SEVERAL    SAMPLES    OR    GROUPS  2O1 

8.26     THE  PROBLEM  OF  SEVERAL  SAMPLES  OR  GROUPS 

In  this  section,  a  topic  of  considerable  importance  will  be  discussed. 
It  is :  Given  several  samples  or  groups  of  observations,  may  all  the  data 
be  pooled  into  one  large  sample?  This  sort  of  problem  has  arisen  earlier 
in  this  book  and  it  is  not  surprising  that  it  also  arises  when  dealing 
with  regression  analyses. 

Although  the  problem  can  arise  regardless  of  the  form  of  the  regres 
sion  function,  the  discussion  here  will  be  limited  to  the  case  of  simple 
linear  regression.  If  other  functional  forms  are  pertinent,  a  statistician 
should  be  consulted. 

When  several  sets  of  sample  data  are  available,  the  question  most 
frequently  asked  is:  Can  one  regression  line  be  used  for  all  the  data? 
More  specialized  questions  are : 

(1)  Taking  liberties  with  the  system  of  notation  adhered  to  up  to 
this  point  and  letting  6t-  represent  the  estimate  of  fit,  where  6* 
is  the  estimated  slope  for  the  ith  group  and  /?»  is  the  true  slope 
of  the  regression  function  in  the  population  from  which  the 
ith  group  is  a  sample,  does /3i  ==/32  =   -  -  •  =/3k?  In  other  words, 
are  all  the  sample  slopes  estimates  of  the  same  true  slope? 

(2)  Assuming  /3i  =  /32=   •  •  •  —Pk,  would  a  regression  fitted  to  the 
group  means  be  linear? 

(3)  Assuming  /3i  =  /32=   -  -  •  —  j8*  and  that  the  regression  of  the 
group  means  is  linear,  is  {3w  =  l3M,  where  &M  is  the  true  regres 
sion  coefficient  for  the  means  and  j3w  is  the  true  pooled  within 
groups  regression  coefficient? 

To  mention  one  case  where  it  is  necessary  to  know  the  answers  to  the 
questions  stated  above,  we  cite  the  technique  known  as  analysis  of  co- 
variance  which  we  shall  study  in  detail  in  a  later  chapter.  This  tech 
nique  has  as  one  of  its  basic  assumptions  the  requirement  that  the  same 
regression  coefficient,  /3,  apply  to  all  groups.  Hence,  the  need  for  an 
appropriate  test  is  clear.  Let  us  now  outline  the  general  procedure  to  be 
followed.  Suppose  we  have  k  groups  and  n*  observations  (on  both  X 
and  Y)  in  the  zth  group.  We  may  present  most  of  the  necessary  calcu 
lations  in  Table  8.22,  where 


(8.111) 

"*  *<  V     J  — I  /      \     7-.1 

5*=  S  (Xf  —  J?i)(F^-~- 7*)=  y^jy^-Ft (8.112) 

J-l  y-1 


2O2  CHAPTER    8,    REGRESSION    ANALYSIS 

2 


/      nt  \ 

(2tF«) 


_       -.   . 

:  (F<,  -   F,)2  =  Z)  F*  --  —  -  >  (8-  113) 

—  i  y«-i 


and 


(&  Tli 

2:2:^ 
i*-l  J«=l 


jo»i  i^l    ^=0=1  ~_^ 

>  .    Wi 
t—  1 

f:  (-YO-  - 


k 

n.  ^ 


z;^)  2:2:  F« 

=  2:2:  ^-F,-  -    ±±±^  —  ^^i  —  ,         (8.115) 


/         k  Hi 

( Ti  S  7" 


If  we  designate  by  >S2  the  sum  of  squares  among  the  k  group  regres 
sion  coefficients,  that  is,  if  $2  is  a  sum  of  squares  expressing  the  amount 
of  variation  among  61,  62,  •••,&*,  where  bj  is  the  regression  coefficient 
in  the  jfth  group,  it  may  be  shown  that 

2 


02  ==  ^w  — 

A 

(8.117) 


y;  J!i  _  J 


Similarly,  we  may  designate  the  sum  of  squares  of  the  deviations  of  the 
F-means  from  the  regression  of  y-means  on  X-means  by  ASf3,  where  this 
is  calculated  as  shown  in  Table  8.22.  The  square  of  the  difference  be 
tween  the  regression  coefficient  computed  from  the  "pooled  within" 
values  (6^-)  and  the  regression  coefficient  for  the  regression  of  the 
means  (6^)  is  given  by  (6^—  &jir)2-  If  we  multiply  this  by  a  suitable  fac 
tor,  it  becomes  another  estimate  of  cr^,  assuming,  of  course,  a  constant 
variance  of  Y  for  all  X.  This  may  be  expressed  as 

(8.118) 


CD 
O 

0 

rt 
o 


CD 
>-l 
bJD 


w 


O 

'5 


(U 

H 


1 

3 


00 


§ 

7 

cr 

CO 

7    -w.l 

1 

-W3  \  w 

>  i  S 

Degrees  of 
Freedom 

<N      CS               CS 

1         1                   1 

s    £         S 

rH 
^                      | 

I          1      „ 

-W3  «W3  I 

-W3 

w 

°i    4 

w 

L           ^      Pq 

4 

J 

^^      «> 

*W3   i   ^3 

i 

w 

L  L      L 

Cj    0?         O 

I     ~h    I 

^0            ^O        ^-O 

ii 

6-    f 

w 

<3J       3 

*W3  ^ 
II     II 

^ 

^   ^5 

w 

*«       < 

-W3  p§ 

ii     ii 

fe     ^ 

pq      pq^ 

EH 

^  f 

w 

^     X              ^ 

|  | 

^ 

Degrees  of 
Freedom 

rH      r-H                 -rH 

L    '           L 

i_ 

-W3  I 

tH 

I 

*wr« 

CL. 
o 

M 

I 

0 

^   M  •  •  •-t* 

giji 

204  CHAPTER    3,    REGRESSION    ANALYSIS 

and  noting  that  in  Table  8.22  we  defined  Si  as  the  pooled  sum  of 
squares  of  deviations  from  regression,  we  can  show  that 

ST  ==  .Si  +  S*  +  ^3  +  6V  (8.  119) 

Furthermore,  the  degrees  of  freedom  associated  with  ST  may  be  sub 
divided  in  the  following  fashion : 

"   +  (k  -  1)  +  (k  -  2)  +  1,        (8.  120) 


JT)  n.  _  2  =  (  32  »<  - 

ml  \    i«l 


and  these  are  associated  with  S±,  S%,  /S3,  and  S4,  respectively. 

Now  we  are  in  a  position  to  answer  the  questions  posed  at  the  begin 
ning  of  this  section.  Let  us  consider  these  in  order  and  indicate  the 
proper  test  procedures. 

1.   Can  one  regression  line  be  used  for  all  the  observations? 

(8.121) 


/ '(£»  - 


2.  To  test  11:01—  •  •  •  =/3&:  This  test  and  the  succeeding  ones  are 
usually  cozisidered  if  F  in  Equation  (8.121)  turns  out  to  be  significant. 
That  is,  we  are  curious  as  to  the  reason  for  the  significance. 

S*/(k  —  1) 


3.    To     test     whether     regression     of     means     is     linear     (assuming 

S*/(k  —  2) 

(8.123) 


(Si 

'      \  *— i 

4.    To  test  H\$W  —  $M   (assuming  regression  of  means  is  linear  and 


It  should  bo  clear  that  the  order  in  which  those  tests  arc  performed  is 
very  important  since  the  assumptions  necessary  for  the  later  tests  are 
tested  as  hypotheses  in  the  earlier  tests.  Note  also  that  if  a  sequence  of 
tests  is  applied,  the  critical  level  (true  probability  of  Type  I  error)  of 
the  sequence  is  not  known  though  it  is  needed  for  proper  interpretation 
of  the  results. 


8.27       SOME    USES    OF    REGRESSION    ANALYSIS 


205 


Example  8.9 


Consider  Table  8.23.  To  test  H:/3i  =/32=/33  using  a:  =  0.01,  we  calculate 
F==  150/15.  667  =  9.  5  with  j>i=2  and  7^  =  300  degrees  of  freedom.  Since 
F  —  9.5  >  7^0.99(2,300),  the  stated  hypothesis  is  rejected.  The  performance 
of  the  remaining  tests  is  left  as  an  exercise  for  the  reader. 

TABLE   8.23-Hypothetical  Data  to  Illustrate  the  Procedure 
for  Testing  the  Hypothesis 


Group 

Degrees 
of  Free 
dom 

Z*2 

!!C  *y 

Z^2 

Z:y2-(I>}02/:£*2 

Degrees 
of  Free 
dom 

Mean 
Square 

A  

101 
101 
101 

400 
200 
400 

800 
600 
600 

4000 
3000 
2000 

2400 
1200 
1100 

100 
100 
100 

B  

C  

Total 

303 

1000 

2000 

9000 

4700 
5000 

300 
302 

15.667 

Difference  for  testing  Hi  /3i=/32=/33 

300 

2 

150 

Granting  the  assumption  that  two  populations  have  a  common  vari 
ance,  the  hypothesis  HifBi  —  pz  may  be  tested  against  the  alternative 
by  examining 

*=(fti-  6aOA^^  (8.125) 

(8.126) 


where 


(-ar«  - 


y—  i 


and 


y—  i 


n\ 


—  4 


(8.127) 


The  value  of  i  specified  by  Equation  (8.125)  is,  of  course,  distributed  as 
"Student's"  t  with  *>  =  ni+n2  —  4:  degrees  of  freedom.  Confidence  limits 
for  01*—  02  may  also  be  obtained  by  use  of  the  foregoing  equations.  It  is 
to  be  noted  that  economies  in  calculation  may  be  achieved  by  select 
ing,  whenever  possible,  the  same  -X"  values  for  both  samples. 

8.27     SOME  USES  OF    REGRESSION  ANALYSIS 

The  uses  to  which  regression  techniques  may  be  put  are  numerous.  A 
few  of  the  more  important  are : 


206 


CHAPTER    8,    REGRESSION    ANALYSIS 


(1)  To  reduce  the  length  of  a  confidence  interval  when  estimating 
some  population  mean  (or  total)  by  considering  the  effect  of 
concomitant  variables. 

(2)  To  eliminate  certain  "environmental"  effects  from  our  esti 
mates  of  treatment  effects;  that  is,  we  may  wish  to  examine 
adjusted  Y-  values. 

(3)  To  predict  Y  knowing  values  of  X^  •  •  •  ,  Xk  (our  auxiliary 
variables)  whether  or  not  a  causal  relationship  exists. 

(4)  To  influence  the  outcome  of  the  dependent  variables  assum 
ing,  of  course,  that  we  have  a  causal  relationship. 

There  are  many  other  uses  for  regression  methods  which  might  have 
been  listed.  We  have  not  attempted  to  exhaust  the  possibilities,  nor 
have  we  attempted  to  give  our  examples  in  any  order  of  importance. 
The  relative  importance  of  the  different  uses  will  vary  depending  on 
the  subject  matter  being  discussed. 

Problems 

8.1        Derive  the  normal  equations  specified  by  Equation  (8.14). 


8.2 
8.3 
8.4 


8.5 


8.6 


8.7 


Given  the  following  values,  find: 

£>:*  =  121  Z;,Y  «  20 


Derive  Equation  (8.21). 

Derive  Equation  (8.22)  from  Equation  (8.25). 

iX,  (b)  SB,  (c)  sbi. 

-  82 

n  —  10 

Find  the  linear  regression  of  F  on  X  given  the  values; 

X:    3  8  4  11  9 
F:    5~"3~4"""l~2 

Given  that 

n  «  277,     5*  -  65,     F  -  72,     ]£>'  «  1600,     £y  -  3600,     Z^  -  2000, 

compute:  (a)  SB,  (b)  sbii  (c)  sp  for  -Y==45. 

Given  the  abbreviated  analysis  of  variance  shown  below,  perform  the 

following: 

(a)   Test  /7:j5i«0  using  a  =  0.01. 

(6)    Compute  the  standard  error  of  estimate,  $$. 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 

Square 

Due  to  regression  

1 

40 

40 

Deviations  about  regression  

SO 

200 

4 

Total 

51 

240 

8.8       Given  that 

n  -  38,     3?  «  5,     "F  «  40, 

answer  the  following: 
(a)   Determine  ?*==»( 


-  100, 


»*  10,000, 


-  -  800 


PROBLEMS  207 

(6)    Test  #  :/3i  =  0  using  a  =  0.05. 

(c)  Partition  ^y2  into  two  parts,  one  associated  with  the  slope  of  the 
linear  regression  and   the   other   associated   with  the   deviations 
about  regression. 

(d)  For    the    observation     (X  =  S,     F  =  36),     compute    the    adjusted 
value  of  Y. 

(e)  Interpret  both  61  and  /3i. 

8.9  Given  that 

n  =  62,     X  =  10,      Y  =  20,      £>*  =  40,      ^y*  -=  250,      £>;y  =  —  80, 

solve  the  following: 

(a)    Determine  y  =  60  + &i-X^. 

(6)  Compute  a  99  per  cent  confidence  interval  for  /3i.  State  all  as 
sumptions. 

(c)  Estimate  the  gain  in  information  from  using  X  as  a  statistical 
control  (the  regression  of  Y  on  -XT)  in  estimating  the  population 
mean  of  Y.  (NOTE:  Information  is  here  used  as  a  synonym  for 
"the  reciprocal  of  the  variance/7) 

8.10  Given  that  61  =  0.2  grams  of  gain  per  gram  of  feed  eaten,  find  the  net 
difference  between  the  gains  of  two  rats  where  one  animal  consumed 
200  grams  of  feed  and  gained  60  grams  while  the  other  animal  con 
sumed  300  grams  and  gained  90  grams. 

8.11  The  data  in  the  table  given  below  represent  the  heights  (Jf)  and  the 
weights  (F)  of  several  men.  We  selected  the  heights  in  advance  and 
then  observed  the   weights   of  a  random   group  of  men  having  the 
selected  heights. 


X 


60  in.  HOlbs. 

61  135 

60  120 

61  126 

62  140 
60  130 
62  135 
65  158 
64  145 
70  170 
72  185 
70  180 


(a)  Plot  a  scatter  diagram. 

(6)  Obtain  the  estimated  regression  line  F"  = 

(c)  Calculate  and  interpret  a  90  per  cent  confidence  interval  for  /3i. 

(d)  Calculate  and  interpret  a  98  per  cent  confidence  interval  for  /50- 

(e)  Calculate   and   interpret   a   95   per   cent   confidence   interval   for 


*~  66* 

(/)  Test  the  hypothesis  £T:/3i  =  0. 
(g)  Test  the  hypothesis  H:0i  =  G. 
W  Test  the  hypothesis  H:(3<>=  —30. 


2O8  CHAPTER    8,    REGRESSION    ANALYSIS 

(i)  Predict  the  weight  of  an  individual  who  is  66  inches  in  height. 
Give  a  "prediction  interval." 

(j)  Estimate  the  height  of  a  man  whose  recorded  weight  is  170  pounds. 
Give  both  point  and  interval  estimates. 

(&)    Test  for  "linearity  of  regression." 

(Z)  What  proportion  of  the  variation  in  Y  is  "explained"  by  the  re 
gression  of  weight  on  height? 

(NOTE:   Give  all  assumptions  and  use  a  probability  of  Type  I 
error  equal  to  .05  in  each  test.) 

8.12  Assuming  the  data  given  in  Problem  4.3  to  be  a  random  sample  from  a 
bivariate    normal   population,    (a)    calculate   the   regression   for   esti 
mating  weight  from  height,  (b)  calculate  the  regression  for  estimating 
height  from  weight,  (c)  plot  a  scatter  diagram  and  show  both  regres 
sion  lines  thereon. 

8.13  The  Consumer  Market  Data  Handbook,  1939  edition,  U.S.  Department 
of  Commerce,   lists   consumer  market  data  by  states,   counties,   and 
cities.  Among  the  types  of  information  listed  are   "Population  and 
Dwellings/'    "Volume   and  Type   of   Business   and   Industry,    1935," 
"Employment  and  Payrolls,  1935/'  "Retail  Distribution  by  Kinds  of 
Business,   1935/'  and   "Related  Indicators  of  Consumer  Purchasing 
Power. "  Among  the  latter  are  numbers  of  income  tax  returns,  auto 
mobile  registrations,  radios,  telephones,  electric  meters,  and  magazine 
subscribers. 

Such  information  as  listed  above  might  be  used  by  national  ad 
vertising  agencies,  large  sales  organizations,  and  by  individual  retail 
or  manufacturing  agencies  for  various  purposes  in  planning  their 
business  activities.  The  numbers  and  kinds  of  analyses  which  might 
bo  considered  for  such  data  arc  largo.  We  have  selected  only  a  small 
portion  for  study  in  this  problem. 

The  data  given  here  present  the  filling  station  sales  per  capita  (F) 
and  the  automobile  registrations  per  1000  persons  (-XT)  for  a  group  of 
Iowa  counties. 


PROBLEMS 


209 


IOWA  CONSUMER  DATA 


County 

Per  Capita 
Sales  of  Fill 
ing  Stations 
(yearly'} 

Automobile 
Registrations 
per  1OOO 

Persons 

Adair  

$17 

206 

Adams  ... 

25 

233 

Allamakee  . 

16 

237 

Appanoose  

13 

183 

Audubon  

28 

243 

Benton  

27 

230 

Black  Hawk 

20 

272 

Boone  

21 

214 

Bremer  

22 

314 

Buchanan  .               ... 

16 

263 

Buena  Vista  .                  .  . 

32 

314 

Butler  

27 

295 

Calhoun  

27 

273 

Carroll  .  .  . 

21 

279 

Cass    . 

30 

283 

Cedar  

21 

276 

Cerro  Gordo  

23 

265 

Cherokee  

43 

254 

Chickasaw 

23 

264 

Clarke  

32 

194 

Clay  

23 

285 

Clayton  

14 

255 

Clinton  ...            . 

21 

232 

Crawford 

19 

238 

Dallas  

24 

271 

Davis  

18 

224 

Decatur  

12 

203 

Delaware 

22 

23O 

(a)  Plot  these  data  on  an  8X11  sheet  of  graph  paper.  On  the  abscissa 
or  -X"-axis  place  automobile  registrations  anc^or^the  ordinate  or  Y- 
axis  plot  sales  per  capita.  Plot  the  point  (X,  "F)  from  the  results 
to  be  obtained  below.  

(6)  Calculate  the  means,  'X  and  T,  and  the  standard  deviations  for 
X  and  Y. 

(c)  Fit  a  straight  line  to  the  plotted  points  by  obtaining  the  regression 
of  Y  on  X  as  a  least  squares  fit.  What  is  the  model  in  this  case — 
i.e.,   for   a  single   observation,    county  per   capita  sales  by  filling 
stations?  What  parameters  do  the  statistics  feo  and  61  estimate? 
Explain  in  words  the  meaning  of  60  and  61,,  that  is,  interpret  the 
results  of  your  analysis. 

(d)  Plot  f*  =  bQ  +  biX  on  the  scattergram. 

(e)  Calculate  &  and  Y  —  &  f  or  each  X, 

(/)     Calculate  (F— ]P")2  for  each  X  and  thus  obtain  S(F— F)2-  Com- 


210 


CHAPTER    8,    REGRESSION    ANALYSIS 


pare  this  value  with  X)?/2""^i5D^2/-  Then  obtain  Sg  and  explain  its 

relationship  to  the  values  of  Y  —  ?. 
(0)    Determine  95  per  cent  confidence  limits  for  fti. 

8.14  On  the  basis  of  the  following  tabulations  comparing  years  of  service 
with  ratings,  the  management  seeks  to  discover  whether  or  not 
there  is  a  distinct  tendency  to  rate  old  employees  higher  than 
more  recent  additions  to  the  working  force. 


Employee* 

Service 
(in  years) 

Rating 

Employee 

Service 
(in  years) 

Rating 

A 

1 

5 

K 

6 

9 

B  

9 

6 

L           .     . 

7 

4 

C  

8 

8 

M  

1 

2 

D    .        ... 

3 

8 

N 

1 

3 

E  

3 

6 

O  

3 

8 

F    .     .    . 

2 

7 

P.  .    . 

1 

6 

G  

4 

5 

O  

2 

5 

H 

5 

6 

R  . 

2 

3 

I  

5 

4 

S.  .    , 

4 

4 

J  

6 

5 

T  

2 

7 

*  Source:  G.  R.  Da  vies,  Business  Statistics,  p.  338. 

(a)  Plot  a  scatter  diagram  ( X  =  service,  F  =  rating). 

(6)  Obtain  the  regression  line  J^  —  fto  +  friAT. 

(c)  Compute  s/^ 

(d)  Compute  s^. 

(tf)    Set  your  results  out  in  an  analysis  of  variance  table. 

(/)    Test  //:/3i  =  0  using  (1)  a  S-tcst  and  (2)  an  latest. 

(0)  Estimate  the  average  rating  which  might  be  given  an  employee 
with  (1)  4  years'  service,  (2)  15  years'  service.  Give  both  point 
and  interval  estimates.  Discuss  the  validity  of  these  estimates. 

(/&)  Kstimate,  hy  interval,  what  rating  an  individual  employee  with 
4  years'  service  might  be  expected  to  receive. 

(NOTE:  Whenever  necessary,  state  all  the  assumptions  made  in  order 
to  use  the  techniques  involved.) 

8,15  Assume  you  are  an  investment  counselor  for  a  large  insxirancc  com 
pany.  As  one  of  yoxir  duties,  you  woxild  need  to  have  some  idea  of  the 
amount  of  policy  loans  per  year,  i.e.,  loans  to  policyholders,  using 
their  life  insurance  policies  as  collateral.  Suppose  you  wish  to  estimate 
the  total  amount  of  policy  loans  your  company  will  make  during  the 
coming  year.  Assume  the  date  to  be  January  I,  1948.  You  are  given 
the  data  sheet  shown  below.  (I)  What  methods  of  estimation  might 
you  use  and  what  would  your  estimates  be?  (2)  What  further  informa 
tion  might  you  request  hi  order  to  do  a  better  job?  (3)  Give  reasons  for 
the  answers  you  make  to  (2). 


PROBLEMS 


21  1 


Year 

National  In 
come*  (in  mil 
lions  of  dollars) 

Estimated 
Population 
of  ILS.A.t 
(in  thousands) 

Policy  Loans 
Made  by  U.S. 
Life  Insurance 
CompaniesJ 
(in  millions 
of  dollars) 

1929.  .  . 

87,355 

121,770 

2,379 

1930 

75   003 

123,077 

2,807 

1.  ... 

58,873 

124,040 

3,369 

2  

41,690 

124,840 

3,806 

3. 

39,584 

125,579 

3,769 

4  

48,613 

126,374 

3,658 

5.  .  . 

56,789 

127,250 

3,540 

6  

64,719 

128,053 

3,411 

7 

73,627 

128,825 

3,399 

8  

67,375 

129,825 

3,389 

9.    . 

72,532 

130,880 

3,248 

1940  

81,347 

131,970 

3,091 

1 

103,834 

133,203 

2,919 

2.  . 

136,486 

134,665 

2,683 

3    . 

168,262 

136,497 

2,373 

4.  .  . 

182,407 

138,083 

2,134 

5  

181,731 

139,586 

1,962 

6.  .  . 

179,289 

141,235 

1,891 

7  

202,500 

144,034 

1,937 

8 

224,  400  § 

146,571 

*  Statistical  Abstract  of  the  United  States,  1949,  p.  281. 

t  Op.  cit.,  p.  7. 

j  Life  Insurance  Fact  Book,  1949,  p.  67. 

§  Estimated. 

8.16  Let  us  assume  that  one  of  your  duties  is  that  of  preparing  reports  for 
the  managing  director  of  the  firm.  They  are  engaged  in  manufacturing 
and  are,  of  course,  interested  in  the  average  cost  per  unit  of  production. 
Obviously,  units  of  production  are  easily  measured,  but  average  cost 
requires  lengthy  and  difficult  computations.  If  some  relationship 
between  these  two  quantities  can  be  determined  empirically,  an 
estimation  procedure  may  be  employed.  From  past  records,  the 
following  data  are  available: 


212  CHAPTER    8,    REGRESSION    ANALYSIS 


F  =  Average  Cost  X  =  Units  Produced 

(in  cents')  (in  thousands*) 


1.1  

9 

1.9  

13 

3.5  

5 

5.9  

17 

7.4  

18 

1.4  

8 

2.6  

14 

1.4  

12 

1.9  

7 

3.5  

15 

1.0  

10 

1.1  

11 

4.6  

4 

17.9  

23 

(a)    What  methods  would  you  employ  to  have  available  a  means  of 

estimating  values  of  Y  if  X  were  known? 
(6)    Describe  briefly  what  devices  you  would  use  to  determine  the  type 

of  curve  that  would  best  fit  a  given  set  of  data. 

8.17  An  advertising  concern  is  interested  in  prorating  sales  by  counties 
for  Maryland.  In  hopes  of  using  magazine  circulation  per  1000  popu 
lation  to  aid  them,  they  obtained  the  following  data: 


PROBLEMS 


213 


Magazine  Circulation  per 
1000  Population 


Per  Capita 
Sales  (F) 


159 279 

114 184 

67 137 

79 126 

112 213 

124 184 

129 181 

58 133 

85 161 

127 228 

64 129 

131 182 

75 142 

116 199 

141 268 

133 189 

76 161 

48 105 

68 102 

127 235 

150 259 

136 232 

114 216 


=  4245 
=  841 , 133  ]T .X" F  =  482  , 786 


(a)    Determine  the  regression  equation. 

(6)    If  the  circulation  in  County  A  were  90,  what  would  you  estimate 
the  per  capita  sales  to  be?  What  is  the  standard  error  of  f"? 

(c)  Is  the  regression  coefficient  significant? 

(d)  What  are  your  assumptions?  Are  they  justified? 

8.18  Given  the  following  data  satisfying  the  normality  and  homogeneous 
variance  assumptions,  do  you  believe  that  the  true  regression  is 
actually  linear? 


X 


4 
3 
6 

7 


18 
19 
18 
13 


26 

25 
24 
21 


38 
35 
28 
31 


44 
43 
39 
38 


214 


CHAPTER    S,    REGRESSION    ANALYSIS 


8.19  For  the  following  data,  test  the  hypothesis  that  /3i=/32  =  ^3  =  /34, 
where  we  assume  normality,  etc.,  as  required  for  simple  linear  re 
gression. 


Samples 

Degrees  of 
Freedom 

Z*2 

S  xy 

Z:v2 

A  

67 

300 

312 

550 

B 

75 

500 

515 

758 

C  . 

115 

200 

216 

375 

D  

34 

200 

300 

500 

8.20  Using  the  data  given  below,  fit  a  second  degree  polynomial  (parabola) 
for  gross  profits  per  farm  against  months  of  labor  to  obtain  a  curve 
for  Iowa  farms. 


Farm  No.* 

Gross 
Profits 
Y 

Months 
of  Labor 
X 

Farm  No.* 

Gross 
Profits 
Y 

Months 
of  Labor 
X 

1.  . 

16.7 

20 

18      

11.2 

14 

2.  .  . 

17.9 

19 

19  

9.4 

15 

3  

17.4 

24 

20  

8.7 

12 

4  

14.9 

15 

21  

12.2 

17 

5 

16.2 

24 

22 

7.7 

14 

6.  .  .     . 

14.0 

15 

23  

11.5 

14 

7  

15.1 

24 

24  

7.3 

13 

8 

18.3 

24 

25 

11.8 

16 

9    . 

11.3 

16 

26.  . 

15.1 

23 

10.  .  . 

18,3 

26 

27  

10.5 

33 

11  

17.1 

24 

28  

17.0 

29 

12 

12.0 

16 

29 

15.6 

30 

13 

15.2 

25 

30 

13.2 

31 

14,    .  .     . 

16.2 

28 

31  

17.2 

22 

15  

14.9 

24 

32  

14.6 

32 

16  

10.5 

15 

33  

12.2 

34 

17  

16.5 

27 

34  

9.8 

36 

*  Source:  Selected  values  from  Iowa  farm  records  plus  some  supplementary  hypothetical 
observations. 


8.21 


with 


=  244, 


8.22 


Given     the     linear     regression: 

Z^^"2  =  58,000,  and  n  =  100: 

(a)   What  is  the  standard  error  of  ^  =  254? 

(&)    For  what  ]?"  value  is  its  variance  a  minimum? 

(c)  Given  that  information  is  the  reciprocal  of  the  variance,  how  may 
we  maximize  our  information  about  the  unknown  parameter  /?i 
in  estimating  a  linear  regression  similar  to  the  above? 

In   a   regression   study  the   following   preliminary    calculations   were 

made:  3?  =  20;  T  =  22;  23  (X—  JT)2 


(a)   What  is  the  estimate  for  the  population  regression  coefficient? 


PROBLEMS          215 

(6)    How  do  you  interpret  the  population  regression  coefficient  /3i? 

(c)  Obtain  the  regression  equation  in  the  form  Y  =  b^  +  b^X. 

(d)  Test  the  hypothesis  H:@i  =  l. 

8.23  An  economist  from  the  University  of  Hawaii  and  an  economist  from 
the  University  of  Chicago  were  comparing  their  studies  of  income  and 
the  consumption  of  various  goods.  Among  the  items  studied  was 
gasoline  for  use  in  private  automobiles.  Each  had  used  a  sample  of 
about  100  university  employees  chosen  to  cover  the  range  of  salaries 
and  wages.  The  Chicago  economist  reported  that  he  had  observed  an 
increase  in  gas  consumption  of  10  gallons  per  $100  increase  in  income 
while  the  Hawaii  economist  noted  an  increase  of  only  4  gallons  per 
$100  increase  in  income.  They  then  looked  at  the  variances  of  their 
regression  coefficients  and  gave  these  figures,  V(bc)  =  2.41  and 


(a)  Could  the  observed  difference  between  the  regression  coefficients 
be  expected  to  occur  more  than  once  in  20  times  by  chance  if  we 
consider  the  necessary  assumptions  for  such  a  test  to  be  fulfilled? 

(6)  Would  your  conclusion  be  changed  if  the  change  in  gas  consump 
tion  had  been  reported  as  0.1  gallon  per  $1  increase  for  Chicago 
and  .04  gallon  per  $1  increase  for  Hawaii?  Or  what  would  the 
variance  of  bH  be,  if  the  regression  coefficients  had  been  reported 
in  the  latter  unit,  per  dollar  increase  in  income? 

(c)  When  the  Hawaii  economist  tested  the  hypothesis  (/S^  —  0),  he 
obtained  a  tf-  value  of  2.828=  (4  —  0)/-\/2.  Approximately  what  F- 
value  would  he  have  obtained  if  he  had  examined  the  reduction  in 
sum  of  squares  due  to  linear  regression  by  preparing  an  analysis 
of  variance? 

[NOTE:    V(§)   is  another  way  of  expressing  s|,  for  example,    t^(6c) 

=<j 

8.24      The  performance  of  a  tensile  strength  test  on  a  specific  metal  yielded 
the  following  results: 

Brinell  Hardness  Tensile  Strength 

Number  (1OOO  psi) 

X  Y 

104  38.9 

106  40.4 

106  39.9 

106  40.8 

102  33.7 

104  39.5 

102  33.0 

104  37.0 
102  33.2 

102  33.9 
101  29.9 

105  39.5 

106  40.6 

103  35.1 


216  CHAPTER    8,    REGRESSION    ANALYSIS 

(a)  Determine  the  best  linear  regression  equation  by  least  squares 
and  obtain  confidence  limits  for  estimating  the  mean  tensile 
strength  associated  with  a  specified  Brinell  number. 

(6)  Is  any  functional  form  other  than  a  linear  equation  indicated  by 
these  data?  Make  the  appropriate  test  and  discuss  your  results. 

8.25  Using  the  data  of  Table  8.10,   obtain  the  following  regression  equa 
tions  : 

(a)  Y  =  bQ  4-  biXi 
(Z>)  y  «  &0  4.  b2x* 
<V)  Y  =  b0  4-  £3^3 

(d)       Y    =    60   +    &4^4- 

For  each  regression  equation,  perform  a  complete  analysis.  Comment 
on  the  four  different  values  of  60.  Also,  compare  the  results  of  this 
problem  with  the  multiple  regression  analysis  obtained  in  Example  8.5. 

8.26  The  solubility  of  nitrous  oxide  in  nitrogen  dioxide  was  investigated 
with  the  following  results: 


Reciprocal  Temperature 

(=  10OO/  degrees  absolute) 

3.801         3.731 

3.662        3.593        3.533 

Solubility 

1.28           1.21 

1.11           0.81           0.65 

(per  cent  by  weight) 

1.33           1.27 

1.04          0.82          0.59 

1.52 

0.63 

Perform  a  complete  regression  analysis  and  interpret  your  results. 

8.27  A  Rockwell  hardness  test  is  fairly  simple  to  perform.  However,  the 
determination  of  abrasion  loss  is  difficult.   In  an  attempt  to  find  a 
way  of  predicting  abrasion  loss  from  a  measurement  of  hardness,  an 
experiment  was  run  and  data  collected  on  30  samples.  The  following 
results  were  obtained: 

JT  -  70.27,          7  =  175.4,          I>2  -  4300,          **T,y*  =  225,011, 
£>;y  =  —  22,946,          s%  —  3663,     and      f  =-  550.4  —  5.336X. 

Estimate  the  abrasion  loss  when  hardness  is  70.  Discuss  the  usefulness 
of  the  prediction  equation. 

8.28  A  gauge  is  to  be  calibrated  using  dead  weights.  If  X  represents  the 
standard  and  Y  the  gauge  reading,  perform  a  linear  regression  analysis 
based  on  the  following  results  from  10  observations: 

2*  =  230,          7  =  226,          2Z«y  =  1532,          £>2  «  1561,          Z^y2  =  1539. 
Test  H:  #L=1  using  «  =  0.01 


PROBLEMS  217 

8.29     Elongation  of  steel  plate  (F)  is  related  to  the  applied  force  in  psi 
Given  the  data 


X 


1.33 

26 

2.68 

51 

3.50 

66 

4.40 

84 

5.35 

101 

6.27 

117 

7.11 

133 

8.93 

150 

9.76 

182 

10.81 

202 

perform  a  complete  regression  analysis  and  interpret  your  results. 

8.30  It  is  desired  to  determine  the  relationship  of  a  twisting  movement  to 
the  amount  of  strain  imposed  on  a  piece  of  test  metal.  Eight  samples 
were  obtained  and  the  following  data  observed: 

Twisting  Movement  (X)  Strain  (F) 

100  112 

300  330 

500  546 

700  770 

900  1010 

1000  1100 

1200  1323 

1300  1515 

Determine  the  "best"  relationship  between  X  and  F.  Interpret  your 
results. 

8.31  The  data  given  below  and  identified  as  F,  Xi,  and  X*  represent  annual 
figures  for  1919  to  1943,  a  25-year  period,  for  three  adjacent  counties 
in  the  semiarid  central  area  of  South  Dakota.  F*  is  the  average  yield 
of  oats  in  the  ith  year.  XM  is  preseason  precipitation  in  inches,  e.g., 
9.82  for  JSTii  is  the  rainfall  from  August,  1918,  to  March  31,  1919,  etc. 
X%{  is  the  growing  season  precipitation  in  inches.  This  rainfall  covers 
the  period  April  1  to  July  31  for  each  crop  year  listed.  Due  to  the 
nature  of  weather  and  yield  data,  we  may  assume  that  these  data 
fulfil  our  necessary  assumptions  for  multiple  linear  regression.   Do 
a  complete  analysis  and  interpretation,  of  the  data.  The  reader  should 
note,  though,  that  these  are  time  series  data,  and  thus  an  ordinary 
multiple  linear  regression  analysis  may  be  of  doubtful  validity. 


21  8 


CHAPTER    8,    REGRESSION    ANALYSIS 


Year 

F 

^Ti 

X2 

1919  

30.8 

9.82 

14.85 

1920  

34.2 

9.12 

17.30 

1  .  . 

14.3 

6.24 

9.92 

2  

34.5 

14.06 

9.33 

3  

32.7 

5.29 

12,01 

4.  .  .          . 

36.0 

7.74 

10.87 

5  . 

33.8 

9.40 

11.78 

6  

3.7 

4.22 

7.14 

7  

26.1 

8.11 

14,44 

8  

18.6 

6.30 

8.95 

9.  .  . 

15.0 

10.58 

6.15 

1930  

23.8 

8.62 

8.63 

1  

4.4 

10.53 

6.19 

2  

23.5 

7.05 

8.86 

3  

0.1 

7.75 

7.97 

4  

0.0 

4.41 

4.93 

5  

19.7 

7.05 

11.27 

6  

0.0 

6.90 

5.37 

7  

4.5 

7.97 

8.78 

8  

14.4 

5.41 

10.37 

9  

13.4 

7.30 

8.78 

1940  

11.8 

5.94 

7.06 

1  

22.2 

6.77 

10.44 

2  

42.9 

11.23 

14.58 

3  

24.6 

8.55 

9,57 

8.32 


A  study  of  18  regions  gives  the  following  data  on  suicide  rate,  age, 
per  cent  male,  and  business  failures.  Fit  an  equation  for  the  linear 
regression  of  Y  on  JSTi,  X%,  and  X3,  where 

F    =  suicide  rate 
Xi  «  age 

X%  *»  per  cent  male 
X$  =  business  failures 

and  analyze  completely.  The  summary  of  the  data  follows: 

53  F      «         285.3  ^YXi    *     8536.6165 

—  531.09  53^-Ya    —  14500.1161 
=»         911.95  23FAr3    -  29644.847 

—  1800  53-^1^2  —  26913,822 
-<  4905.6904  2^Yi-Y3  »  53614.575 
=^  15731,2223  23^Xa  =  91U31 .630 
«  46218.4473 

»  199843.52 


8.33     Do  a  complete  multiple  linear  regression  analywin  of  the  following  data. 
Interpret  yoxir  results. 


PROBLEMS 


219 


T>  o  KHif- 

Choles 
terol 
Dosage 
(gm.  per 
day) 

Average 
Blood 
Total 
Choles 
terol 
(m#.) 

Initial 
Weight 

(&00 

Ratio  of 
Final 
Weight  to 
Initial 
Weight 

Average 
Food  In 
take  per 
kg.  Initial 
Weight 
(gm.  per 
day) 

Degree 
of  Athero 
sclerosis 

No. 

Xi 

X* 

X* 

X4 

X5 

Y 

1  

30 

424 

2.46 

0.90 

18 

2 

2  

30 

313 

2.39 

0.91 

10 

0 

3.  ... 

35 

243 

2.  75 

0.95 

30 

2 

4  

35 

365 

2.19 

0.95 

21 

2 

5  

43 

396 

2.67 

1.00 

39 

3 

6  

43 

356 

2.74 

0.79 

19 

2 

7  

44 

346 

2.55 

1,26 

56 

3 

8 

44 

156 

2   58 

0,95 

28 

0 

9.  .    . 

44 

278 

2.49 

1,  10 

42 

4 

10.  .    . 

44 

349 

2,52 

0.88 

21 

1 

11.  .    . 

44 

141 

2.36 

1.29 

56 

1 

12  

44 

245 

2.36 

0.97 

24 

1 

13  

45 

297 

2.56 

1.11 

45 

3 

14  

45 

310 

2.62 

0.94 

20 

2 

15  

45 

151 

3.39 

0.96 

35 

3 

16  

45 

370 

3.57 

0.88 

15 

4 

17 

45 

379 

1.98 

1.47 

64 

4 

18    . 

45 

463 

2.06 

1.05 

31 

3 

19    . 

45 

316 

2.45 

1.32 

60 

4 

20    . 

45 

280 

2.25 

1.08 

36 

4 

21  .  . 

44 

395 

2.15 

1.01 

27 

1 

22.  .. 

49 

139 

2.20 

1.36 

59 

0 

23  

49 

245 

2.05 

1.13 

37 

4 

24  

49 

373 

2.15 

0.88 

25 

1 

25  

51 

224 

2,15 

1.18 

54 

3 

26      . 

51 

677 

2.10 

1.16 

33 

4 

27.. 

51 

424 

2.10 

1.40 

59 

4 

28 

51 

150 

2.10 

1  .05 

30 

0 

8.34  You  arc  presented  with  farm  records  for  one  year  for  a  sample  of  89 
dairy  farms  located  in  a  fairly  homogeneous  area  in  the  same  milk  shed. 
The  records  contain  the  following  information: 

Y    —  milk  sold  per  cow  (Ibs.) 

Xi  =  amount  of  concentrates  fed  per  cow 

X%  =  silage  fed  per  cow 

X 3  =  pasture  cost  per  cow 

Xi,  —  amount  of  other  roughage  fed. 

You  first  decide  to  fit  a  multiple  linear  regression  of  F,  milk  sold,  on  the 
four  independent  variatcs,  the  X's  given  above.  Thus,  the  regression 
equation  is  of  the  form 

Y  «=  &0  +  biXi  +  bzX*  + 


220 


CHAPTER    8,    REGRESSION    ANALYSIS 


8.35 


(a)  List  the  numerical  quantities  and  statistics  you  would  compute  to 
obtain  this  regression  equation  for  Y.  You  need  not  give  detailed 
formulas.  In  particular,  you  will  wish  to  compare  b2  and  b4,  or 
silage  with  other  roughage  fed  in  effect  on  milk  production.  Also,, 
pasture  is  quite  homogeneous  in  the  area,  so  you  suspect  /33  may 
not  be  different  from  zero.  Include  in  your  list  such  items  as 
needed  for  examination  of  the  indicated  regression  coefficients. 
(6)  Supposing  you  obtain  61  =+0.30,  what  interpretation  would  you 

make  of  this  statistic? 
(c)    Can  you  suggest  any  other  form  for  this  regression  function,  using 

only  the  given  J^T's?  If  so,  write  it  out. 

Using  the  data  given  below,  obtain  a  multiple  linear  regression  equa 
tion.  (Do  a  complete  analysis.)  Then,  consider  other  possible  analyses 
and  comment  on  the  "best"  functional  relationship. 

DATA  FROM  25  IOWA  COUNTIES* 


Corn 

Yield 

Percent 

No. 

No. 

Percent 

Value 

Ob 
serva 
tion 

per 
Acre 
1910- 
1919 

age  Farm 
Land  in 
Small 
Grain 

Improved 
Acres 
per 
Farm 

Brood 
Sows  per 
1,000 
Acres 

age  Far  in 
Land 
in 
Corn 

per  Acre 
of  Land 
Jan.  1, 
1920 

Sum 

T>  um 
ber 

County 

Ar! 

X4 

A"a 

-Y4 

AT5 

F 

W 

I 

Allamakee 

40 

11 

103 

42 

14 

$  87 

297 

2 

Bremer 

36 

13 

102 

58 

30 

133 

372 

3 

Butler 

34 

19 

137 

53 

30 

174 

447 

4 

Calhoun 

41 

33 

160 

49 

39 

285 

607 

5 

Carroll 

39 

25 

157 

74 

33 

263 

591 

6 

Cherokee 

42 

23 

166 

85 

34 

274 

624 

7 

Dallas 

40 

22 

130 

52 

37 

235 

516 

8 

Davis 

31 

9 

119 

20 

20 

104 

303 

9 
10 

Fayette 
Fremont 

36 
34 

13 
17 

106 
137 

53 
59 

27 
40 

141 
208 

376 
495 

11 

Howard 

30 

18 

136 

40 

19 

115 

358 

12 

Ida 

40 

23 

185 

95 

31 

271 

645 

13 
14 
15 

Jefferson 
Johnson 
Kossuth 

37 
41 
38 

14 
13 
24 

98 
122 
173 

41 
80 
52 

25 
28 
31 

163 
193 
203 

378 
477 
521 

16 
17 

Lyon 
Madison 

38 
34 

31 
16 

182 
124 

71 
43 

35 
26 

279 
179 

636 
422 

18 

Marshall 

45 

19 

138 

60 

34 

244 

540 

19 

Monona 

34 

20 

148 

52 

30 

165 

449 

20 

Pocahontas 

40 

30 

164 

49 

38 

257 

578 

21 

Polk 

41 

22 

96 

39 

35 

252 

485 

22 
23 
24 

Story 
Wapello 
Warren 

42 
35 
33 

21 
16 
18 

132 
96 
118 

54 
41 
38 

41 
23 
24 

280 
167 
168 

570 
378 
399 

25 

Winnesluek 

36 

18 

113 

61 

21 

115 

364 

Sums 

937 

488 

3342 

1361 

745 

4955 

11828 

!Means 

37.48 

19.52 

133.68 

54.44 

29.80 

198  .  20 

473.12 

*  Reproduced  from;  H.  A.  Wallace  and  G.  W.  Snedecor,  Correlation  and  Machine  Calcu 
lation  (revised  ad,;  Ames,  Iowa;  The  Iowa  State  College  Press,  1931).  By  permission  of  the 
authors  and  publisher. 


REFERENCES    AND    FURTHER    READING  221 

References  and  Further  Reading 

1.  Anderson,    R.    L.,   and  Houseman,    E.   E.    Tables   of  orthogonal  polynomial 
values  extended  to  Ar=  104.  Res.  Bui.  297,  Agr.  Exp.  Sta.,  Iowa  State  Univ., 
April,  1942. 

2.  Bowker,   A.  H.,  and  Lieberman,   G.  J.   Engineering  Statistics.  Prentice-Hall, 
Inc.,  Englewood  Cliffs,  N.J.,  1959. 

3.  Brownlee,  K.  A.  Statistical  Theory  and  Methodology  in  Science  and  Engineer 
ing.  John  Wiley  and  Sons,  Inc.,  New  York,   1960. 

4.  Chew,  V.   (editor)  Experimental  Designs  in  Industry.  John  Wiley  and  Sons, 
Inc.,  New  York,  1958. 

5.  Davies,  O.  L.  (editor)  Statistical  Methods  in  Research  and  Production.  Oliver 
and  Boyd,  Edinburgh,  1949. 

6.  Dixon,  W.  J.,  and  Massey,  F.  J.  Introduction  to  Statistical  Analysis.  Second 
Ed.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1958. 

7.  Eisenhart,  C.  The  interpretation  of  certain  regression  methods  and  their  use 
in  biological  and  industrial  research.  Ann.  Math.   Stat.,  10    (No.   2):162— 86, 
1939. 

8.  Ezekiel,    M.   Methods   of  Correlation   Analysis.   John    Wiley  and   Sons,    Inc., 
New  York,   1941. 

9.  Gray  bill,   F.  A.  An  Introduction  to  Linear  Statistical  Models.   McGraw-Hill 
Book  Company,  Inc.,  New  York,  1961. 

10.  Hader,    R.    J.,    and    Grandage,    A.    H.    E.    Simple    and    multiple    regression 
analyses.  In  Experimental  Designs  in  Industry   (edited  by  V.   Chew).  John 
Wiley  and  Sons,  Inc.,  New  York,  1958. 

11.  Hunter,  J.  S.  Determination  of  optimum  operating  conditions  by  experimen 
tal   methods,    Part   II-l,    Models   and   Methods.    Industrial   Quality    Control, 
15  (No.  6):  16-24,  Dec.,  1958. 

12.  Kempthorne,   O.   The  Design  and  Analysis  of  Experiments.  John  Wiley  and 
Sons,  Inc.,  New  York,  1952. 

13.  Kendall,   M.   G.   The  Advanced  Theory  of  Statistics.  Vols.  I  and  II.   Charles 
Griffin  and  Co.,  Ltd.,  London,  1946. 

14.  Kramer,    C.   Y.  Simplified  computations  for  multiple  regression.   Industrial 
Quality  Control,  13  (No.  8):8-ll,  Feb.,  1957. 

15.  Prater,   N.   H.  Estimate  gasoline  yields  from  crudes.  Petroleum  Refiner,  35 
(No.  5):236-38,  May,  1956. 

16.  Snedecor,  G.  W.  Statistical  Methods,  Fifth  Ed.  The  Iowa  State  University 
Press,  Ames,  1956. 


C  H  APT  E  R    9 

CORRELATION  ANALYSIS 

IN  CHAPTER  8,  methods  of  estimating  functional  relationships  among 
variables  were  presented.  Such  methods  have  many  uses  in  experi 
mental  work.  However,  there  is  a  related  matter  which  also  deserves 
attention  when  discussing  the  joint  variation  of  two  or  more  variables* 
It  is:  How  closely  are  the  variables  associated?  Or,  in  other  words, 
what  is  the  degree  (or  intensity)  of  association  among  the  variables? 

9.1  MEASURES  OF  ASSOCIATION 

The  techniques  that  have  been  developed  to  provide  measures  of  the 
degree  of  association  between  variables  are  known  as  correlation  meth 
ods.  This  name  reflects  the  universal  practice  of  speaking  about  "meas 
ures  of  correlation' *  rather  than  about  "measures  of  the  degree  (or  in 
tensity)  of  association."  Consequently,  when  an  analysis  is  performed 
to  determine  the  amount  of  correlation,  it  is  referred  to  as  a  correla 
tion  analysis.  The  resulting  measure  of  correlation  is  usually  called  a 
correlation  coefficient. 

In,  this  chapter  some  of  the  more  frequently  used  measures  of  corre 
lation  will  be  presented.  However,  because  of  the  close  ties  between  this 
chapter  and  some  of  the  preceding  chapters  (particularly  Chapter  8), 
it  will  be  sufficient  to  give  only  a  minimum  of  discussion. 

9.2  AN    INTUITIVE  APPROACH   TO   CORRELATION 

Because  of  the  nature  of  the  concept  of  correlation,  it  is  clear  that 
(in  most  cases)  it  is  closely  related  to  the  concept  of  regression.  In  fact, 
for  a  given  regression  equation,  it  seems  reasonable  to  expect  that  a 
correlation  coefficient  will  measure  how  well  the  regression  equation 
fits  the  data  or,  stating  this  in  reverse  fashion,  how  closely  the  sample 
points  baig  the  regression  curve.  Thus,  a  correlation  coefficient  will 
undoubtedly  be  related  to  the  standard  error  of  estimate  ($#)  which 
measures  the  dispersion  of  the  points  about  the  regression  curve. 

Pursuing  this  idea  and  denoting  the  correlation  coefficient  by  the 
symbol  R,  we  express  R  as  a  function  of  s^}  for  example, 

*-/(*,).  (9.D 

If  R  is  to  perform  satisfactorily  as  a  measure  of  correlation,  it  is  desir 
able  that  it  exhibit  two  characteristics: 

(1)  It  should  be  large  when  the  degree  of  association  is  high  and 
small  when  the  degree  of  association  is  low. 

(2)  It  should  be  independent  of  the  units  in  which  the  variables 
are  measured. 

C2223 


9.4       CORRELATION    IN    SIMPLE    LINEAR    REGRESSION  223 

One   way  to  achieve  the  desired  properties  is  to    (approximately) 
define  R  by 

JV^l  -4/4  (9.2) 


where 

^E 


2     ^ST^    /  -rr  T>\  9  //„  ~\  /Q     '2\ 


4  =  13  (F  ~   F)V(»  -  1),  (9-4) 

and  g  is  the  number  of  parameters  in  the  true  regression  function  that 
were  estimated  by  the  regression  equation  symbolized  by  F.  If  n  is 
large  relative  to  q,  another  approximation  is 

R*G*  1  -  ]C  (F  -   f)V  ]C  (F  -   F)2.  (9.5) 


Since  53(F-  I^)2<  22  (F-  F)2,  it  is  clear  that  0  <B2<1.  Further,  if 
the  sample  points  hug  the  regression  curve  closely  (i.e.,  the  correlation 
is  high),  R2  will  be  close  to  1.  Similarly,  if  the  regression  curve  is  a  poor 
fit,  the  sample  points  "will  be  widely  dispersed  about  the  estimated 
regression  and  R2  will  be  close  to  0,  reflecting  a  low  correlation, 

Having  given  the  foregoing  intuitive  approach  to  correlation,  it  is 
necessary  that  a  more  precise  approach  be  formulated.  This  will  now 
be  done.  It  is  hoped  that  the  remarks  given  earlier  in  this  section  will 
aid  the  reader  in  appreciating  the  discussions  to  follow. 

9.3      THE   CORRELATION    INDEX 

Rewriting  Equation  (9.5)  as 


(F  -   F)2 
and  referring  to  Sections  8.8,  8.15,  and  8.16,  it  is  seen  that 

sum  of  squares  due  to  regression 
R*  =  -  -  -  ~  -- 
corrected  sum  of  squares 


(9.7) 


Since  the  ratio  defined  by  Equation  (9.7)  may  be  calculated  for  any 
estimated  regression  equation,  it  is  a  most  general  and  useful  measure 
of  correlation.  It  is  referred  to  as  the  correlation  index.  In  succeeding 
sections,  special  cases  will  be  examined  in  detail. 

9.4      CORRELATION    IN    SIMPLE    LINEAR    REGRESSION 

In  Section  8.8,  the  partitioning  of  the  sum  of  squares  of  the  depend- 
ent  variable  was  discussed  and  the  results  presented  in  Table  8.2. 
Referring  to  Table  8.2  and  invoking  Equation  (9.6),  we  obtain 

,  _   {  i:  Fg  -  CE 


224  CHAPTER   9,    CORRELATION    ANALYSIS 

(9'8) 


where  r2  is  used  instead  of  R2  to  conform  with  standard  practice.  It  is 
customary  to  talk  about  r  rather  than  r2.  Thus,  we  have 


=  (9,9) 


which  assumes  the  same  sign  as   ^2xy,  and  hence  the  same  sign  as  bi. 
It   is  readily  seen  that  the  correlation   coefficient  associated  with 
simple  linear  regression  is  easily  obtained  once  a  regression  analysis  has 
been  performed.  Further,  it  is  clear  that 

—  l<r<l  (9.10) 

where  —  1  represents  perfect  negative  linear  association  in  the  sample 
and  +1  represents  perfect  positive  linear  association  in  the  sample.  A 
value  of  0  is  interpreted  to  mean  that  no  linear  association  between  X 
and  Y  exists  in  the  sample.  Since  r  is  only  a  sample  value,  any  infer 
ence  to  the  sampled  population  must  be  carefully  stated.  More  will  be 
said  concerning  this  a  little  later. 

Example  9.1 

Referring  to  Tables  8.3  and  8.4,  the  coefficient  of  linear  correlation 
between  X  and  Y  for  the  Schopper-Riegler  data  is  determined  as 
follows: 

r*  **  10,177.59/(51,712.00  —  41,217.23)  •»  0.9698 
r    -  V0.9698  =  0.98. 
Example  9.2 

For  the  following  data, 


X 


-2  4 

•1  1 

0  0 

1  1 

2  4 


it  may  be  verified  that  r  =  0,  indicating  no  linear  association.  Please 
note  carefully  the  word  "linear/7  for  a  moment's  reflection  will  reveal 
that  X  and  Y  are  perfectly  associated,  the  relationship  being  Y**X*. 
What  we  calculated  was  a  measure  of  linear  correlation  when  the 
indicated  relationship  is  actually  quadratic.  This  simple  example  should 
call  to  your  attention  one  of  the  greatest  potential  trouble  spots  in 


9.5       SAMPLING    FROM    A    BIVARIATE    NORMAL    POPULATION  225 

correlation  analysis,   namely,   the  use  of  an  inappropriate  measure  of 
correlation. 

The  preceding  discussion  and  interpretation  of  r,  or  perhaps  we 
should  say  of  r2,  is  most  valuable  in  regression  analyses.  Examination 
of  Equations  (9.7)  and  (9.8)  reminds  us  that  lOOr2  is  the  percentage  of 
the  corrected  sum  of  squares  that  is  "explained  by7'  the  fitting  of  the 
simple  linear  regression  Y=bQ+biX.  If  this  percentage  is  not  large 
enough  to  satisfy  us,  a  better  fitting  regression  equation  should  be 

found.  . 

Some  terms  associated  with  the  coefficient  of  correlation  that  are 

sometimes  encountered  are: 

r2  =  coefficient  of  determination,  (9.11) 

1  —  r2  =  coefficient  of  nondetermination,  (9.12) 

and 

—  r2  ==  coefficient  of  alienation.  (9.13) 


Example  9.3 

For  the  Schopper-Riegler  data,  7*2  =  0.9698  and  1  -r2  =  0.0302.  Thus, 
96.98  per  cent  of  the  variation  in  Y  (Schopper-Riegler  rating)  is  "ex 
plained  by"  the  linear  regression  of  Y  on  X  (hours  of  beating). 

9.5     SAMPLING   FROM   A   BIVARIATE  NORMAL  POPULA 
TION 

The  interpretation  of  r  given  in  the  preceding  section  is  valid  for  any 
simple  linear  regression  regardless  of  what  assumptions  are  made  con 
cerning  the  variables  X  and  Y.  However,  if  a  random  sample  is  drawn 
from  a  bivariate  normal  population,  then  r  [defined  by  Equation  (9.9)  J 
is  a  sample  estimate  of  the  population  parameter 


p  « 


(9.14) 


The  reader  should  note  that  this  is  the  same  correlation  coefficient 
specified  by  Definition  3.31,  and  thus  it  is  not  surprising  that  r  is  some 
times  referred  to  as  the  sample  product-moment  correlation  coeffic^ent. 

When  sampling  from  a  bivariate  normal  population,  it  is  natural  to 
want  to  test  hypotheses  about  the  true  value  of  p.  Since  such  tests  are 
simply  further  examples  of  the  general  techniques  introduced  in  Chap 


ter  7,  only  a  brief  explanation  will  be  given. 

To  test  H:p  =  Q  versus  the  alternative  A:p?*Q,  we  calculate 


(9.15) 


226  CHAPTER    9,    CORRELATION    ANALYSIS 

and  reject  H  if  t>t a~.«/2)  cn— 2)  or  if  £<  —  Z(i_a/2)  0-2).  However,  a  mini 
mum  amount  of  simple  algebra  will  show  that 

t  =  —  =  —  >  (9.16) 

*r  *6l 

and  thus  the  test  just  detailed  is  identically  equivalent  to  the  test  of 
.ff:/3i  =  0  versus  A:/3i^O  as  given  in  Section  8.13.  A  review  of  that 
section  will  remind  you  that  the  hypothesis  might  also  be  tested  using 
an  J^-ratio  [see  Equation  (8.43)].  Consequently,  three  equivalent 
methods  of  testing  are  available,  the  choice  being  determined  by  the 
form  of  the  analysis. 

Example  9.4 

Given  the  sample  observations 


X 


11 


F 

it  is  easily  verified  that  r—  —0.98.  Using  Equation  (9.15),  we  obtain 
£=(  —  0.98)  V3/VO.  0413=  —8.49.  Since  t=  —8.49  <  —£.995(3)  =  —5.841, 
the  hypothesis  7/:p  —  0  is  rejected  in  favor  of  the  alternative  A  :p  ^0. 
Clearly,  a  1  per  cent  significance  level  was  used.  It  is  suggested  that 
the  reader  consider  //:/3i==0  versus  A:/3i  -p^O  and  compare  the  resulting 
test  statistic  with  that  computed  above. 


If  the  hypothesis  to  be  tested  is  //:p  =  p0  versus  A:p^p^y  where 
p0r^0,  the  test  procedure  is  more  complicated.  The  complication  arises 
because  (r—  p0)/5r  is  not  distributed  as  "Student's"  t  unless  p0  =  CK 
When  p0?^07  an  approximate  test  is  provided  by 

*r  -  G)[log.  (l  +  r)  -  log,  (1  -  r)] 

-   (1.1513)  [ioglo(l  +  r)  -  loglo(l  -  r)].  (9.17) 

Fisher  (4)  has  shown  that  zr  is  approximately  normally  distributed 
with  mean  #PO  and  variance  0%  —  l/(n  —  3).  The  approximate  test  pro 
cedure  is  to  calculate 

z  -   (*r-  *Pa)/cr,  (9.18) 

and  compare  this  quantity  with  fractiles  of  the  standard  normal  dis 
tribution.  The  hypothesis  //:p  =  p()  would  be  rejected  if 


or  if 

*<  -*(i-*,a>-  (9-20) 

The  research  worker  may  also  be  interested  in  obtaining  a  confidence 
interval  estimate  of  p.  This  may  be  obtained  by  calculating 


9.6       CORRELATION    IN    MULTIPLE    LINEAR    REGRESSION 


227 


and  then  using  Equation  (9.17)  to  solve  for  rL  and  ru, 

Quite  frequently  the  research  worker  has  several  independent 
samples,  each  randomly  selected  from  a  bivariate  normal  population, 
from  which  estimates  rx,  -  -  -  ,  rk  are  obtained.  If  the  research  worker 
can  accept  the  hypothesis  Hip^—  •  -  -  =  pA,  it  is  permissible  to  obtain 
a  pooled  estimate  of  the  common  population  correlation  coefficient, 
and  this  pooled  estimate  should,  of  course,  be  more  reliable  than  any 
of  the  individual  estimates.  If  calculations  are  carried  out  as  in  Table 
9.1  and  the  observed  chi-square  is  not  judged  significant  at  the  lOOa 
per  cent  significance  level,  a  pooled  estimate  of  p  (corresponding  to  the 
''average  z"}  may  be  found. 

TABLE  9. 1-Cal dilations  for  Testing  the  Hypothesis  p±  =   •  -  -   =:PA(^==3) 


Sample 

Size  of 

Sample 

n  —  3 

r 

z 

(v,  •a'Nrr 

\'U  ^^  O  J  At 

(n  —  3)s2 

A  .  .     . 

102 
102 
102 

99 
99 
99 

.  63245 
.  77459 
.  67082 

.74551 
1.03168 
.81223 

73  .  80549 
102.13632 
80.41077 

55.02273 
105,37200 
65.31204 

B.    . 

C  

Total 

306 

297 

256.35258 
.86314 

225.70677 

Jc                                      /         k 

Average  z  =    23  (^  —  3)#;  /     S  (w»  —  3) 

t—  1                                  '            i—  1 

(Averages)       23  (^  -~  3)zt- 

221.26817 

v2        for  testinc:  //:p,   —     •  •  •    ==  p,.  

4.43860 

9.6      CORRELATION    IN    MULTIPLE   LINEAR   REGRESSION 

When  a  multiple  linear  regression  equation  has  been  fitted  to  a  set  of 
data,  as  in  Section  8.15,  it  is  natural  to  seek  a  measure  of  correlation 
which  reflects  the  "goodness  of  the  fit."  The  correlation  index  defined 
in  Section  9.3  may  be  used  to  give  us  what  we  desire.  Referring  to  Equa 
tion  (8.68),  it  is  seen  that 

sum  of  squares  due  to  regression 


corrected  sum  of  squares 


(9.22) 


This  may  also  be  expressed  as 


x*>y 


Sy' 


(9.23) 


228  CHAPTER    9,    CORRELATION    ANALYSIS 

which  is  analogous  to  the  expression 


as  given,  in  Equation  (9.8).  If  we  calculate  R  =  \/R2,  where  R*  is  defined 
by  Equation  (9.22)  or  Equation  (9.23),  then  R  is  known  as  the  mul 
tiple  correlation  coefficient.  The  significance  of  R  may  be  assessed  by 
the  F-test  specified  in  Equation  (8.76).  No  example  will  be  given  at 
this  time  since  nothing  new  and  different  is  involved.  However,  some 
of  the  problems  at  the  end  of  the  chapter  will  require  the  calculation 
and  interpretation  of  the  coefficient  of  multiple  correlation. 

It  is  also  worth  noting  that  R3  as  defined  by  Equation  (9.22),  may 
be  thought  of  as  a  simple  linear  correlation  between  Y  and   Y  where 


Closely  allied  to  the  topic  of  multiple  correlation  is  that  of  partial 
correlation.  By  partial  correlation  is  meant  the  correlation  between  two 
variables  in  a  multivariable  problem  under  the  restriction  that  any 
common  association  with  the  remaining  variables  (or  some  of  them) 
has  been  "eliminated."  Clearly,  many  partial  correlation  coefficients 
may  be  calculated.  For  example,  a  first  order  partial  correlation  coeffi 
cient  is  one  which  measures  the  degree  of  linear  association  between 
two  variables  after  taking  into  account  their  common  association  with 
a  third  variable.  Symbolically, 

— 

(9  '  25) 


, 
Vl  — 


— 


* 


where  the  subscripts  refer  to  the  three  variables  JTi,  X%,  and  X$.  Here, 
of  course,  r^.z  is  attempting  to  measure  the  correlation  between  -XTi 
and  -XT  2  independent  of  JT3.  It  should  also  be  clear  that  r*v  (i,  j  =  1,  2,  3) 
are  simple  linear  correlation  coefficients  measuring  the  correlation  be 
tween  Xi  and  Xj.  A  second  order  partial  correlation  coefficient  may  be 
illustrated  by 

^12,3    —    ^14.3^24.3 

===  (9.26) 


.  a 

Vl  -  rl 

which  measures  the  correlation  between  .X\  and  X%  independent  of  X$ 
and  X&. 

Before  proceeding  to  another  topic,  it  will  be  worth  digressing  for  a 
moment  to  discuss  a  related  matter  (related  to  partial  correlation, 
that  is)  in  regression.  In  Section  8.15,  the  equation 

&  -  60  +  blXl  +  -  —  +  bkXk  (9,27) 


9.7      THE    CORRELATION    RATIO  229 

was  discussed  for  the  case  k  =  4.  At  that  time,  had  we  so  desired,  it 
would  have  been  appropriate  to  call  attention  to  a  different  system  of 
notation  which  is  sometimes  encountered.  For  fc  =  4,  Equation  (9.27) 
would  appear  as 

Y  =  bQ  +  biXi  +  b2X2  +  bzX*  +  £4X4.  (9.28) 

An  alternative  notation  is 

Y    ==5     bo    +    byi,  234^1    +    &F2.  134^2    +    #^3.  124^3    +    &r4.  123-^4,  (9.29) 


and  in  this  form  the  analogy  with  partial  correlation  is  evident.  Strictly 
speaking,  the  coefficients  should  be  called  partial  regression  coefficients 
where,  for  example,  &rx.234  represents  how  Y  would  vary  per  unit  change 
in  XT.  if  X%,  Xz,  and  X±  were  all  held  fixed.  Thus,  6  y  1.234  (or,  as  we  usually 
denote  it,  61)  gives  only  a  partial  picture  of  what  happens  to  Y  as  Xi 
changes.  Hence  the  adjective  "partial."  It  should  be  clear  that  the  less 
cumbersome  notation  was  used  (at  the  risk  of  not  clearly  defining  the 
meaning)  solely  to  simplify  the  writing  of  the  equations. 

9.7     THE  CORRELATION    RATIO 

Closely  related  to  the  correlation  index  is  a  quantity  known  as  the 
correlation  ratio.  Denoted  by  E*,  it  is  defined  by 


(9.30) 


where  T\-  is  the  mean  of  the  ith  group  consisting  of  n*  observations  and 
7  is  the  mean  of  all  observations.  Expressing  Equation  (9.30)  in  words, 

among  groups  sum  of  squares 

£2    ^    -  ?_^  -  £  -  ^  -  (9.31) 

corrected  sum  of  squares 

where  the  quantity  labeled  "among  groups  sum  of  squares"  is  most 
easily  found  using  the  identity 


(9.32) 

z—  1  ,  i«*l 


where 


(9.33) 
=  total  of  the  observations  in  the  ith  group 


230  CHAPTER    9,    CORRELATION    ANALYSIS 

and 

k 

T  =   ]T  a*  =  nY 

~1  (9.34) 

=  total  of  all  observations. 
It  should  be  clear,  of  course,  that 


(9.35) 
=  total  number  of  observations. 

A  moment's  reflection  will  indicate  that  the  value  of  E2  is  highly 
dependent  on  the  choice  of  groups.  For  example,  if  there  is  only  one 
observation  in  each  group,  the  value  of  ffi  is  unity;  if  all  the  observa 
tions  are  in  one  group,  the  value  of  E*  is  0.  Great  care,  then,  must  be 
exercised  when  grouping  the  observations. 

Another  point  of  interest  is  the  following  :  Once  the  observations  are 
assembled  in  groups,  the  value  of  E2  is  determined  solely  from  the 
values  of  the  "dependent"  variable.  Consequently,  the  "independent" 
variable  need  not  be  a  quantitative  variable.  It  can  be  a  qualitative 
variable.  That  is,  subject  to  the  dangers  implicit  in  the  groxiping,  the 
correlation  ratio  may  be  used  to  measure  the  correlation  between  a 
quantitative  variable  and  a  qualitative  variable. 

Since  grouping  is  so  important,  some  guidance  is  necessary.  One  rule 
of  thumb  is  to  have  three  to  five  groups,  each  containing  a  large  num 
ber  of  observations  (say  100).  Strict  rules  of  procedure  are  hard  to  de 
fine,  but  the  preceding  rule  may  prove  helpful.  Denoting  the  popula 
tion  correlation  ratio  by  ?72,  Woo  (13)  gives  tables  for  use  in  testing  the 
hypothesis  J/:?7  =  0  when  we  are  willing  to  assume  that  the  K»y  are 
normally  and  independently  distributed  (with  common  variance)  in 
each  group. 

Because  the  analysis  of  variance  form  of  presenting  results  is  so  often 
encountered,  it  should  not  be  surprising  to  find  it  helpful  in  the  present 
situation.  Referring  to  Table  9.2,  it  is  seen  that  the  sums  of  squares 
needed  in  Equation  (9.31)  arc  easily  accessible.  (NOTE:  Now  that  the 
opportunity  has  presented  itself,  we  shall  take  a  moment  to  review 
the  symbolism  introduced  in  Section  7.20.  It  seems  almost  unnecessary 
to  remark  that  the  letters  M,  (7,  and  W  in  the  symbols  Mvv,  Gyv,  and 
W^j,  were  chosen  to  stand  for  the  words  "Mean,  Groups,  and  Within," 
respectively.  However,  since  this  abbreviated  method  of  representing 
various  sums  of  squares  will  be  used  extensively  in  later  chapters,  it  is 
a  good  idea  to  become  well  acquainted  with  the  notation  as  early  as 
possible.) 


9.8       BISERIAL    CORRELATION 


231 


TABLE   9.2-Analysis  of  Variance  Associated  With  the 
Calculation  of  a  Correlation  Ratio 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of  Squares 

Mean  Square 

Mean    . 

1 

jj//_     =:  T^/M. 

Af«,/l 

Among  groups  .  . 
Within  groups  .  .  . 

k  —  1 

k 
/   „    \W>i   —    1) 

Gyy  =    23  G?M  -  T*/n 
Wvv  =    Z  ^2  -  Myy  -  Gvv 

<?«,/(*  -  1) 

/           * 

Total 

M 

y-  F2 

^—' 

9.8      BISERIAL   CORRELATION 

A  measure  of  correlation  encountered  frequently  in  such  areas  of 
specialization  as  education,  psychology,  and  public  health  is  the  bi- 
serial  correlation  coefficient.  Only  a  brief  discussion  will  be  given  in  this 
text.  Those  persons  interested  in  more  detail  are  referred  to  McNemar 
(7),  Pearson  (10),  and  Treloar  (12). 

The  biserial  correlation  coefficient,  usually  denoted  by  rb,  is  used 
where  one  variable,  Y,  is  quantitatively  measured  while  the  second 
variable,  X,  is  dichotomized,  that  is,  defined  by  two  groups.  The  as 
sumptions  necessary  for  a  meaningful  interpretation,  of  rb  are  : 

(1)  Y  is  normally  distributed  and  suffers  little  due  to  broad  group 
ing  (if  grouping  is  necessary) . 

(2)  The  true  distribution  underlying  the  dichotomized  variable  X 
should  be  of  normal  form. 

(3)  The  regression  of  Y  on  X  is  linear. 

(4)  The  mean  value  of  Y  in  the  minor,  or  smaller,  category  as 
specified  by  X,  denoted  by  Fi,  is  to  be  on  the  regression  line. 
This  assumption  implies  a  large  number  of  observations  in  the 
minor  segment. 

If  we  define: 

p  =  proportion  of  observations  in  the  major  category 
q  =  proportion  of  observations  in  the  minor  category 
z  =  ordinate  of  the  standard  normal  curve  at  the  point  cutting  off  a 

tail  of  that  distribution  with  area  equal  to  q 
T2  =  mean  of  the  Y  values  in  the  major  category 
SF  =  standard  deviation  of  all  the  Y  values, 
then 


(9.36) 


232  CHAPTER    9,    CORRELATION    ANALYSIS 

and  this  gives  us  a  measure  of  the  degree  of  linear  association  between 
X  and  F. 

It  should  be  mentioned  that,  in  a  manner  analogous  to  the  way  in 
which  we  developed  the  correlation  ratio,  Pearson  (10)  introduced  the 
concept  of  a  biserial  correlation  ratio,  denoted  by  Eb}  which  extends  the 
biserial  correlation  concept  to  cover  any  postulated  regression  function. 
We  shall  not  go  into  detail  here,  but  the  reader  is  referred  to  Pearson 
(10)  and  Treloar  (12)  if  he  is  interested  in  such  problems. 

9.9  TETRACHORIC  CORRELATION 

Another  measure  frequently  encountered  in  some  areas  of  research  is 
the  tetrachoric  correlation  coefficient.  This  is  generally  denoted  by  rt  and 
is  used  to  measure  the  degree  of  linear  association  between  two  vari 
ables,  X  and  Y,  where  both  are  dichotomized  and  the  true  underlying 
distributions  are  assumed  to  be  normal.  That  is,  if  we  have  samples 
from  a  bivariate  normal  population  but  the  measurements  are  not 
available  (we  know  only  to  which  cell  of  a  2X2  contingency  table  each 
observation  belongs),  we  can  obtain  a  measure  of  the  correlation  be 
tween  X  and  F.  It  is  not  feasible  to  present  a  formula  for  rt,  but  refer 
ence  to  McNemar  (7),  Treloar  (12),  and  other  works  will  indicate  cal- 
culational  methods  for  those  interested  in  this  particular  statistic. 

9.10  COEFFICIENT  OF  CONTINGENCY 

Of  some  interest,  also,  is  a  measure  of  the  degree  of  association  be 
tween  two  characteristics  where  our  observational  data  are  classified  in 
an  rXc  contingency  table.  In  Chapter  7  we  gave  a  method  for  testing 
the  hypothesis  that  these  two  characteristics,  or  classifications,  were 
independent  of  one  another.  Suppose,  however,  that  we  are  more  inter 
ested  in  estimating  the  degree  of  association  between  them  than  test 
ing  the  hypothesis  of  independence.  How  may  we  do  thin?  Pearson  (9) 
proposed  for  this  purpose  a  measure  known  as  the  coefficient  of  con 
tingency  defined  by 


C=     ^/—T-1'  C^-37) 

where  x2  is  the  usual 


as  given  in  Chapter  7,  In  the  case  of  a  2X2  table,  this  may  seem  to  be 
analogous  to  a  tetrachoric  correlation  coefficient,  but  the  coefficient  of 
contingency  is  of  wider  generality  because  wo  no  longer  require  the 
assumption,  of  normality  of  the  underlying  distributions.  Any  distri 
bution,  discrete  or  continuous,  is  acceptable.  However,  there  is  a  dis 
advantage  to  this  measxire  of  association;  its  maximum  possible  value 


9.11        RANK    CORRELATION  233 

varies  with  the  number  of  rows  and  columns,  and  thus  two  different 
values  of  C  are  not  directly  comparable  unless  computed  from  tables 
of  the  same  size.  For  further  remarks  on  this  measure,  the  reader  is  re 
ferred  to  McNemar  (7)  and  Treloar  (12). 

9.11       RANK   CORRELATION 

Let  us  now  consider  a  slightly  different  problem  but  one  that  arises 
quite  frequently  in  certain  areas  of  research.  The  problem  is  as  follows: 
n  individuals  are  ranked  from  1  to  n  according  to  some  specified  char 
acteristic  by  m  observers.,  and  we  "wish  to  know  if  the  m  rankings  are 
substantially  in  agreement  with  one  another.  How  may  we  answer  such 
a  query?  Kendall  and  Smith  (6)  have  proposed  a  measure  known  as  the 
coefficient  of  concordance,  W,  for  answering  this  question  which  is 
defined  by 


W  =  -  ,  (9.38) 

m?(w?  —  n) 

where  S  equals  the  sum  of  the  squares  of  the  deviations  of  the  total  of 
the  ranks  assigned  to  each  individual  from  m(n  +  l)/2.  The  quantity 
m(n  +  l)/2  is,  of  course,  the  average  value  of  the  totals  of  the  ranks, 
and  hence  3  is  the  usual  sum  of  squares  of  deviations  from  the  mean. 
W  varies  from  0  to  1,  0  representing  no  community  of  preference,  while 
unity  represents  perfect  agreement.  The  hypothesis  that  the  observers 
have  no  community  of  preference  may  be  tested  using  tables  given  in 
Kendall  (5)  or,  more  simply  (for  n>7),  by  calculating 

(9.39) 


<mn(n  -J-  1) 

which  is  approximately  distributed  as  chi-square  with  v  =  n—  1  degrees 
of  freedom.  If  there  are  "ties"  in  some  of  the  rankings,  it  may  be  neces 
sary  to  modify  our  formulas  somewhat;  if  such  a  case  is  encountered, 
the  researcher  is  referred  to  Kendall  (5). 

If  we  find  W  to  be  significant,  the  next  step  is  to  estimate  the  true 
ranking  of  the  n  individuals.  This  is  done  by  ranking  them  according  to 
the  sum  of  the  ranks  assigned  to  each,  the  one  with  the  smallest  sum 
being  ranked  first,  the  one  with  the  next  smallest  sum  being  ranked 
second,  and  so  on.  If  two  sums  are  equal,  we  rank  these  two  individuals 
by  the  sum  of  the  squares  of  the  ranks  assigned  to  them,  the  one  with 
the  smaller  sum  of  squares  obviously  being  ranked  ahead  of  the  other. 
If  W  is  not  significant,  we  are  not  justified  in  attempting  to  find  an 
"average/'  or  "pooled,"  estimate  of  a  true  ranking,  for  we  are  not  at 
all  certain  that  such  a  true  ranking  even  exists. 

When  m  =  2,  that  is,  when  only  two  rankings  are  available,  a  slightly 
different  approach  is  often  used.  In  this  case,  a  measure  known  as 
Spearman's  rank  correlation  coefficient  is  computed.  Spearman's  rank 
correlation  coefficient,  denoted  by  r89  is  defined  by 


234 


CHAPTER    9,    CORRELATION    ANALYSIS 


!_ 


n 


3    


n 


(9.40) 


where  c^  equals  the  difference  between  the  two  ranks  assigned  to  the  ith 
individual.  It  can  easily  be  seen  that  rs  varies  from  —  1  to  +1,  whereas 
W  varied  only  from  0  to  1,  —  1  signifying  perfect  disagreement  and 
+  1  signifying  perfect  agreement  between  the  two  rankings.  A  test  of 
the  null  hypothesis  H  :  ps  —  0  may  be  made  using  tables  provided  by 
Olds  (8).  We  must  remember,  however,  that  the  same  conclusion, 
namely,  to  accept  or  reject  H,  could  be  reached  by  computing  W  and 
comparing  with  the  tabulated  values  for  m  =  2.  Incidentally,  we  should 
remark  that  Kendall  (5)  does  not  tabulate  W  itself  but  only  the  associ 
ated  value  of  S.  This,  of  course,  cuts  down  the  amount  of  arithmetic 
required  since  it  is  not  necessary  actually  to  compute  the  value  of  W 
in  order  to  perform  our  statistical  test.  Similarly,  Olds  (8)  only  tabu 
lates 

±4. 

t—1 

Example  9.5 

Consider  the  data  of  Table  9.3.  Calculations  yield 


0.771     with     £)  d 

t—  1 


8. 


Using  01  =  0.05,  the  hypothesis  H:ps  =  Q  is  rejected.  [NOTE:  This  con 
clusion  was  reached  after  consulting  the  tables  provided  by  Olds  (8).] 
Thus,  it  is  concluded  that  the  two  judges  are  in.  quite  good  agreement. 


TABLE  9.3~Preferences  for  Six  Lemonades  as  Expressed  by  Two  Judges 


Lemonade 

Ranking  Given 
by  Judge  No.  1 

Ranking  Given 
by  Judge  No.  2 

Difference  in 
Ranks  =  d 

A    

4 

4 

0 

B  

1 

2 

—  1 

c  

6 

5 

1 

D  

5 

6 

—  1 

JS  

3 

1 

2 

F  

2 

3 

—  .  i 

Example  9.6 

Consider  the  data  of  Table  9.4.  It  may  be  verified  that  m(n+l)/2 
=  10.5,  S  ==25, 5,  and  W  =  0.162.  Examination  of  the  tables  in  Kendall 
(5)  leads  us  to  accept  the  hypothesis  of  no  community  of  preference 


9.12       INTRACLASS    CORRELATION  235 

TABLE  9.4— Preferences  for  Six  Lemonades  as  Expressed  by  Three  Judges 


Lemonade 

Ranking 
Given  by 
Judge  No.  1 

Ranking 
Given  by 
Judge  No.  2 

Ranking 
Given  by 
Judge  No.  3 

Sum  of  Ranks 

A  

5 

2 

4 

11 

B  

4 

3 

1 

8 

C  

1 

1 

6 

8 

D 

6 

5 

3 

14 

E  

3 

6 

2 

11 

F  

2 

4 

5 

11 

among  our  three  judges,  and  thus  we  shall  not  attempt  to  estimate 
any  "true  order  of  preference." 

9.12      INTRACLASS  CORRELATION 

a> 

The  measure  of  correlation  to  be  discussed  in  this  section  was  devised 
to  assess  the  degree  of  association  (or  similarity)  among  individuals 
within  classes  or  groups.  For  this  reason,  the  measure  is  known  as  the 
intraclass  correlation  coefficient.  (NOTE :  Some  authors  have  referred  to 
the  intraclass  correlation  coefficient  as  the  coefficient  of  homotypic  corre 
lation  but  the  former  term  is  more  common.) 

As  an  example  of  a  situation  in  which  the  intraclass  correlation  co 
efficient  is  the  proper  measure,  consider  the  problem  of  measuring  the 
correlation  between  heights  of  brothers.  Because  all  that  is  desired  is  a 
measure  of  similarity  between  heights  of  brothers,  any  attempt  to 
label  one  as  X  and  the  other  Y  (for  example,  by  age)  would  introduce 

TABLE  9.5-Symbolic  Representation  of  Data  To  Be  Used  in  Calculating 
the  Intraclass  Correlation  Coefficient 


Groups 

1 

2 

k 

Fu 
F12 

F21 
F22 

Ykl 
Ffc2 

Observations* 


Fln 

F2n 

Ykn 

Total 

Gf*                                                                                            S~** 
\                                                         ^-'"2                                                          ^fc 

*  Each  observation  is  assumed  to  be  of  the  form  F#  — ^u+ £i+€t-j  where  M  is  a  constant,  g^ 
is  a  random  variable  with  mean  0  and  variance  <r#,  and  eif  is  a  random  variable  with 
mean  0  and  variance  <r2.  That  is,  a  linear  model  has  been  postulated  which  states  that  any 
observation  is  a  linear  combination  of  three  contributing  factors:  an  over-all  mean  effect, 
an  effect  due  to  the  particular  group  to  which  the  observation  belongs,  and  an  "error" 
effect  representing  all  extraneous  sources  of  variation. 


236 


CHAPTER    9,    CORRELATION    ANALYSIS 


TABLE  9.6—  General  Analysis  of  Variance  for  Calculating  the  Intraclass 
Correlation  Coefficient  Using  the  Data  of  Table  9,5 


Source 
of 

Variation 

Degrees 
of 
Freedom 

Sum 
of 
Squares 

Mean 
Square 

Expected 
Mean 
Square 

Mean  

1 

M-uu 

Among  groups  

k—1 

Cr,rt/ 

s^-^-yiSQ 

(T2Jirncr^f 

Within  groups 

&O—  1) 

Ww 

$* 

<r2 

Total 

kn 

Y\  Y2 

a  spurious  element  into  the  correlation.  The  spurious  element,  of 
course,  would  be  that  an  ordinary  (simple  linear)  correlation  would 
measure  the  correlation  between  the  heights  of  older  brothers  and  the 
heights  of  younger  brothers  rather  than  simply  assess  the  "sameness" 
of  heights  of  brothers. 

The  intraclass  correlation  coefficient,  denoted  by  rI}  is  most  easily 
calculated  using  analysis  of  variance  techniques.  Given  the  data  of 
Table  9.5,  the  variation  among  the  kn  observations  may  be  summarized 
as  in  Table  9.6,  where 


T  - 


+  G2  + 


G/n  — 


(9.41) 
(9.42) 

(9.43) 


and 


R  -  Mvv  —  Guv.  (9.44) 

Since  the  population  intraclass  correlation  coefficient  is  defined  by 


'G 


(9.45) 


a  sample  estimate  is  provided  by 

2 


+  4 

MS  a  —  MSV 


where 


MS  a  +  (n  — 
MSa  =  mean  square  among  groups 


(9.46) 


9.12       INTRACLASS    CORRELATION  237 

=  s*  +  ns%  (9.47) 

=  Gvv/(]k  -  1) 
and 

MSW  =  mean  square  within  groups 

=  s*  (9.48) 

=  Wv3f/k(n  —  1). 

It  will  be  seen  that  if  n  =  2;  the  analysis  would  fit  the  situation  described 
earlier,  namely,  the  correlation  between  the  heights  of  brothers. 
(NOTE:  Once  again  we  have  availed  ourselves  of  the  opportunity  to 
introduce  some  new  notation.  This  time  the  concept  of  components  of 
variance,  denoted  by  s*  and  g2^,  has  been  used  as  an  alternative  way  of 
expressing  mean  squares.  The  relationship  between  "expected  mean 
squares'7  and  "mean  squares"  is,  of  course,  simply  the  familiar  rela 
tionship  between  "population  parameters"  and  "sample  statistics." 
The  determination  of  the  form  of  the  various  expected  mean  squares 
will  be  examined  in  detail  in  succeeding  chapters,  where  linear  models 
will  be  the  main  topic  of  discussion.  Those  who  desire  more  informa 
tion  on  this  topic  may  jump  ahead  to  the  appropriate  sections.) 

If  one  is  willing  to  assume  that  the  individuals  within  groups  are 
random  samples  from  normal  populations  (one  population  per  group) 
and  that  each  population  has  the  same  variance,  then  the  hypothesis 
H:pr  =  0  is  equivalent  to  the  hypothesis  .ff:  0-^  =  0,  and  this  may  be 
tested  using 

F  =  MS  a/ MS  w  (9.49) 

with  degrees  of  freedom  i>i  =  fc— 1  and  V2  =  k(n—  1). 

Example  9.7 

Given  the  data  in  Table  9.7,  calculations  will  lead  to  the  analysis  of 
variance  shown  in  Table  9.8.  From  this  we  obtain  r7  =  0.6974.  To  test 
/frpj^O,  we  calculate  F  =  30.857/5. 500  =  5. 61  with  z>i  =  7  and  i>2=*8 

TABLE  9.7-Heights  of  Eight  Pairs  of  Brothers 

rfeights 
Pair  (inches') 

A 71;  71 

B 69;  72 

C 59;  65 

D 65;  64 

E 66;  60 

P 73;  72 

G 68;  67 

H 70;  68 


238  CHAPTER    9,    CORRELATION    ANALYSIS 

TABLE  9.8-Analysis  of  Variance  for  Data  of  Table  9.7 


Expected 

Degrees  of 

Sum  of 

Mean 

Mean 

Source  of  Variation 

Freedom 

Squares 

Square 

Square 

Mean.    . 

1 

67.5 

67.5 

Among  groups  

7 

216.0 

30.857 

o-2+2<r5 

(Among  pairs  of 

brothers) 

Within  groups  

8 

44.0 

5.500 

<r2 

(Between  brothers 

within  pairs) 

Total 

16 

327.5 

degrees    of    freedom.     Since    F  =  5.61  >FQ.  95^8)  =  3.5,     the     hypothesis 
Hipz  =  Q  is  rejected. 

9.13      CORRELATIONS  OF  SUMS  AND    DIFFERENCES 

Reference  to  Section  5.14  reminds  us  that,  for  any  constants  a»  and 
any  variables  Xt,  the  linear  combination  specified  by 

U  =  22  a^Xi  (9.50) 

has 

vv  =  E[U]   =  jb  <*tf**  (9.51) 

*— i 

and 

where  /*,•  is  the  mean  of  .XT*-,  of  IKS  the  variance  of  X*,  arid  &*$  is  the 
co variance  of  X*  and  JSTj.  .Thus,  if  C7  — JSTi±JXr2, 

Mt/  =  MI  ±  ^  (9.53) 

and 

2  .-  _2  j_  -.2  4.  7<r  ^Q   ^4^1 

"  rr  V  i         [         C/  o      -.1—      **\J  -i  o  •  V,  ^  *  OTPy 

Utilizing  Definition  (3.31),  it  is  easily  verified  that  Equation  (9.54) 
may  be  rewritten  as 

^  —  ^2  ^  ^.2  ±  2p12<rIo-2.  (9,55) 

Rearranging  terms,  we  obtain 


PROBLEMS  239 

0-2     0.2    0-2 

P12  =  —-—^ ~  I  U  =  Xi  +  X*  (9.56) 

or 

2       l  2     _-2 

^  1  2 U    .  T-T-    -y  yr  XQ       c^\ 

P12    — ,  U    —    -A.1   —    ^-2-  ^y-->/>' 

This  leads  to  an  alternative  method  of  obtaining  r12  (the  sample  esti 
mate  of  P12);r  namely: 

TT    -\r      _i       V*  /^O      c:Q^ 

I  c/  ==  j\.  i  — }—  -A.  2  y^y .  ooj 


or 

o2    _|_     ?2  _    C2 

*t    +     *2  ^      .  j,    _    ^    _    ^  (Q     59) 


Before  terminating  our  discussion  of  the  correlation  of  sums  and  dif 
ferences,  attention  must  be  directed  to  the  relationship  between  the 
contents  of  this  section  and  the  "method  of  paired  observations" 
examined  in  Sections  6.9  and  7.9.  Noting  thatZ)  =  JX~  —  Y  is  analogous 
to  U  =  X±  —  X^  we  recognize  that  a  legitimate  pairing  of  related  ob 
servations  will  yield  a  smaller  standard  error  of  the  mean  difference  if 
a  positive  correlation  exists.  Such  a  reduction  in  the  standard  error 
represents  a  gain  in  efficiency  (relative  to  nonpairing)  which  will  be 
reflected  in  a  shorter  confidence  interval,  an  easier  establishment  of 
statistical  significance,  or  a  smaller  sample  size.  Clearly,  the  success  of 
pairing  in  any  situation  depends  upon  the  extent  to  which  the  re 
searcher  can  introduce  positive  correlation  into  an  experiment. 

Problems 

9.1  Using  the  data  of  Example  9.4,  test  H:/3i  =  0  versus  A  :/3i  ^0  using: 
(a)  a  Z-tcst,  (b)  an  ^-test.  In  both  tests,  let  <*  =  0.01. 

9.2  If  U  =  a  +  bX  and  V  =  c  +  dY,  show  that  ruv  —  rXY- 

9.3  Verify  Equation  (9.16). 

9.4  Interpret  a  simple  linear  correlation  coefficient  of  —0.8. 

9.5  If  the  simple  linear  (product-moment)  correlation  coefficient  between 
X  and  F  is  rjsrr  =  0.8,  what  are  the  values  of: 

(a)   rxv,  (6)   rxf  ,          and  (c)  ry$l 

9.6  Using  the  data  of  Problem  4.3  and  the  results  of  Problem  8.12,  com 
pute  and  interpret  the  appropriate  measure  of  correlation. 

9.7  Using  the  data  of  the  problem  indicated,  compute  and  interpret  the 
appropriate  measure  of  correlation: 

(a)  8.4  (e)  8.8  00    8.14  (m)8.20 

(6)  8.5  (/)  8.9  (J)    8.15  (n)   8.21 

(c)  8.6  (g)  8.11  (/e)    8.16  (o)    8.22 

(d)  8.7  (A)  8.13  (0    8.17  (p)  8.24 


24O 


CHAPTER    9,    CORRELATION    ANALYSIS 


9.8 


(?) 

(r) 
0) 

8.26 
8.27 
8.28 

(0    8.29                    O)  8.32 
O)  8.30                    <X>   8.33 
(v)    8.31                     (?)   8.35 

The  following 
F,  selected  at 

table  gives  hypothetical  data  for 
random  from  a  bivariate  normal 

the  covariates  X  and 
distribution. 

X 

F 

X 

Y 

12 

74 

18 

149 

20 

170 

16 

142 

17 

147 

13 

144 

11 

75 

18 

173 

8 

46 

11 

101 

8 

59 

16 

140 

4 

20 

15 

132 

12 

90 

5 

35 

9 

74 

14 

96 

12 

77 

6 

50 

16 

144 

3 

24 

11 

110 

5 

26 

10 

99 

8 

95 

13 

109 

6 

73 

15 

109 

17 

159 

(a) 
(6) 

(c) 


(d) 


Compute  the  means,  the  standard  deviations,  and  the  standard 

errors  of  the  means  of  X  and  F. 

Make  a  scatter  diagram  to  show  the  relation  between  these  two 

series.  Also,  draw  one  line  through  the  plotted  data  showing  the 

mean  of  X  and  another  showing  the  mean  of  F, 

Fit  a  straight  line  to  the  points  on  the  scatter  diagram  in  order  to 

express  mathematically  the  average  relationship  between  these 

two  variables.  The  required  equation  is  f^  —  bo  +  hiX.  This  calls  for 

the  computation  of: 

/   x*v 
(1)    the  regression  coefficient  bi  »  --  > 


—  3T).  Find 


(2)    the  F-intercept  60  —  7  — 

The  regression  equation  may  be  written 

6Q  and  61  geometrically  from  the  graph. 

Calculate  the  estimated  value  of  F  for  each  of  the  30  values  of  X 

from  the  equation  P"~&o  +  &i-Y.  Also,  compute  the  errors  of  esti 

mate  (F-F)  for  each  X. 

Interpret  the  constants  50  and  &i  obtained  for  !?"=*=  bo+hiX. 

Compute  and  interpret  the  standard  error  of  estimate  from  the 

formula 


/j\  (y  - 

-y—^ 


(00    Compute  the  sum  of  squares  of  the  errors  of  estimate  (deviations 
from  regression)  with  the  formula 


PROBLEMS  241 


(70    Test  the  regression  coefficient,  61,  for  significance. 

(i)     Compute  the  correlation  coefficient  using  the  formula 


(f)    Compute  and  interpret  the  coefficient  of  determination,  r2. 

(A?)    Partition  53  2/2  into  two  parts:  that  associated  with  regression,  and 

that  attributed  to  errors  of  estimate. 

(Z)     Compute  the  correlation  coefficient  between  X  and  F. 
(m)  Compute  the  correlation  coefficient  between  Y  and  Y. 
(ri)   Compute  the  correlation  coefficient  between  x  and  y. 
(o)    Compute  and  interpret  the  95  per  cent  confidence  limits  of  /3i. 
(p)   Compute  the  standard  errors  for  the  estimated  values  of   Y  for 

each  of  the  following: 

(1)  the  mean  of  all  F's  whose  X  value  is  equal  to  10. 

(2)  particular  Y's  whose  X  value  is  equal  to  10. 

(q)    Compute  the  sum  of  squares  attributed  to  regression  using  thefor- 

mula  52  (F—  F)2.  The  short-cut  formula  is  (2>2/W2>2>  or  r^y*. 

Show  computationally  that  the  three  formulas  give  the  same  sum 

of  squares. 

(r)    Show  computationally  that  (1  —  r^^^  =  2Z(Y—  F)2. 
(s)    Compute  the  regression  of  X  on  F;  that  is,  compute  the  constants 

in  the  equation  X  =  b'0  +  biY,  where 

and     ft'o  -  3T 

Plot  the  regression  on  the  same  sheet  on  which  the  regression 

^-^^Q+biX  was  plotted. 
(0     Show  that  r2  =  Z>i&I,  where  61  and  bi  are  the  two  regression  coeffi 

cients. 
(u*)   Compute 


(v)    Show  logically,  algebraically,  or  geometrically  that  |  r  \  cannot  be 

less  than  0  nor  greater  than  1. 
9.9        We  have  this  sample  of  X  and  Y  values: 


9 

4 

11 

2 

7 

5 

10 

1 

8 

3 

242 


CHAPTER    9,    CORRELATION    ANALYSIS 


9.10 


9,11 


9.12 


9.13 


(a)    Compute  the  product-moment  correlation  between   Y  and  X  for 

this  sample. 
(&)    What  assumptions  are  required  for  testing  the  significance  of  a 

sample  value  of  rl  What  parameter  is  estimated  by  the  sample 

correlation? 

Indicate  or  describe  three  methods  for  testing  the  hypothesis  that 

the  true  value  of  the  correlation  is  0  in  the  bivariate  population 

from  which  the  above  sample  was  taken.  (Exact  formulas  are  not 

required.) 

Management  seeks  to  discover  a  measure  of  correlation  between  length 
of  service  on  the  part  of  a  certain  type  of  machine  and  the  annual  re 
pair  bills  on  such  machines.  From  the  following  data: 


(c) 


Machine 

Years  of 
Service 

Annual  Repair 
Cost 

A  

1 

$2.00 

B 

3 

1  .50 

C  .,  . 

4 

2,50 

D  

2 

2.00 

&  

5 

3.00 

F.    . 

8 

4.00 

o. 

9 

4.00 

H  

10 

5.00 

/.  .  . 

13 

8.00 

J  

15 

8.00 

(a)   Make  a  scatter  diagram,   designating  years  of  service  as  the  X 
scries  and  annual  repair  costs  as  the  Y  series. 


(6) 
(c) 


Find  the  correlation  coefficient  r. 
Is  the  measure  of  correlation  significant? 
What  are  your  assumptions? 
Given  that 

22  yj  «=  1000 

and  that  the  Rum  of  squares  due  to  regression  is  640,  compute  the  value 
of  r  showing  all  your  steps.  What  assumptions  are  necessary  if  r  is  to  be 
interpreted  as  a  sample  estimate  of  a  population  correlation  coefficient? 
The  correlation  coefficient  between  the  C.A.V.D.  Vocabulary  and  the 
Graduate  Record  Verbal  tests  was  0.60  for  a  sample  of  67  men  students 
and  0.50  for  a  sample  of  39  women  students.  With  a  risk  of  Type  1  error 
of  5  per  cent,  is  this  evidence  that  the  two  groups  are  random  samples 
from  bivariate  normal  populations  of  the  same  correlation? 
Given  the  following  data  and  statistics  for  a  random  sample  from  a 
bivariate  normal  distribution; 


3? 

7 


6 

20 
22 


100 

2500 

-400 


-4 
—0.8 
44 


6.708 
0.6708 


(a)    Give  a  detailed  interpretation  of  the  linear  regression  of  Y  on  X. 


REFERENCES    AND    FURTHER    READING 


243 


Include  all  inferences  that  can  be  made  about  the  population 

regression.  Also,  interpret  all  inferences  made. 
(&)    Interpret  the  above  correlation  coefficient. 

(c)    What  assumptions  are  implicit  in  the  use  of  the  regression  in  (a)  ? 
9.14     Using  the  results  below,  test  the  hypothesis  H:pi=p(i  =  lj  •  •  -  ,  7). 
Also  run  through  the  series  of  tests  outlined  in  Section  8.26.  State  the 
assumptions  made  in  each  case. 

REGRESSION  AND  CORRELATION  DATA  IN  SEVEN  TYPES  or  SHEETING 


Fabrics 

Degrees 
of 
Free 
dom 

Z*' 

Sary 

S;y2 

Correla 
tion 
Coefficient 

Regression 
Coefficient 

Degrees 
of 

Free 
dom 

Sum  of 
Squares 

Mean 
Square 

1 

139 
139 
139 
139 
139 
139 
139 

60357.14 
60357.14 
60357.14 
60357.14 
60357.14 
60357.14 
60357.14 

—  989  .  64 
—  1970.43 
—  1647.50 
—  192.86 
—  5482.14 
—  7605.00 
—  12458.50 

1965  .  89 
2351.43 
3190,85 
3258.61 
2804.04 
2276.79 
4375.60 

—0.0909 
—0.1654 
—0.1186 
—0.0138 
—  0.4214 
—0.6487 
—0.7666 

—0.0164 
—0.0326 
—0.0273 
—0.0032 
-0.0908 
—0.1260 
—0.2064 

138 
138 
138 
138 
138 
138 
138 

1949.66 
2287.10 
3145.88 
3257.99 
2306.11 
1318.56 
1804.00 

2  

3  
4 

7  

Total 

973 

422499.98 

—30346.07 

20223.21 

-0.5028 

966 

16069,30 

16.63 

972 

20201.41 

Difference  for  testing  among  regression  coefficients 
^  =  688.68/16.63=341.41 

6 

4132.11 

688.68 

References  and  Further  Reading 

1.  Bowker,  A.  H.,  and  Lieberman,  G.  J.  Engineering  Statistics.  Prentice-Hall, 
Inc.,  Englewood  Cliffs,  N.J.,  1959. 

2.  Brownlee,   K.  A.  Statistical    Theory  and  Methodology  in  Science   and  Engi 
neering.  John  Wiley  and  Sons,  Inc.,  New  York,  1960. 

3.  Dixon,  W.  J.,  and  Massey,  F.  J.  Introduction  to  Statistical  Analysis.  Second 
Ed.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1957. 

4.  Fisher,  R.  A.  On  the  probable  error  of  a  coefficient  of  correlation  deduced 
from  a  small  sample.  Metron,  1  (No.  4)  :3,  1921. 

5.  Kendall,   M.   G.   Rank  Correlation  Methods.   Charles   Griffin  and   Co.,   Ltd., 
London,  1948. 

6    y  and  Smith,  B.  Babington.  The  problem  of  m  rankings.  Ann.  Math. 

Stat.,  10:275,   1939. 

7.  McNemar,  Q.  Psychological  Statistics.  John  Wiley  and  Sons,  Inc.,  New  York, 
1949. 

8.  Olds,  E.  G.  Distributions  of  sums  of  squares  of  rank  differences  for  small 
numbers  of  individuals.  Ann.  Math.  Stat.,  9:133,  1938. 

9.  Pearson,  K.  Mathematical  contributions  to  the  theory  of  evolution.  XIII. 
On  the  theory  of  contingency  and  its   relation  to   association   and   normal 
correlation.  Drapers'  Co.,  Res.   Me?n.,  Biometric   Series  I.  Cambridge   Uni 
versity  Press,  London,  1904. 

IQ    m  On  a  new  method  of  determining  correlation  when  one  variable  is 

given   by   alternative   and   the    other   by    multiple   categories.    Biometrika, 
7:248,  1910. 

11.  Snedecor,  G.  W.  Statistical  Methods.  Fifth  Ed.  Iowa  State  University  Press, 
Ames,  1956. 

12.  Treloar,  A.  E.  Correlation  Analysis.   Burgess  Publishing  Co.,   Minneapolis, 
1942.  . 

13.  Woo,   T.   L.   Tables  for  ascertaining  the  significance  or  nonsignincance  of 
association  measured  by  the  correlation  ratio.  Biometrika,  21:1,  1929. 


CHAPTER    10 

DESIGN  OF  EXPERIMENTAL 
INVESTIGATIONS 

BEFORE  PROCEEDING  to  the  introduction  and  discussion  of  further 
techniques  of  statistical  analysis,  time  will  be  taken  to  examine  certain 
aspects  of  data  acquisition.  Such  a  digression,  if  it  really  is  a  digres 
sion,  is  justified  because  the  analysis  of  any  set  of  data  is  dictated  (to  a 
large  extent)  by  the  manner  in  which  the  data  were  obtained.  The  truth  of 
the  foregoing  statement  will  be  illustrated  many  times  throughout  the 
remainder  of  this  book. 

10.1  SOME  GENERAL   REMARKS 

It  has  been  well  demonstrated  in  the  preceding  chapters  that  sta 
tistics  (as  a  science)  deals  with  the  development  and  application  of 
methods  and  techniques  for  the  collection,  tabulation,  analysis,  and 
interpretation  of  data  so  that  the  uncertainty  of  conclusions  based 
upon  the  data  may  be  evaluated  by  means  of  the  mathematics  of 
probability.  However,  it  should  also  be  evident  that  there  is  some 
thing  more  to  statistics  than  the  routine  analysis  of  data  using  stand 
ard  techniques,  For  example,  the  reader  should  realize  that  the  anal 
yses  are  exact  only  if  all  the  underlying  assumptions  are  satisfied.  Since 
this  is  rarely  true,  much  depends  on  the  skill  of  the  researcher  in  select 
ing  the  method  of  analysis  which  best  fits  the  circumstances  of  the 
experimental  situation  being  studied.  Thus,  it  seems  safe  to  say  that 
statistics  is  an  art  as  well  as  a  science. 

10.2  WHAT  IS  MEANT  BY  "THE  DESIGN  OF  AN  EXPERI 
MENT"? 

Designing  an  experiment  simply  means  planning  an  experiment  so 
that  information  will  be  collected  which  is  relevant  to  the  problem 
under  investigation.  All  too  often  data  are  collected  which  turn  out  to 
be  of  little  or  no  value  in  any  attempted  solution  of  the  problem.  The 
design  of  an  experiment  is,  then,  the  complete  sequence  of  steps  taken 
ahead  of  time  to  insure  that  the  appropriate  data  will  be  obtained  in  a 
way  which  permits  an  objective  analysis  leading  to  valid  inferences 
with  respect  to  the  stated  problem.  Such  a  definition  of  designing  an 
experiment  implies,  of  course,  that  the  person  formulating  the  design 
clearly  understands  the  objectives  of  the  proposed  investigation. 

10.3  THE  NEED   FOR  AN    EXPERIMENTAL   DESIGN 

That  some  sort  of  design  is  necessary  before  any  experiment  is  per 
formed  may  be  demonstrated  by  considering  an  example. 

[244] 


1O.4       THE    PURPOSE    OF    AN    EXPERIMENTAL    DESIGN  245 

Example  10.1 

It  is  desired  to  determine  the  effect  of  gasoline  and  oil  additives  on 
carbon  and  gum  formation  of  engines.1  Twenty  additives  are  to  be 
tested  in  combination  with  a  "control"  gasoline  and  oil  mixture.  Eighty 
similar  engines  are  available  for  use  in  the  experimental  program. 

As  the  problem  is  now  stated  it  is  far  too  general  to  permit  the  selec 
tion,  of  a  particular  design.  Many  questions  must  be  asked  (and  answers 
obtained)  before  the  statistician  can  propose  a  suitable  design.  Typical 
questions  are: 

(1)  How  is  the  effect  to  be  measured?  That  is,  what  are  the  char 
acteristics  to  be  analyzed? 

(2)  What  factors  influence  the  characteristics  to  be  analyzed? 

(3)  Which  of  these  factors  will  be  studied  in  this  investigation? 

(4)  How  many  times  should  the  basic  experiment  be  performed? 

(5)  What  should  be  the  form  of  the  analysis? 

(6)  How  large  an  effect  will  be  considered  important? 

When  we  recognize  that  the  foregoing  questions  are  only  a  small  sample 
of  those  that  might  be  asked,  it  is  evident  that  much  thought  should  be 
given  to  the  planning  stage  in  any  experimental  investigation.  In  fact, 
the  importance  of  thus  recommendation  cannot  be  overemphasized. 

10.4     THE   PURPOSE  OF  AN    EXPERIMENTAL   DESIGN 

The  purpose  of  any  experimental  design  is  to  provide  a  maximum 
amount  of  information  relevant  to  the  problem  under  investigation. 
However,  it  is  also  important  that  the  design,  or  plan,  or  test  program, 
be  kept  as  simple  as  possible.  Further,  the  investigation  should  be  con 
ducted  as  efficiently  as  possible.  That  is,  every  effort  should  be  made 
to  conserve  time,  money,  personnel,  and  experimental  material.  For 
tunately,  most  of  the  simple  statistical  designs  are  not  only  easy  to 
analyze  but  also  are  efficient  in  both,  the  economic  and  statistical 
senses.  For  this  reason,  a  statistician  should  be  consulted  in  the  early 
stages  of  any  proposed  research  project.  He  can  often  recommend  a 
simple  design  which  is  both  economical  and  efficient. 

Having  said  that  the  purpose  of  any  experimental  design  is  to  pro 
vide  a  maximum  amount  of  information  at  minimum  cost,  it  is  evident 
that  the  design  of  experiments  is  a  subject  which  involves  both  sta 
tistical  methodology  and  economic  analysis.  A  person  planning  an  ex 
periment  should  incorporate  both  of  these  features  into  his  design. 
That  is,  he  should  strive  for  statistical  efficiency  and  resource  economy. 
However,  an  examination  of  books  on  statistical  methods  and  the 
design  of  experiments  will  seldom  reveal  many  explicit  references  to 
the  cost  aspects  of  the  problem.  This  is  unfortunate.  On  the  other  hand, 
the  subject  of  cost  is  implicit  in  most  discussions  of  experimental  design. 
We  have  only  to  note  the  continual  attempts  to  plan  experiments  using 

1  Projects  and  Publications  of  the  National  Applied  Mathematics  Laboratories, 
April  through  June,  1949,  p.  79. 


246  CHAPTER    1O,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

the  smallest  size  sample  possible,  to  realize  that  the  cost  aspect  has  not 
been  overlooked.  Fortunately,  as  we  have  already  observed,  most 
simple  designs  are  both  economical  and  efficient,  and  thus  the  statis 
tician's  efforts  to  achieve  statistical  efficiency  usually  also  lead  to 
economy  of  experimentation. 

10.5  BASIC   PRINCIPLES  OF   EXPERIMENTAL   DESIGN 

It  has  been  stated  many  times  that  there  are  three  basic  principles 
of  experimental  design:  replication,  randomization,  and  local  control. 
Because  of  the  fundamental  nature  of  these  concepts,  each  will  be  dis 
cussed  separately.  Further,  it  is  recommended  that  the  reader  strive  for 
as  complete  an  understanding  and  appreciation  of  these  ideas  as  pos 
sible,  for  they  will  play  a  very  important  role  in  much  of  the  remainder 
of  this  book. 

10.6  REPLICATION 

By  replication  we  mean  the  repetition  of  the  basic  experiment.  The 
reasons  why  replication  is  desirable  are:  (1)  It  provides  an  estimate  of 
experimental  error  which  acts  as  a  "basic  unit  of  measurement"  for 
assessing  the  significance  of  observed  differences  or  for  determining  the 
length  of  a  confidence  interval.  (2)  Since,  under  certain,  assumptions, 
experimental  error  may  be  estimated  in  the  absence  of  replication,  it  is 
also  fair  to  state  that  replication  sometimes  provides  a  more  accurate 
estimate  of  experimental  error.  (3)  It  enables  us  to  obtain  a  more  pre 
cise  estimate  of  the  mean  effect  of  any  factor  since  cr-?  =  cr*/n.  (In  the 
formula  just  quoted,  o-2  represents  the  true  experimental  error  and  n 
the  number  of  replications.) 

It  must  be  emphasized  that  multiple  readings  do  not  necessarily 
represent  true  replication.  This  statement  may  best  be  substantiated 
by  an  example. 

Example  10.2 

Two  manufacturing  processes  are  used  to  produce  thermal  batteries" 
Sample  batteries  are  obtained  from  each  of  two  production  lots,  one 
lot  being  produced  by  process  A  and  the  other  by  process  B.  The 
batteries  are  then  tested  and  the  activated  life  of  each  battery  is 
recorded. 

If  an  analysis  of  the  above  experiment  were  attempted,  it  would  be 
discovered  that  no  valid  estimate  of  error  is  available  for  testing  the 
difference  between  processes.  The  variation  among  batteries  within 
lots  yields  a  valid  estimate  of  error  for  assessing  only  the  lot-to-lot 
variability.  True  replication  would  require  that  batteries  be  tested 
from  each  of  several  lots  manufactured  by  each  process.  (NOTE:  In 
the  example  just  given,  the  effects  of  lots  and  processes  are  said  to  be 
confounded.  This  term  will  be  discussed  more  fully  a  little  later.) 

Sometimes  the  absence  of  true  replication  is  more  easily  recognized 
than  in  Example  10.2.  For  instance,  if  multiple  measurements  of  acti 
vated  life  had  been  obtained  by  connecting  several  clocks  to  a  single 


10.7       EXPERIMENTAL    ERROR    AND    EXPERIMENTAL    UNITS  247 

battery,  the  researcher  would  easily  have  recognized  that  the  observed 
data  were  not  true  replications  but  only  repeated  measurements  on  the 
same  experimental  unit.  Another  example  of  the  same  type  of  spurious 
replication  (i.e.,  multiple  measurements  rather  than  true  replication) 
would  be  multiple  determinations  of  the  silicon  content  of  a  particular 
batch  of  pig  iron  where  the  variability  among  processes  was  to  be  as 
sessed. 

10.7   EXPERIMENTAL   ERROR   AND   EXPERIMENTAL 
UNITS 

In  the  preceding  discussion  of  replication,  the  terms  experimental 
error  and  experimental  unit  were  used.  Because  of  their  wide  usage,  it 
is  necessary  to  have  a  clear  understanding  of  their  meanings.  An  experi 
mental  unit  is  that  unit  to  which  a  single  treatment  (which  may  be  a 
combination  of  many  factors)  is  applied  in  one  replication  of  the  basic 
experiment.  The  term  experimental  error  describes  the  failure  of  two 
identically  treated  experimental  units  to  yield  identical  results. 

At  the  risk  of  saying  too  much  and  thus  confusing  the  reader,  it  is 
my  belief  that  some  discussion  of  the  preceding  definitions  is  in  order. 
In  one  respect,  the  term  "experimental  error"  is  unfortunate,  especially 
the  word  '  'error. "  This  word  is  probably  a  legacy  from  the  physical  sci 
ences,  particularly  astronomy,  where  the  investigators  (observers)  were 
concerned  with  errors  in  both  measurement  and  observation.  However, 
the  influence  of  experimenters  in  both  the  biological  and  physical  sci 
ences  should  not  be  discounted  entirely.  The  adoption  of  the  word 
"error"  could  just  as  easily  be  attributed  to  them,  for  they  clearly 
recognized  the  existence  of  errors  of  technique  in  the  performance  of 
their  experiments.  But  whatever  the  history  of  the  word  "error/ '  a 
thoughtful  examination  of  the  definition  of  the  terra  "experimental 
error"  will  reveal  that  its  meaning  to  the  statistician  is  much  more 
general.  In  each  particular  situation,  it  reflects:  (1)  errors  of  experi 
mentation,  (2)  errors  of  observation,  (3)  errors  of  measurement, 
(4)  the  variation  of  the  experimental  material  (i.e.,  among  experi 
mental  units),  and  (5)  the  combined  effects  of  all  extraneous  factors 
which  could  influence  the  characteristics  under  study  but  which  have 
not  been  singled  out  for  attention  in  the  current  investigation. 

There  is  another  item  related  to  the  term  experimental  error  which  is 
sometimes  confusing  to  the  statistical  novice.  This  is  the  practice  of  the 
professional  statistician  of  referring  to  "the  experimental  error  for 
testing  a  particular  effect."  Such  a  phrase  suggests  that,  in  a  given  ex 
periment,  there  may  be  more  than  one  experimental  error  even  though 
examination  of  the  assumed  statistical  model  will  reveal  only  one  such 
term.  As  confusing  as  this  practice  may  be  to  the  uninitiated,  it  serves 
a  useful  purpose.  As  the  reader  progresses  through  the  remainder  of  this 
book,  he  will  become  more  familiar  with  the  way  in  which  the  expres 
sion  is  used  and  thus,  I  hope,  become  more  tolerant  of  what  seems  at 
the  moment  to  be  an  unwise  use  of  words  that  have  been  carefully 


248  CHAPTER    TO,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

defined.  In  an  attempt  to  give  a  somewhat  more  specific  defence  at  this 
time,  let  me  say  that  all  the  statistician  is  really  doing  is  reminding  you 
of  the  fact  that  every  statistic  has  its  own  standard  error.  Perhaps  his 
choice  of  wrords  is  not  the  best,  but  it  is  a  firmly  entrenched  part  of  the 
language  of  experimental  design.  Thus,  I  strongly  recommend  that  you 
forgive  the  statistician  his  choice  of  words  and  that  you  concentrate  on 
the  more  important  task  of  learning  how  and  when  to  use  statistical 
methods. 

Before  terminating  this  discussion  of  experimental  error,  ways  of 
reducing  its  magnitude  should  be  indicated.  The  following  statements 
are,  of  course,  only  general  recommendations,  for  specific  recommenda 
tions  can  be  made  only  "when  a  particular  design  problem  is  being  con 
sidered.  Experimental  error  may  usually  be  reduced  by  adoption  of 
one  or  more  of  the  following  techniques:  (1)  using  more  homogeneous 
experimental  material  or  by  careful  stratification  of  available  material, 
(2)  utilizing  information  provided  by  related  variates,  (3)  using  more 
care  in  conducting  the  experiment,  (4)  using  a  more  efficient  experi 
mental  design. 

10.8      CONFOUNDING 

In  Section  10.6,,  the  word  "confounded"  was  introduced  to  describe 
a  certain  phenomenon  which  is  fairly  common  in  experimentation. 
Since  this  phenomenon  is  so  important  in  the  design  of  experiments,  it 
is  appropriate  that  time  be  taken  to  investigate  and  describe  it  more 
thoroughly.  This  will  best  be  done  through  the  use  of  examples. 

Example  10.3 

A  chemist  has  developed  a  new  synthetic  fertilizer  and  wishes  to 
compare  it  with  an  established  product.  He  contacts  a  nearby  university 
and  they  agree  to  run  an  experiment  on  two  available  experimental 
plots.  The  established  product  will  be  applied  to  one  plot  of  ground  and 
the  experimental  product  to  the  other.  The  characteristic  to  be  meas 
ured  and  used  as  the  index  of  performance  will  be  the  yield  (converted 
to  bushels  per  acre)  of  a  specified  cereal  crop.  However,  when  the  two 
yields  are  compared,  we  are  unable  to  say  how  much  of  the  difference 
is  due  to  fertilizers  and  how  much  is  due  to  inherent  differences  (in  fer 
tility,  soil  type,  etc.)  between  the  two  plots.  That  is,  any  comparison  of 
fertilizers  is  said  to  be  confounded  with  a  comparison  of  plots  or,  in 
slightly  different  words,  the  effects  of  fertilizers  and  plots  are  con 
founded. 

Example  10.4 

An  analyst  is  engaged  in  determining  the  percentage  of  iron  in  chemi 
cal  compounds.  Two  different  procedures  are  to  be  compared.  The 
analyst  takes  a  sample  of  the  first  chemical  compound  and  makes  a 
determination  of  the  iron  content  using  procedure  A.  Then  he  makes  a 
determination  using  procedure  J5.  This  sequence  (that  is,  first  A  and 
then  jB)  of  steps  is  repeated  several  times,  each  time  on  a  new  sample 
from  a  different  compound.  But  here  again,  as  in  Example  10.3,  we  are 


1O.9       RANDOMIZATION  249 

troubled  by  the  existence  of  confounding.  Any  comparison  of  the  two 
procedures  (A  and  B)  will  be  confounded  with  a  comparison  of  the  first 
and  second  determinations  made  (on  each  compound)  by  the  analyst. 
That  is,  if  there  is  any  improvement  in  technique  (due  to  a  learning 
process)  from  the  first  to  the  second  determination,  this  effect  will  be 
confounded  with  the  difference  between  procedures. 

Examination  of  the  preceding  examples  will  show  that  the  word  "con 
founded^  is  simply  a  synonym  for  "mixed  together."  That  is,  two  (or 
more)  effects  are  said  to  be  confounded  in  an  experiment  if  it  is  impos 
sible  to  separate  the  effects  when  the  subsequent  statistical  analysis  is 
performed. 

Since  one  of  the  purposes  of  experimental  design  is  to  provide  unam 
biguous  results,  it  would  seem  almost  obvious  that  a  good  design  should 
avoid  confounding.  It  is,  therefore,  disconcerting  to  the  uninitiated  to 
learn  that  the  statistician  frequently  deliberately  introduces  confound 
ing  into  a  design.  However,  as  you  will  see  later,  such  a  procedure  is 
not  followed  indiscriminately.  When  confounding  is  introduced  into  a 
design  it  is  done  so  for  a  good  reason,  and  the  reason,  is,  as  often  as  not, 
to  achieve  economy  through  reduction  of  the  size  of  the  experiment. 

10.9      RANDOMIZATION 

It  was  noted  in  Section  10.6  that  replication  provides  an  estimate  of 
experimental  error  which  can  be  used  for  assessing  the  significance  of 
observed  differences.  That  is,  replication  makes  a  test  of  significance 
possible.  But  what  makes  such  a  test  valid?  We  have  seen  that  every 
test  procedure  has  certain  underlying  assumptions  which  must  be  satis 
fied  if  the  test  is  to  be  valid.  Perhaps  the  most  frequently  invoked  as 
sumption  is  the  one  which  states  that  the  observations  (or  the  errors 
therein)  are  independently  distributed.  How  can  we  be  certain  that  this 
assumption  is  true?  We  cannot,  but  by  insisting  on  a  random  sample 
from  a  population  or  on  a  random  assignment  of  treatments  to  the  ex 
perimental  units,  we  can  proceed  as  though  the  assumption  is  true. 
That  is,  randomization  makes  the  test  valid  by  making  it  appropriate  to 
analyze  the  data  as  though  the  assumption  of  independent  errors  is 
true.  Note  that  we  have  not  said  randomization  guarantees  independ 
ence,  bxit  only  that  randomization  permits  us  to  proceed  as  though 
independence  is  a  fact.  The  reason  for  this  distinction  should  be  clear; 
Errors  associated  with  experimental  units  that  are  adjacent  in  space 
or  time  will  tend  to  be  correlated,  and  all  that  randomization  does  is  to 
assure  us  that  the  effect  of  this  correlation  on  any  comparison  among 
treatments  will  be  made  as  small  as  possible.  Some  degree  of  correlation 
will  still  remain,  for  no  amount  of  randomization  can  ever  eliminate  it 
entirely.  That  is,  in  any  experiment,  true  and  complete  independence  of 
errors  is  an  ideal  that  can  never  be  achieved.  However,  such  independ 
ence  should  be  sought,  and  randomization  is  the  best  technique  de 
vised  so  far  to  attain  the  desired  end. 

Sometimes  the  concept  of  randomization  is  introduced  as  a  device 
for  "eliminating"  bias.  To  illustrate  the  thinking  back  of  this  approach, 


250  CHAPTER    TO,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

consider  again  Example  10.4.  There,  any  comparison  of  procedures  A 
and  B  would  be  biased  in  favor  of  B  if  a  learning  effect  existed.  How 
ever,  if  each  time  a  new  compound  "was  to  be  investigated  the  analyst 
had  decided  at  random  which  procedure  to  use  first,  the  bias  would 
have  been  reduced,  perhaps  even  eliminated.  But  even  more  would 
have  been  accomplished.  If  there  were  other  biases  operating,  these 
would  also  have  had  their  effects  eliminated  (or  at  least  reduced)  by 
the  randomization.  That  is,  by  randomly  assigning  treatments  to  the 
experimental  units,  we  try  to  make  certain  that  treatments  will  not  be 
continually  favored  or  handicapped  by  extraneous  sources  of  variation 
over  which  the  experimenter  has  no  control  or  over  which  he  chooses 
not  to  exercise  control.  In  other  words,  randomization  is  like  insurance; 
it  is  always  a  good  idea,  and  sometimes  it  is  even  better  than  we  expect. 

Regardless  of  the  foregoing  arguments  in  favor  of  randomization, 
there  have  been  (in  the  past)  persons  who  have  spoken  out  in  favor  of 
systematic  (nonrandom)  designs.  "Can  we  not,"  they  ask,  "obtain  a 
more  accurate  measurement  of  differences  among  treatments  if  such 
treatments  are  applied  to  the  experimental  units  in  a  systematic  man 
ner?"  The  only  honest  answer  to  this  query  is,  "Possibly."  Why,  then, 
does  the  statistician  insist  on  randomization?  The  reason  is,  of  course, 
the  same  as  expressed  earlier:  It  is  because  the  statistician  wishes  to 
make  certain  inferences  from  the  observed  data  and  he  desires  to  at 
tach  a  measure  of  reliability  to  these  inferences.  If  randomization  is 
not  employed,  the  quoted  measure  of  reliability  may  be  biased.  Fur 
ther,  any  inference  would  be  unsupported  by  a  meaningful  probability 
statement.  (NOTE:  The  reader  is  reminded  of  the  discussion  of  judg 
ment  versus  random  samples  presented  in  Section  4.2.) 

There  are,  of  course,  situations  in  which  complete  randomization  is 
either  impossible  or  uneconomical.  The  statistician  should  not,  there 
fore,  adopt  the  unrelenting  position  of  insisting  on  complete  randomiza 
tion  in  every  case.  On  the  other  hand,  neither  should  he  agree  to  the 
use  of  a  completely  systematic  design,  for  the  experimenter  must 
reconcile  himself  to  the  fact  that  some  degree  of  randomization  is  re 
quired  for  the  valid  application  of  most  statistical  analyses.  Clearly, 
some  intermediate  position  between  the  two  extremes2  of  complete 
randomization  or  a  strictly  systematic  design  is  often  most  realistic. 
Once  the  experimenter  and  the  statistician  recognize  one  another's 
problems,  a  compromise  plan  can  usually  be  found  which  is  mutually 
satisfactory. 

10.10      LOCAL  CONTROL 

In  Section  10.5,  it  was  stated  that  the  three  basic  principles  of  ex 
perimental  design  are  replication,  randomization,  and  local  control. 

2  The  question  of  which  is  better,  a  systematic  or  a  randomized  design,  has 
never  been  completely  settled.  Most  likely  it  never  will  be  settled.  Most  designs 
in  common  use  today  involve  both  systematic  and  random  elements,  and  this 
seems  a  reasonable  state  of  affairs.  For  the  person  who  wishes  to  pursue  this 
point  farther,  the  literature  offers  many  papers  discussing  the  argument,  both 
pro  and  con.  See  references  (2,  27,  35,  36,  44). 


1O.11        BALANCING,    BLOCKING,    AND    GROUPING  251 

The  first  two  of  these  basic  principles  have  already  been  discussed  and 
it  is  now  appropriate  that  time  be  devoted  to  the  third. 

In  one  sense,  local  control  is  synonymous  with  experimental  design. 
However,  this  interpretation  of  experimental  design  is  very  narrow,  and 
not  consistent  with  our  earlier  definition.  If  we  agree,  then,  that  experi 
mental  design  is  as  defined  in  Section  10.2,  then  local  control  is  only  a 
part  of  the  total  complex.  In  this  sense,  local  control  refers  to  the 
amount  of  balancing,  blocking,  and  grouping  of  the  experimental  units 
that  is  employed  in  the  adopted  statistical  design.  It  was  observed 
earlier  (Section  10.9)  that  replication  and  randomization  make  a  valid 
test  of  significance  possible.  What,  then,  is  the  function  of  local  con 
trol?  The  function,  or  purpose,  of  local  control  is  to  make  the  experi 
mental  design  more  efficient.  That  is,  local  control  makes  any  test  of 
significance  more  sensitive  or,  in  the  language  of  Section  7.1,  it  makes 
the  test  procedure  more  powerful.  This  increase  in  efficiency  (or  sensi 
tivity  or  power)  results  because  a  proper  use  of  local  control  will  reduce 
the  magnitude  of  the  estimate  of  experimental  error.  (NOTE:  The 
reader  should  recognize  that  local  control  can  be  exerted  in  several 
ways.  The  more  common  methods  have  been  suggested  above  and  in 
the  last  paragraph  of  Section  10.7.) 

10.11       BALANCING,    BLOCKING,   AND   GROUPING 

In  the  preceding  section,  the  terms  balancing,  blocking,  and  grouping 
were  introduced  in  connection  with  the  principle  of  local  control. 
Rather  than  leave  these  words  undefined,  a  few  sentences  of  explana 
tion  will  be  given  so  that  the  researcher  will  understand  what  is  im 
plied.  Actually,  it  is  possible  to  say  that  the  three  terms  are  synony 
mous.  However,  in  this  text  we  shall  use  them  to  describe  different 
aspects  of  design  philosophy.  It  is  hoped  that  this  will  not  lead  to  con 
fusion  when  other  references  are  consulted. 

By  grouping  will  be  meant  the  placing  of  a  set  of  homogeneous  exper 
imental  units  into  groups  in  order  that  the  different  groups  may  be 
subjected  to  different  treatments.  These  groups  may,  of  course,  con 
sist  of  different  numbers  of  experimental  units. 

Example  10.5 

A  pharmaceutical  company  is  investigating  the  comparative  effects 
of  three  proposed  compounds.  The  experiment  will  consist  of  injecting 
rats  with  the  compounds  and  recording  the  pertinent  reaction.  A  litter 
consisting  of  11  rats  (experimental  units)  is  available.  Each  of  the  11 
rats  is  assigned  at  random  to  one  of  three  groups  subject  only  to  the 
restriction  that  the  three  groups  contain  4,  4,  and  3  rats,  respectively. 
The  animals  in  the  first  group  are  then  injected  with  compound  A, 
those  in  the  second  group  with  compound  B,  and  those  in  the  third 
group  with  compound  C. 

By  blocking  will  be  meant  the  allocation  of  the  experimental  units  to 
blocks  in  such  a  manner  that  the  units  within  a  block  are  relatively 
homogeneous  while  the  greater  part  of  the  predictable  variation  among 


252  CHAPTER    TO,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

units  has  been  confounded  with  the  effect  of  blocks.  That  is,  using  the 
researcher's  prior  knowledge  concerning  the  nature  of  the  experimental 
units,  the  statistician  can  design  the  experiment  in  such  a  way  that 
much  of  the  anticipated  variation  will  not  be  a  part  of  experimental 
error.  In  this  way,  a  more  efficient  design  is  provided. 

Example  10.6 

Consider  again  the  problem  outlined  in  Example  10.5.  This  time, 
however,  let  us  assume  that  12  rats  are  available  and  that  the  pedigrees 
show  6  of  them  are  from  litter  X,  3  are  from  litter  F,  and  3  from  litter  Z. 
Since  it  may  well  be  expected  that  rats  in  the  same  litter  will  perform 
more  nearly  alike  than  rats  from  different  litters  (due  to  inherited 
characteristics),  it  would  seem  natural  to  form  three  blocks.  The  first 
block  would  contain  the  6  rats  from  litter  X,  the  second  block  would 
contain  the  3  rats  from  litter  Y,  and  the  third  block  would  contain  the  3 
rats  from  litter  Z.  The  three  treatments  (A,  B7  and  C)  would  then  be 
assigned  at  random  to  the  rats  within  blocks.  Since  each  rat  is  subjected 
to  only  one  treatment,  the  block  containing  6  rats  would  undoubtedly 
end  up  with  2  rats  seeing  treatment  A,  2  seeing  treatment  B,  and  2 
seeing  treatment  C.  The  other  two  blocks  would  have  single  rats  seeing 
each  treatment. 

By  'balancing  will  be  meant  the  obtaining  of  the  experimental  units, 
the  grouping,  the  blocking,  and  the  assignment  of  the  treatments  to  the 
experimental  units  in  such  a  way  that  a  balanced  configuration  results. 
(Circular  though  the  preceding  definition  is,  I  feel  it  projects  the 
thought  I  wish  to  impart.  Consequently,  I  hope  you  will  forgive  the 
poor  logic.)  It  should  be  clear  that  we  can  have  little  or  no  balance, 
partial  balance,  approximate  balance,  or  complete  balance  in  any  par 
ticular  design.  For  instance,  Example  10.5  illustrates  a  case  of  approxi 
mate  balance,  while  Example  10.6  might  be  construed  as  an  illustra 
tion  of  partial  balancing.  Rather  than  go  on  to  manufacture  further 
examples  at  this  time,  let  us  defer  the  matter  until  later.  As  you  pro 
gress  through  the  chapters  on  various  designs  which  follow,  it  will 
become  abundantly  clear  that  the  statistician  continually  strives  for 
balanced  designs.  Thus,  examples  of  completely  balanced  designs  will 
be  available  in  excess. 

10.12     TREATMENTS  AND  TREATMENT  COMBINATIONS 

Several  times  in  the  preceding  sections,  the  word  "treatments"  has 
been  used  with  little  or  no  explanation.  Just  what  is  meant  by  this 
word?  Like  so  many  other  terms  in  statistics,  the  word  "treatments" 
entered  the  literature  because  of  its  use  in  agronomic  experimentation. 
However,  the  word  "treatments"  (like  "blocks"  and  "plots")  has  long 
since  lost  its  strict  agronomic  connotation.  In  fact,  the  three  phrases 
mentioned  in  the  preceding  sentence  are  now  an  accepted  part  of  the 
language  of  statistics,  regardless  of  the  area  of  application. 

To  the  statistician,  the  word  treatment  (or  treatment  combination) 
implies  the  particular  set  of  experimental  conditions  which  will  be  im- 


10.13       FACTORS,    FACTOR    LEVELS,    AND    FACTORIALS  253 

posed  on  an  experimental  unit  within  the  confines  of  the  chosen  design. 
By  way  of  explanation,  several  illustrations  will  now-  be  given : 

(1)  In  agronomic  experimentation,  a  treatment  might  refer  to: 

(a)  a  brand  of  fertilizer,  (b)  an  amount  of  fertilizer,  (c)  a  depth 
of  seeding,  or  (d)  a  combination  of  (b)   and   (c).  The  latter 
example  would  more  properly  be  termed  a  treatment  combi 
nation. 

(2)  In  animal  nutrition  experimentation,  a  treatment  might  refer 
to:  (a)  the  breed  of  sheep,  (b)  the  sex  of  the  animals,  (c)  the 
sire  of  the  experimental  animal,  or  (d)  the  particular  ration 
fed  to  an  animal. 

(3)  In  psychological  and  sociological  studies,  a  treatment  might 
refer  to:  (a)  age,  (b)  sex,  or  (c)  amount  of  education. 

(4)  In  an  investigation  of  the  effects  of  various  factors  on  the 
efficiency  of  washing  clothes  in  the  home,  the  treatments  were 
various  combinations  of:  (a)  the  type  of  water  (hard  or  soft), 

(b)  temperature  of  water,  (c)  length  of  wash  time,  (d)  type  of 
washing  machine,  and  (e)  kind  of  cleansing  agent. 

(5)  In  an  experiment  to  study  the  yield  of  a  certain  chemical  proc 
ess,  the  treatments  might  be  all  combinations  of:  (a)  the  tem 
perature   at   which   the   process   was   operated    and    (b)    the 
amount  of  catalyst  used. 

(6)  In  a  research  and  development  study  concerned  with  batteries, 
the  treatments    could   be   various    combinations    of:    (a)    the 
amount  of  electrolyte  and  (b)  the  temperature  at  which  the 
battery  was  activated. 

Many  more  examples  could  be  cited  from  every  field  in  which  experi 
mentation  is  performed.  However,  later  chapters  will  abound  with 
such  examples.  Thus,  it  seems  best  that  we  move  on  to  other  matters. 

10-13      FACTORS,    FACTOR    LEVELS,   AND   FACTORIALS 

In  any  discussion  of  experimental  design,  the  word  "factorial"  is 
almost  certain  to  be  heard.  Frequently,  the  reference  is  to  a  "factorial 
design."  However,  this  is  actually  a  misnomer.  There  is  no  such  thing 
as  a  factorial  design.  The  adjective  "factorial"  refers  to  a  special  way 
in  which  treatment  combinations  are  formed  and  not  to  any  basic  type 
of  design.  Thus,  if  a  randomized  complete  block  design3  has  been 
selected  and  the  treatment  combinations  are  of  a  factorial  nature,  a 
more  correct  expression  would  be  "a  randomized  complete  block  design 
involving  a  factorial  treatment  arrangement."  Some  writers,  such  as 
Yates  (46),  have  recognized  this  situation  and  they  speak  of  factorial 
experiments  rather  than  factorial  designs.  This  shift  in  terminology, 
while  in  the  proper  direction,  does  not  completely  resolve  the  difficulty 
since  the  word  "experiment"  seems  to  imply  that  survey  data  are  to  be 
excluded.  To  avoid  any  such  implication,  we  shall  speak  not  of  factorial 

*  See  Chapter  12  for  a  definition  of  this  type  of  design. 


254  CHAPTER    TO,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

designs  nor  of  factorial  experiments,  but  simply  of  factorials.  It  is  to  be 
understood,  of  course,  that  this  is  only  an  abbreviation  for  a  more 
lengthy  expression  describing  the  nature  of  the  treatments. 

Having  introduced  the  subject  of  factorials,  it  is  desirable  that  spe 
cific  terms  be  defined  in  an  explicit  manner.  This  will  now  be  done. 

In  most  investigations,  the  researcher  is  concerned  with  more  than  one 
independent  variable  and  in  the  changes  that  occur  in  the  dependent 
variable  as  one  or  more  of  the  independent  variables  are  permitted  to 
vary.  In  the  language  of  experimental  design,  an  independent  variable 
is  referred  to  as  a  factor.  Referring  to  the  illustrations  in  the  preceding 
section,  it  is  noted  that  five  factors  were  listed  for  the  home  washing 
study,  while  the  battery  study  involved  only  two  factors.  The  reader 
can  easily  find  many  more  examples  of  investigations  involving  several 
factors  by  consulting  various  technical  journals. 

Before  proceeding  to  the  definition  of  the  next  term  arising  in  con 
nection  with  factorials,  it  will  be  wise  to  indicate  the  generally  accepted 
notation  used  to  represent  factors.  Most  writers  use  lower  case  Latin 
letters  to  represent  factors.  As  an  illustration,  the  five  factors  in  the 
home  washing  experiment  might  be  represented  by 

OT  — =  type  of  washing  machine 
a  =  kind  of  cleansing  agent 
b  —  type  of  water 
c  =  temperature  of  water 
d  =  length  of  wash  time. 

A  second  illustration  is  provided  by  an  investigation  conducted  by 
Ratner  (40).  His  experiment  involved  a  study  of  how  long  it  took  to  per 
form  a  certain  move,  and  the  factors  investigated  were 

d  =  distance 
w  =  weight 
o  =  operator-pair 

It  was  mentioned  earlier  that  the  researcher  is  generally  interested 
in  experimental  results  (observations  on  the  dependent  variables)  as 
one  or  more  factors  are  allowed  to  vary.  It  will  be  seen  in  Ratner's  study 
that  he  considered  3  distances  (d^,  d2,  <33),  10  weights  (wi}  -  -  -  ,  1^10), 
and  4  operator-pairs  (ox,  o2,  o3,  o4).  In  the  home  washing  experiment, 
the  investigator  used  2  types  of  machine,  2  kinds  of  cleansing  agent,  2 
types  of  water,  2  temperatures  of  water,  and  2  lengths  of  wash  time. 
These  various  values,  or  classifications  of  the  factors,  are  known  as  the 
levels  of  the  factors.  That  is,  there  were  10  levels  of  weight,  3  levels  of 
distance,  and  4  levels  of  operator-pairs  in  Ratner's  experiment.  In  the 
home  washing  study,  each  factor  appeared  at  2  levels.  These  two  ex 
amples  should  indicate  that  the  word  "level"  is  a  very  general  term 
which  may  be  applied  in  many  varied  situations.  Ratner's  investiga- 


10.13       FACTORS,    FACTOR    LEVELS,    AND    FACTORIALS 


255 


tion  of  move  times  provides  an  excellent  example  of  this  diversity,  for 
tlie  3  levels  of  distance  (6,  12,  and  18  inches)  are  values  of  a  continuous 
variable,  while  the  4  levels  of  operator-pairs  (i.e.,  4  distinct  pairs  of 
operators  formed  from  8  individuals)  are  classifications  of  a  qualitative 
variable. 

Since  so  many  experiments  involve  factorial  treatment  arrange 
ments,  it  is  necessary  that  some  notation  be  adopted  to  represent  the 
various  treatment  combinations.  Unfortunately,  several  systems  of 
notation  appear  in  the  literature.  These  are  summarized  in  Table  10.1 

TABLE   10.1-niustrations  of  Notations  Used  To 
Represent  Factorial  Treatment  Combinations 


HF^f/ir*  4-rvt  m-*i  -f- 

Method 

j_  teatment 
Combination 

I 

II 

III 

IV 

V* 

1  

Q>\O\G\ 

111 

daboCQ 

000 

(1) 

2 

Qf\b\Ci 

112 

dob^c^ 

001 

c 

3  

a\b\cz 

113 

ao&o£2 

002 

c* 

4  

CLib^Ci 

121 

dobiCQ 

010 

b 

5  

CLlbzCz 

122 

G^biCi 

Oil 

be 

6  

ciibzCs 

123 

aobiCs, 

012 

be2 

7  

CL^b^Ci 

211 

ciibQCQ 

100 

a 

8  

dJ}\C2 

212 

CLlboCi 

101 

ac 

9  

aJb^Cz 

213 

dibaCz 

102 

ac2 

10                   

dob^Ci 

221 

a\b\CQ 

110 

ab 

11      . 

CLzbtCz 

222 

a^biCi 

111 

abc 

12 

CLzb^Cz 

223 

GlbiCz 

112 

abc* 

*  In  this  representation,  the  absence  of  a  letter  implies  that  the  factor  which  it  represents 
is  at  the  lowest  level.  In  general,  the  exponents  on  the  letters  agree  with  the  subscripts 
used  in  Method  III.  Thus,  ao&tffc  becomes  a°blc*  =  bc*.  The  symbol  (1)  is  used  to  signify  that 
each  factor  is  at  its  lowest  level,  that  is,  oo&o^o  is  equivalent  to  a°60c°=*  (1). 

for  a  case  involving  12  treatment  combinations  where  the  12  combina 
tions  were  formed  from  2  levels  of  factor  a,  2  levels  of  factor  6,  and  3 
levels  of  factor  c.  In  this  rep  resent  at  ion,  using  Method  I  as  an  ex 
ample,  the  symbol  a.-^-c*  (i=l,  2;  j=l,  2;  &=1,  2,  3)  represents  the 
treatment  combination  formed  by  using  the  iih  level  of  factor  a,  the 
jth  level  of  factor  b,  and  the  fcth  level  of  factor  c. 

There  is  another  item  of  terminology  that  should  be  mentioned  in 
the  present  context.  This  item  is  best  explained  by  example.  The  fac 
torial  arrangement  of  the  treatments  used  in  Table  10.1  would  be  re 
ferred  to  by  the  statistician  as  a  2X2X3  factorial.  Similarly,  Ratner's 
investigation,  would  be  termed  a  3X10X4  factorial,  while  the  home 
washing  study  was  a  2X2X2X2X2  =  25  factorial. 

Before  leaving  (for  the  time  being)  the  subject  of  factorials,  it  is  only 
fair  that  the  reader  be  warned  of  a  double  use  of  certain  symbols  which 
could  (but  should  not)  lead  to  confusion.  The  situation  is  as  follows: 


256  CHAPTER    TO,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

It  is  common  practice  to  use  the  letters  a,  6,  c,  •  •  •  to  denote  not  only 
the  various  factors  but  also  the  number  of  levels  of  the  factors.  For 
example,  a  statistical  model  might  be  written  as 

Y#  =  M  +  <*i  +  ftj  +  aj;         i  =  1,  -  -  -  ,  a  (10. 1) 

j  =  1,  •  •  •  ,  * 

where 

fj,  =  mean  effect 

<xt  =  effect  of  the  ith  level  of  factor  a 
$3  =  effect  of  thejth  level  of  factor  b 
eij  =  experimental  error 

and 

!>;=    Z&   =    0 
»=l  y=i 

while  the  6»/  are  NID  (0,  cr) .  In  this  and  similar  situations,  the  decision 
to  use  a  and  6  to  denote  not  only  the  factors  but  also  the  number  of 
levels  of  each  factor  should  not  lead  to  any  confusion.  The  sense  in 
which  a  letter  is  being  used  in  any  particular  instance  should  always 
be  perfectly  clear  from  the  context. 

10.14      EFFECTS  AND    INTERACTIONS 

Whenever  a  statistician  undertakes  the  design  of  an  experiment,  he 
must  first  ascertain  the  objectives  of  the  researcher.  Frequently,  the 
objectives  may  be  very  simple.  For  example,  the  researcher  may  wish 
to  determine  the  effect  on  the  yield  of  a  chemical  reaction  of  changing 
the  operating  temperature  while  all  other  factors  (variables)  are  held 
constant  at  predetermined  levels.  On  the  other  hand,  he  may  have  no 
interest  whatsoever  in  temperature;  his  concern  might  be  only  with 
pH.  In  this  case,  an  experiment  would  be  planned  to  determine  the 
effect  of  pH  under  the  restriction  that  all  other  factors  (including  tem 
perature)  are  held  constant. 

Experiments  such  as  those  referred  to  in  the  preceding  paragraph  are 
fine  if  the  effects  of  pH  and  temperature  (on  the  response  variable)  are 
independent.  However,  if  we  know  that  the  factors  are  interdependent, 
or  if  we  are  doubtful  of  the  validity  of  an  assumption  of  independence, 
then  an  experiment  which  estimates  both  main  effects  and  interactions 
should  be  recommended.  Such  an  experiment  would,  of  course,  utilize 
a  factorial  arrangement  of  the  treatments. 

Example  10.7 

It  is  suggested  that  the  effects  of  pTEL  and  temperature  on  the  yield  of 
a  certain  chemical  reaction  are  not  independent.  It  is,  therefore,  recom 
mended  that  a  design  be  adopted  which  utilizes  treatment  combinations 


10.14       EFFECTS    AND    INTERACTIONS  257 

formed  by  combining  different  levels  of  the  two  factors  involved.  It  is 
decided  that  two  levels  of  each  factor  will  be  investigated.  Denoting 
plS.  by  a  and  temperature  by  b,  the  four  treatment  combinations  might 
be: 


=  pEL  of  4.0  and  a  temperature  of  30°C. 
=  pTEL  of  4.0  and  a  temperature  of  40°C. 
=  pH  of  4.4  and  a  temperature  of  30°C» 
of  4.4  and  a  temperature  of  40°C. 


Before  we  can  say  how  the  performance  of  an  experiment  involving  a 
factorial  set  of  treatment  combinations  will  help  answer  our  questions 
concerning  independence  of  the  factors,  it  will  be  necessary  to  define 
certain  terms.  These  terms  (effect,  main  effect,  and  interaction)  have 
already  been  used  without  explanation.  The  time  has  now  arrived  when 
specific  definitions  must  be  given. 

We  shall  consider  first  a  22  factorial  such  as  the  one  used  in  Example 
10.7.  If  we  agree  that  the  symbols  a^bj  (i  =  Q,  l;y  =  0,  1)  can  represent 
not  only  the  treatment  combinations  but  also  the  average  yields  from 
all  experimental  units  subjected  to  the  similarly  designated  treatment 
combinations,  it  is  possible  to  define  effect,  main  effect,  and  interaction 
as  noted  below.  (NOTE  :  To  avoid  complicating  the  discussion,  it  has 
been  assumed  that  each  average  yield  was  obtained  from  the  same 
number  of  experimental  units.) 

Effect  of  a  at  level  b0  of  6  =  a-Lbo  —  a0&o  (10.2) 

Effect  of  a  at  level  &i  of  b  =  a^bi  —  a06i  (10.3) 

Main  effect  of  a  —  [(ai&o  —  #o&o)  +  (#i&i  —  a0&i)]/2 

=  G*i  -  ao)(Si  +  6o)/2  (10.4) 

=  A. 

Similarly, 

Effect  of  b  sit  level  aQ  of  a  =  ao&i  —  aQb0  (10.5) 

Effect  of  6  at  level  a±  of  a  =  aj>^  —  #160  (10.6) 

Main  effect  of  6  =   [(a05i  —  #0&o)  +  (^1^1  ~~  ^160)  3/2 

=  (ai  +  a0)(6i  -  6o)/2  (10.7) 

=  B. 

If  a  and  6  were  acting  independently,  the  effect  of  a  at  60  and  the  effect 
of  a  at  61  should  be  the  same.  (A  similar  statement  holds  for  the 
effects  of  b  at  ao  and  ai.)  Thus,  any  difference  in  these  two  effects  is  a 
measure  of  the  degree  of  interdependence  between  the  factors,  that  is, 
of  the  extent  to  which  a  and  b  interact.  Accordingly,  we  define  the 
interaction  between  a  and  b  by 


258 


CHAPTER    1O,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 
AB   = 


(10.8) 


If  the  symbols  used  in  the  preceding  definitions  are  simplified  by  re 
placing  ao  and  60  by  unity,  and  a\  and  61  by  a  and  6,  the  effects  and  in 
teractions  may  be  defined  by 

4M  =  (a  +  1)(6  +  1)  (10.9) 

2,4  =  (a  -  1)(5  +  1)  (10.10) 

2J5  =  (a  +  1)(6  -  1)  (10.11) 

2^LJ3  =  (a  —!)(&  —  1)  (10.12) 

where  M  represents  the  mean  effect  (i.e.,  the  mean  yield  of  all  experi 
mental  units). 

Example  10.8 

Let  us  assume  that  an  experiment  has  been  performed  involving 
treatments  such  as  described  in  Example  10.7.  To  illustrate  the  compu 
tation  of  main  effects  and  interactions;  three  hypothetical  cases  will  be 


examined. 


II 


III 


O-Q 


bo 

61 


63 

67 

69 

73 

61 


63 

67 

69 

78 

Z>0 
Jl 


63 

67 

69 

70 

Case       I:  71^  =  68,  A  =4,  B  =  6,  and  AB  =  0. 

Case     II:  M  =  69.25,  A  =  6.5,  B  =  8.5,  and 

Case  II I:  M  =  67.25,  A  =  2.5,  B  =  4.5,  and  A B 1.5. 

Having  defined  and  illustrated  (for  a  22  factorial)  the  concepts  of 
effects,  main  effects,  and  interactions,  it  is  appropriate  that  an  attempt 
be  made  to  put  these  ideas  into  words  rather  than  symbols.  However, 
the  reader  is  reminded  (again)  that  the  understanding  of  a  concept  is 
much  more  important  than  the  memorization  of  any  definition,  whether 
it  be  in  words  or  in  mathematical  symbolism.  With  that  reminder,  let 
us  now  attempt  definitions  of  the  two  terms,  "interaction"  and  "main 
effect."  Utilizing  the  earlier  definitions  and  the  illustrations  in  Example 
10.8,  we  may  say  that: 

(1)  Interaction  is  the  differential  response  to  one  factor  in  combination 
with  varying  levels  of  a  second  factor  applied  simultaneously.  That 
is,  interaction  is  an  additional  effect  due  to  the  combined  influence 
of  two  (or  more)  factors. 

(2)  The  main  effect  of  a  factor  is  a  measure  of  the  change  in  the  response 


7  O.I  4       EFFECTS    AND    INTERACTIONS  259 

variable  to  changes  in  the  level  of  the  factor  averaged  over  all  levels  of 
all  the  other  factors. 

It  should  be  clear  that  the  concepts  described  as  effects  and  interac 
tions  will  also  be  present  in  situations  involving  more  than  two  factors. 
For  example,  in  a  case  involving  four  factors,  there  would  be  four  main 
effects,  six  two-factor  interactions  involving  the  combined  effect  of  two 
factors  averaged  over  the  other  two  factors,  four  three-factor  inter 
actions  involving  the  combined  effect  of  three  factors  averaged  over 
the  one  remaining  factor,  and  one  four-factor  interaction  involving  the 
combined  effect  of  all  four  factors.  Extensive  discussion  of  these  ideas 
will  be  deferred  until  a  later  chapter. 

Before  terminating  the  discussion  of  effects  and  interactions,  how 
ever,  two  additional  topics  will  be  mentioned*  One  is  a  convenient 
method  of  determining  the  effects  in  2n  factorials;  the  other  is  the 
definition  of  effects  and  interactions  for  3n  factorials. 

To  illustrate  the  method  of  calculating  effects  in  2n  factorials,  let  us 
consider  a  23  factorial.  Using  the  abbreviated  notation  for  treatment 
combinations  given  in  Table  10.1,  and  letting  these  symbols  also  repre 
sent  the  average  yields  of  experimental  units  subjected  to  the  similarly 
designated  treatment  combinations,  the  main  effects  and  interactions 
may  be  found  by  adding  and  subtracting  yields  according  to  the  signs 
given  in  Table  10.2.  It  can  easily  be  verified  that  this  procedure  is 
simply  a  tabular  device  for  calculating  the  effects  and  interactions 
defined  by 

X  =   (a  ±   1)(6  ±  l)(c  ±  l)/22  (10.13) 

where  the  sign  in  each  set  of  parentheses  is  plus  if  the  corresponding 
capital  letter  is  not  contained  in  X  and  negative  if  it  is  contained  in  X, 
and  the  right-hand  side  is  to  be  expanded  and  the  yields  substituted  for 
the  appropriate  treatment  combination  symbols.  Equation  (10.13)  may 
be  extended  to  the  2n  factorial  case  by  simply  adding  more  multipli- 

TABLE    10.2-Schematic  Representation  of  Effects  and 
Interactions  in  a  23  Factorial 


Treatment  Combination 

Effect 

C\T 

(1)             a              b              ab              c              ac             be            abc 

Interaction 

+             +             +             +             +             +             +             + 

SM 

_|_             _|_             —             _|_             —             -|- 

4:A 

____^_|__-_-|-~|_ 

4B 

_|_             _(__(__             —             _]_ 

4AB 

__[__^_^-_|_ 

4C 

_|_             _)-             —              +             —             _|- 

4  AC 

^_1____              —             —             +             + 

4BC 

-             +             +             -             +             --             + 

4ABC 

260  CHAPTER    TO,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

cative  factors  as  shown  in  Equation  (10.14), 

X  =   [O  ±  1)(6  ±   l)(c  ±  1)0*  ±   1)  •  •  •  j/2"-1.          (1O.14) 


Wlien  factors  are  investigated  at  only  two  levels,  the  best  the  re 
searcher  can  do  (apart  from  a  simple  test  of  significance)  is  to  deter 
mine:  (1)  whether  the  effect  of  a  factor  is  positive  or  negative  and 
(2)  whether  the  factors  are  independent.  However,  when  factors  are 
investigated  at  more  than  two  levels,  the  researcher  can  probe  more 
deeply.  He  now  has  the  opportunity  to  see  if  the  effect  of  a  factor  is 
linear  or  nonlinear.  In  most  experimental  work,  this  is  a  very  impor 
tant  item  of  information,  and  thus  the  researcher  should  give  serious 
consideration  to  factorials  involving  more  than  two  levels  of  the  fac 
tors  when  planning  an  investigation. 

If  an  experiment  is  designed  involving  two  factors,  each  at  three 
levels,  the  main  effects  and  interactions  may  be  used  to  study  the  non- 
linearity  of  the  response  variable.  Rather  than  go  into  excessive  detail 
at  this  time,  only  the  pertinent  formulas  will  be  presented.  In  these  for 
mulas  we  have  again  used  the  symbols  a^bj  (i  =  0,  1,  2;  j  =  Q,  1,  2)  to 
represent  both  the  treatment  combinations  and  the  yields  from  the 
treatment  combinations. 

Linear  effect  of  a  =  AL  =  (a2  —  a0)(6o  +  61  +  £2)/3  (10.15) 

Quadratic  effect  of  a  =  AQ  =   (a2  —  2#i  +  #0)(£o  +  &i  •+-  t>z)/6  (10.16) 

Linear  effect  of  b  =  BL  =  (<z0  +  #1  +  #2)(&2  —  £o)/3  (10.17) 

Quadratic  effect  of  6  =  BQ  =  (a0  +  ai  +  a2)(Z>2  —  2iL  +  J0)/6  (1O.18) 

Linear  X  Linear  interaction  =  A^B^  =   (a2  —  #0)(&2  ~~  *o)/2  (10.19) 

Linear  X  Quadratic  interaction 

(10.20) 
=   ALBQ  =   (a*  —  a0)(£2  —  26i  +  60)/4 


Quadratic  X  Linear  interaction 

(10.21) 


Quadratic  X  Quadratic  interaction 

(10.22) 
=   AQBQ  =   (a*  —  2ai  +  a0)(62  —  26i  +  60)/8 


Example  10.9 

Consider  an  experiment  similar  to  that  described  in  Example  10.7 
but  involving  three  levels  of  pH  and  three  levels  of  temperature.  As  in 
Example  10.8,  three  cases  will  be  considered. 


I 

£7o  <Zl 


10.15       TREATMENT    COMPARISONS 

II  III 


10 

13 

16 

13 

16 

19 

16 

19 

22 

60 
61 
62 


Case       I:  ^£,  =  6, 

and  j4.QjBQ  =  0. 
,  ——       o,      -^IQ 

=  03  and 

r  T  T  .    j .  Q 

J.  J.  -L   .      -tTL  L  O  , 


22 

10 

14 

25 

13 

17 

30 

18 

22 

10 

12 

11 

14 

17 

21 

19 

25 

35 

J  and 


—  1/8. 


It  should  IOQ  noted  that  the  "no  interaction"  result  in  cases  I  and  II 
could  have  been  predicted  by  observing  that  the  pattern  of  differences 
between  yields  at  varying  levels  of  b  is  the  same  for  each  level  of  a. 
(NOTE:  We  could  just  as  easily  have  examined  the  differences  between 
yields  at  varying  levels  of  a  for  each  level  of  &). 

From  the  preceding  discussion,  it  should  be  evident  that  there  is  a 
great  deal  to  be  said  about  effects  and  interactions.  As  a  matter  of  fact, 
what  started  out  to  be  a  short  section  exposing  the  reader  to  general 
concepts  has  grown  (necessarily,  I  believe)  into  a  rather  detailed  dis 
cussion  of  the  topic.  On  tlie  other  hand,  the  surface  has  only  been 
scratched.  There  is  much  more  that  can  be  said.  Some  of  this  additional 
material  will  be  discussed  in  later  chapters,  while  the  remainder  will  be 
left  to  books  devoted  to  experimental  design.  For  those  who  wish,  to 
read  further  on  these  topics,  the  following  references  are  recommend 
ed:  Cochran  and  Cox  (13),  Cox  (14),  Davies  (16),  Federer  (20),  Finney 
(21  and  22),  Kempt  home  (28),  Quenouille  (39),  and  Yates  (46). 

10.15     TREATMENT  COMPARISONS 

In  most  experiments  involving  several  treatments,  the  researcher  will 
be  interested  in  certain  specific  comparisons  among  the  treatment 
means.  To  aid  in  making  such,  comparisons,  the  statistician  finds  it 
convenient  to  talk  in  terms  of  "contrasts,"  Algebraically,  a  contrast 
among  the  quantities  TI,  -  -  -  ,  T&  (where  2\-  is  the  sum  of  nt  observa 
tions)  is  defined  by 


ck5Tk 


(10.23) 


where 


(10.24) 


If  each  ni  =  n,  that  is,  if  each  Tt  is  the  sum  of  the  same  number  of  ob- 


262  CHAPTER    TO,    DESIGN    OF   EXPERIMENTAL    INVESTIGATIONS 

servations,  then  the  necessary  condition  for  a  contrast  reduces  to 

,    =  0.  (10.25) 


Example  10.10 

Consider  an  experiment  involving  batteries  in  which  four  treatments 
are  to  be  investigated.  The  four  treatments  happen  to  be  four  different 
electrolytes.  However,  it  is  noted  that  electrolytes  No.  1  and  No.  2  are 
quite  similar  in  composition,  that  No.  3  and  No.  4  are  also  similar,  but 
that  Nos.  1  and  2  differ  considerably  from  Nos.  3  and  4.  It  would,  then, 
be  reasonable  to  plan  comparisons  of:  (1)  treatments  1  and  2  versus  treat 
ments  3  and  4,  (2)  treatment  1  versus  treatment  2,  and  (3)  treatment 
3  versus  treatment  4.  Assuming  that  20  batteries  (experimental  units) 
are  used  and  that  they  are  allocated  to  the  treatments  in  the  ratio 
4:2:5:9,  what  would  be  the  form  of  the  contrasts  for  the  selected  treat 
ment  comparisons?  Denoting  the  treatment  totals  by  T^i  =  ly  2,  37  4), 
the  desired  contrasts  are: 


=  7rz  +  7T2  -  3T3  —  3T* 
C2  -  (l)Ti  +  (-2)2-2  +  (0)2*8 
C3  -  (0)2-1  +  (0)2*a  +  (9)T,  +  (-5)  2V 

One  might  ask  how  we  obtained  the  coefficients  c*j  (i  =  1,  2,  3,  4;  j"  =1,2, 
3)  used  in  the  above  comparisons.  A  short  explanation  at  this  moment 
should  serve  to  clear  up  any  difficulties.  Consider  the  case  of  compari 
son  Ci:  What  we  are  actually  attempting  to  do  is  to  compare  the  mean 
of  6  observations  (4  +  2)  with  the  mean  of  14  observations  (5+9).  It  is, 
of  course,  necessary  to  adjust  for  the  spurious  weighting  given  by  our 
comparison  of  treatment  totals  based  on  unequal  numbers  of  ob 
servations.  Since  the  smallest  integer  which  may  be  divided  evenly  by 
both  6  and  14  is  42,  we  see  that  7  and  3  are  the  indicated  weights  to 
be  used  if  our  comparison  is  to  be  unaffected  by  the  differing  numbers 
of  observations  associated  with  the  various  treatments.  The  remaining 
coefficients  are  found  in  a  like  manner. 

Example  10.11 

Consider  a  research  situation  similar  to  that  described  in  Example 
10.10,  but  involving  five  treatments.  Suppose  that  four  batteries  are 
allocated  to  each  treatment.  If  treatment  No.  2  represents  a  commonly 
used  electrolyte,  while  Nos.  1,  3,  4,  and  5  are  newly  developed  electro 
lytes  in  which  Nos.  1  and  3  are  of  type  A  and  Nos.  4  and  5  are  of  type 
B,  the  contrasts  specified  in  Table  10.3  are  appropriate  for  the  obvious 
treatment  comparisons. 

Let  us  now  take  note  of  another  item  of  importance.  If  two  con 
trasts, 

(10.26) 


10.15       TREATMENT    COMPARISONS  263 

TABLE   10.3-Symholic  Representation  of  the  Contrasts  for  the 
Treatment  Comparisons  Specified  in  Example  10.11 


Electrolyte 

Contrast 

1                        2 

3                        4 

5 

Ci  

—  1                    +4 

i                     i 

—  1 

C2  - 

+  1                         0 

-4-1                    ~1 

—  1 

C3  

+  1                         0 

—  1                        0 

0 

C4  

0                        0 

0                    +1 

—  1 

and 

C_  ^T"*          I        ,-        T"'          F  I        ^-        HT^  /  "1  f~\      O  *7\ 

q    —     Clq**-   1    ~T~    ^2^-t  2    ~T~     *     *     "    ~T~    Ckq^-  ky  \1-\J .  £  I ) 

are  such  that 

]C  «**pcf«  =  0  (p^  g),  (10.28) 

then  the  contrast  CP  is  orthogonal  to  the  contrast  Cff.  (NOTE:  It  is 
common  practice  to  speak  of  orthogonal  contrasts  or  orthogonal  treat 
ment  comparisons.)  If  Ui  =  n  (for  all  i),  the  orthogonality  condition 
reduces  to 

23<^<^=0,          pr*q.  (10,29) 

The  reader  can  easily  verify  that  the  contrasts  specified  in  Examples 
10.10  and  10.11  are,  in  each  case,  orthogonal.  In  addition,  the  percep 
tive  student  will  have  noted  that  the  effects  and  interactions  discussed 
in  Section  10.14  were  also  orthogonal  contrasts. 

At  this  time?  the  following  question  might  well  be  asked,  namely, 
"Are  orthogonal  contrasts  better  than  nonorthogonal  contrasts?"  In 
tuitively,  orthogonal  contrasts  seem  to  be  preferable.  (NOTE :  Actually, 
they  are  preferable  if  one  wishes  the  estimates  derived  from  the  dif 
ferent  contrasts  to  be  uncorrelated.)  However,  occasionally  it  is  desir 
able  to  design  an  experiment  with  the  expressed  intent  of  analyzing  a 
set  of  nonorthogonal  contrasts.  In  such  cases,  the  probability  state 
ments  accompanying  the  associated  tests  of  significance  are  of  an  am 
biguous  nature  (due  to  the  correlation  between  the  contrasts),  and 
much  care  should  be  exercised  in  interpreting  the  experimental  results. 

One  final  remark  needs  to  be  made  and  then  we  may  move  on  to 
another  topic.  The  remark  is  the  following:  Regardless  of  the  desira 
bility  of  orthogonal  contrasts,  the  statistician  should  not  let  his  prefer 
ence  for  such  a  state  of  affairs  override  the  needs  of  the  researcher.  By 
this  is  meant  that,  as  nice  as  it  is  to  have  a  set  of  orthogonal  contrasts, 


264  CHAPTER    1O,    DESIGN    OF   EXPERIMENTAL    INVESTIGATIONS 

only  those  contrasts  which  are  meaningful  to  the  researcher  should  be 
analyzed. 

10.16      STEPS   IN    DESIGNING  AN    EXPERIMENT 

Each  statistician  has  his  own  list  of  steps  which  he  follows  when 
designing  an  experiment.  However,  a  comparison  of  various  lists 
reveals  that  they  all  cover  essentially  the  same  points. 

According  to  Kempthorne  (28),  a  statistically  designed  experi 
ment  consists  of  the  following  steps : 

(1)  Statement  of  the  problem. 

(2)  Formulation  of  hypotheses, 

(3)  Devising  of  experimental  technique  and  design. 

(4)  Examination  of  possible  outcomes  and  reference  back  to  the  reasons 
for  the  inquiry  to  be  sure  the  experiment  provides  the  required  in 
formation  to  an  adequate  extent. 

(5)  Consideration  of  the  possible  results  from  the  point  of  view  of  the 
statistical  procedures  which  will  be  applied  to  them,  to  ensure  that 
the  conditions  necessary  for  these  procedures  to  be  valid  are  satis 
fied. 

(6)  Performance  of  experiment. 

(7)  Application  of  statistical  techniques  to   the  experimental  results. 

(8)  Drawing  conclusions  with  measures  of  the  reliability  of  estimates  of 
any    quantities    that    are    evaluated,    careful    consideration    being 
given  to  the  validity  of  the  conclusions  for  the  population  of  objects 
or  events  to  which  they  are  to  apply. 

(9)  Evaluation  of  the  whole  investigation,  particularly  with  other  in 
vestigations  on  the  same  or  similar  problems.4 

In  a  later  section,  these  steps  will  be  illustrated  through  the  considera 
tion  of  some  design  problems. 

Since  the  designing  of  an  experiment  or  the  planning  of  a  test  pro 
gram  is  such  an  important  part  of  any  investigation,  the  statistician 
must  make  every  effort  to  obtain  all  the  relevant  information.  This 
will  usually  require  one  or  more  conferences  with  the  researcher,  and 
the  asking  of  many  questions.  It  has  been  my  experience  that  the 
amount  of  time  consumed  in  this  phase  can  be  materially  reduced  if,  at 
the  preliminary  meeting  between  the  researcher  (e.g.,  a  development 
engineer)  and  the  statistician,  time  is  taken  to  explore  the  relationship 
between  research  and/or  development  experimentation  and  the  sta 
tistical  design  of  experiments.  (NOTE:  Frequently,  there  is  a  formid 
able  communications  barrier  which  must  be  overcome.)  One  of  the 
best  ways  to  convince  the  researcher  of  the  need  for  the  multitude  of 
questions  posed  by  the  statistician  is  to  give  him  (in  the  first  meeting) 
a  "check  list"  which  specifies  various  stages  in  the  planning  of  a  test 
program.  (An  even  more  efficient  arrangement  if  you  are  the  statisti 
cian  in  an  industrial  organization  is  to  distribute  copies  of  such  a  list  to 
all  persons  who  may  at  some  time  have  need  of  your  services.)  One 

4  O.  Kempthorne,  The  Design  and  Analysis  of  Experiments,  John  Wiley  and 
Sons,  Inc.,  New  York,  1952,  p.  10. 


10.16       STEPS    IN    DESIGNING    AN    EXPERIMENT  265 

such  list,  prepared  by  Bicking  (3),  is  reproduced  below  for  your  con 
sideration. 

Check  List  for  Planning  Test  Programs 

A.  Obtain  a  clear  statement  of  the  problem 

1.  Identify  the  new  and  important  problem  area 

2.  Outline  the  specific  problem  within  current  limitations 

3.  Define  exact  scope  of  the  test  program 

4.  Determine  relationship   of  the  particular  problem  to   the  whole  re 
search  or  development  program 

B.  Collect  available  background  information 

1.  Investigate  all  available  sources  of  information 

2.  Tabulate  data  pertinent  to  planning  new  program 

C.  Design  the  test  program 

1.  Hold  a  conference  of  all  parties  concerned 

a.  State  the  propositions  to  be  proved 

b.  Agree  on  magnitude  of  differences  considered  worthwhile 

c.  Outline  the  possible  alternative  outcomes 

d.  Choose  the  factors  to  be  studied 

e.  Determine   the  practical  range  of  these  factors   and  the  specific 
levels  at  which  tests  will  be  made 

f .  Choose  the  end  measurements  which  are  to  be  made 

g.  Consider  the  effect  of  sampling  variability  and  of  precision  of  test 
methods 

h.  Consider  possible  inter-relationships  (or  "interactions")  of  the 
factors 

i.  Determine  limitations  of  time,  cost,  materials,  manpower,  instru 
mentation  and  other  facilities  and  of  extraneous  conditions,  such, 
as  weather 

j.    Consider  human  relation  angles  of  the  program 

2.  Design  the  program  in  preliminary  form 

a.  Prepare  a  systematic  and  inclusive  schedule 

b.  Provide   for  step-wise   performance  or   adaptation   of  schedule  if 
necessary 

c.  Eliminate    effect    of    variables    not   under    study   by    controlling, 
balancing,  or  randomizing  them 

d.  Minimize  the  number  of  experimental  runs 

e.  Choose  the  method  of  statistical  analysis 

f.  Arrange  for  orderly  accumulation  of  data 

3.  Review  the  design  with  all  concerned 

a.  Adjust  the  program  in  line  with  comments 

b.  Spell  out  the  steps  to  be  followed  in  unmistakable  terms 

D.  Plan  and  carry  out  the  experimental  work 

1.  Develop  methods,  materials,  and  equipment 

2.  Apply  the  methods  or  techniques 

3.  Attend  to  and  check  details;  modify  methods  if  necessary 

4.  Record  any  modifications  of  program  design 

5.  Take  precautions  in  collection  of  data 

6.  Record  progress  of  the  program 


266  CHAPTER    10,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

E,  Analyze  the  data 

1.  Reduce  recorded  data,  if  necessary,  to  numerical  form 

2.  Apply  proper  mathematical  statistical  techniques 

F.  Interpret  the  results 

1.  Consider  all  the  observed  data 

2.  Confine  conclusions  to  strict  deductions  from  the  evidence  at  hand 

3.  Test  questions  suggested  by  the  data  by  independent  experiments 

4.  Arrive  at  conclusions  as  to  the  technical  meaning  of  results  as  well 
as  their  statistical  significance 

5.  Point  out  implications  of  the  findings  for  application  and  for  further 
work 

6.  Account  for  any  limitations  imposed  by  the  methods  used 

7.  State  results  in  terms  of  verifiable  probabilities 

G*    Prepare  the  report 

1.  Describe  work  clearly  giving  background,  pertinence  of  the  problems 
and  meaning  of  results 

2.  Use  tabular  and  graphic  methods  of  presenting  data  in  good  form  for 
future  use 

3.  Supply^  sufficient  information  to  permit  reader  to  verify  results  and 
draw  his  own  conclusions 

4.  Limit  conclusions  to  objective  summary  of  evidence  so  that  the  work 
recommends  itself  for  prompt  consideration  and  decisive  action.5 

The  reader  should  realize,  of  course,  that  the  two  lists  (of  steps  in 
designing  experiments)  presented  in  this  section  are  only  guides.  Very 
seldom  will  the  various  steps  be  tackled  and  settled  in  the  particular 
order  given.^The  statistician  does  not  operate  in  such  a  mechanical  and 
routine  fashion.  Questions  will  be  asked  and  answers  received  which  will 
trigger  new  lines  of  thought,  and  thus  the  planning  conference  will  find 
itself  jumping  from  one  step  to  another  in  a  seemingly  haphazard  man 
ner.  Furthermore,  it  is  not  surprising  to  find,  as  the  conference  pro 
gresses  and  new  information  is  brought  forth,  the  same  step  being  con 
sidered  several  times.  Regardless  of  the  repetition  inherent  in  such  a 
procedure,  it  is  a  good  procedure. 

In  summary,  then,  the  designing  of  an  experiment  can  be  a  time- 
consuming  and,  occasionally,  a  painful  process.  Thus,  the  use  of  check 
lists  such,  as  those  presented  earlier  can  be  most  helpful  (as  a  supple 
ment  to  common  sense)  in  making  relatively  certain  that  nothing  has 
been  overlooked. 

10.17      ILLUSTRATIONS     OF     THE     STATISTICIAN'S     AP 
PROACH   TO   DESIGN    PROBLEMS 

To  illustrate  the  manner  in  which  a  statistician  approaches  a  design 
problem,  a  series  of  examples  will  be  considered.  The  first  of  these  will 
demonstrate  the  application  of  Kempthorne's  nine  steps,  while  the 

6  Charles  A.  Bicking,  "Some  uses  of  statistics  in  the  planning  of  experiments  " 
Industrial  Quality  Control,  Vol.  10,  No.  4,  Jan.,  1954,  p.  23. 


10.17       APPROACH    TO    DESIGN    PROBLEMS  267 

remainder   will   illustrate    various    topics   discussed   in    Sections    10.1 
through    10.16. 

Example  10.12 

Suppose  a  machine  is  constructed  for  the  purpose  of  generating  a 
random  series  of  0*s  and  1's.  If  the  machine  is  truly  a  generator  of 
random  binary  elements,  it  should,  among  other  things,  yield  03s  50  per 
cent  of  the  time  and  l*s  50  per  cent  of  the  time*  It  is  proposed  that  an 
experiment  be  devised  to  check  on  this  particular  aspect  of  the  random 
ness  of  the  machine. 

The  preceding  paragraph  illustrates  Kempthorne's  Step  1,  the 
statement  of  the  problem.  If  we  formulate  H:p0  =  %  (where  p0  stands 
for  the  probability  of  a  0)  and  A  :pQ  =?*%,  we  have  taken  care  of  Step  2. 
The  devising  of  an  experimental  technique  and  design  (Step  3)  is  fairly 
simple.  In  this  case  we  shall  operate  the  device  a  certain  number  of 
times,  say  n,  record  the  proportion  of  0Js  (po),  and  !see  if  this  is  in  close 
enough  agreement  with  the  hypothesis  H.  If  the  agreement  is  good, 
we  accept  H]  if  the  agreement  is  poor,  we  reject  H  and  accept  A,  the 
alternative  hypothesis.  The  only  remaining  part  of  Step  4  to  be  taken 
care  of  is  the  determination  of  the  number  of  operations  of  the  device 
that  are  required  before  we  feel  safe  in  making  a  decision.  Suppose  it  is 
desired  that  the  probability  of  rejecting  H  :po  =  i  (when  it  is  really  true) 
should  be  no  greater  than  a:  =  0.05.  This  implies  n>6,  as  can  easily  be 
shown.  Note  carefully  the  concept  of  rejecting  a  true  hypothesis.  The 
value  of  n  would  also  be  influenced  by  fixing  the  probability  of  accept 
ing  a  false  hypothesis,  but  we  choose  to  ignore  this  in  the  present 
example.  Step  5  consists,  in  this  case,  of  recognizing  that  the  results 
will  be  analyzed  using  the  binomial  distribution,  and  thus  we  should 
make  certain  that  the  repeated  events  (operations  of  the  device)  are 
statistically  independent.  Step  6  is  evident,  though  sometimes  trouble 
some.  When  discussing  Steps  3  and  4,  the  content  of  Step  7  was  alluded 
to,  and  all  that  remains  is  the  formalizing  of  the  analysis.  Step  8  implies 
that  we  should  produce  a  confidence  interval  estimate  of  the  true 
probability  of  producing  a  0  with  our  device;  that  is,  a  point  estimate,  pQ> 
is  not  sufficient.  We  must  also  be  very  careful  to  state  that  our  conclu 
sions  only  hold  for  the  particular  device  operated,  unless  this  device  was 
randomly  selected  from  a  larger  group  (or  population)  of  devices. 
Had  other  devices  of  a  similar  nature  been  investigated,  the  results  of 
our  experiment  should  be  evaluated  along  with  all  pertinent  informa 
tion  from  the  allied  studies  (Step  9). 

The  reader  will  probably  have  recognized  the  similarity  of  this  illus 
tration  to  Example  7.4.  It  is,  of  course,  the  same.  All  we  have  done  here 
is  "dress  up"  the  problem  and  use  it  to  illustrate  the  various  steps  in 
the  design  of  an  experiment. 

Example  10.13 

Consider  the  problem  of  an  engineer  who  wishes  to  assess  the  relative 
effects  of  eight  treatments  (for  the  moment  undefined)  on  the  activated 
life  of  a  particular  type  of  thermal  battery.  Assume  that  64  relatively 
homogeneous  batteries  are  available  for  experimentation.  With  only  this 
much  information,  the  most  efficient  design  would  be  to  randomly 
assign  the  batteries  to  the  eight  treatments  (groups)  subject  to  the 


268  CHAPTER    1 0,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

restriction  that  8  batteries  be  allocated  to  each  treatment.  Such  an 
assignment  is  illustrated  in  Table  10.4.  The  reader  should  note  that  the 
major  design  decisions  reached  in  this  example  were  concerned  with 
balancing  and  grouping.  (NOTE:  The  type  of  design  described  above  is 
known  as  a  completely  randomized  design.)^ 

TABLE   10.4-Random  Assignment  of  Batteries  to 
Treatments  as  Described  in  Example  10.13 

Treatments 
ABCDEFGH 


9 

58 

37 

18 

14 

21 

48 

43 

22 

53 

36 

38 

1 

15 

63 

56 

64 

26 

30 

33 

50 

3 

60 

41 

34 

11 

5 

29 

27 

45 

57 

23 

17 

52 

6 

61 

16 

47 

25 

10 

4 

51 

13 

40 

49 

32 

59 

12 

31 

8 

2 

35 

46 

19 

7 

20 

28 

14 

54 

39 

44 

62 

55 

42 

Numbers  in  the  table  represent  serial  numbers  of  units;  a  random  order  of  testing  would 
also  be  determined. 

Example  1O.14 

As  a  second  illustration  of  a  completely  randomized  design,  consider 
the  agronomist  who  has  28  homogeneous  experimental  plots  available 
for  testing  the  relative  effects  of  4  different  fertilizers  on  the^  yield  of  a 
particular  variety  of  oats.  A  reasonable  design  would  be  to  impose,  at 
random,  a  different  fertilizer  on  each  plot.  If  the  restriction  is  imposed 
that  7  of  the  experimental  plots  be  allocated  to  each  fertilizer  (treat 
ment),  complete  balance  will  have  been  achieved. 

Example  10.15 

Referring  to  Example  10.13,  suppose  you  are  now  advised  that  the 
64  batteries  consist  of  8  batteries  from  each  of  8  different  production 
lots.  How  will  this  additional  information  affect  the  design?  If  it  is 
suspected  that  there  are  real  differences  among  the  lots,  the  precision 
of  the  experiment  can  be  improved  by  removing  the  lot-to-lot  variation 
from  the  estimate  of  experimental  error.  Such  an  improvement  in  de 
sign  may  be  accomplished  by  assigning  the  treatments  to  the  batteries 
at  random  within  each  lot.  Such  a  restricted  randomization  is  illustrated 
in  Table  10.5.  (NOTE:  The  type  of  design  described  above  is  known  as 
a  randomized  complete  block  design.)7 

The  major  benefit  resulting  from  this  type  of  blocking  is  a  gain  in 
efficiency  in  analysis.  That  is,  more  sensitive  tests  of  significance  for 
treatment  differences  can  be  made  and  shorter  confidence  interval 
estimates  of  treatment  effects  can  be  obtained. 

6  See  Chapter  11  for  further  discussion  of  completely  randomized  designs. 

7  See  Chapter  12  for  further  discussion  of  randomized  complete  block  designs. 


1O.T7       APPROACH    TO    DESIGN    PROBLEMS  269 

TABLE   10.5— Random  Assignment  of  Treatments  to  Batteries 
Within  Lots  as  Described  in  Example  10.15 

Lots 


1-ff 

9-H 

17  -R 

25-C 

33-J5 

41-G 

49-3 

57-Z> 

2-C 

1CKE 

18-  A 

26-D 

34-F 

42-H 

50-.E 

58-  A 

3-F 

ll-D 

19-B 

27  -E 

33-D 

43-B 

51-6= 

59-F 

4-B 

12-F 

2Q-G 

2S-B 

36-G 

44-C 

52-F 

60-G 

5-& 

13-G 

21-C 

29-H 

37-C 

45-JS 

53-H 

61-C 

6-G 

14-C 

22-F 

30-G 

3S-A 

46-A 

54-A 

62-H 

7~D 

15-B 

23-D 

31-F 

39-J3 

47-F 

55~D 

63-B 

S-A 

16-  A 

24r-H 

32-^1 

4Q-H 

48-£> 

56-C 

64-JS 

Numbers  in  the  table  represent  serial  numbers  of  units;  the  letters  represent  treatments. 
It  will  be  observed  that  we  have  assumed  Lot  ISTo.  1  contains  batteries  1  to  8,  Lot  No.  2 
contains  batteries  9  to  16,  etc. 

Example  10.16 

Another  illustration  of  a  randomized  complete  block  design  is  pro 
vided  by  the  following  problem  in  nutrition  research..  A  nutritionist 
wishes  to  assess  the  relative  effects  of  four  newly  developed  rations  on 
the  weight-gaining  ability  of  rats.  He  has  20  rats  available  for  experi 
mentation.  Examination  of  the  pedigrees  of  the  experimental  animals 
indicates  that  the  20  rats  consist  of  4  rats  from  each  of  5  litters.  The 
statistician  would,  under  these  circumstances,  recommend  that  the 
rations  be  assigned  to  the  rats  at  random  within  each  litter  (block). 

Example  10.17 

Consider  next  a  somewhat  more  complex  problem.  Assume  that  we 
are  again  concerned  with  testing  batteries.,  but  this  time  the  problem 
arises  during  the  development  phase.  The  development  engineer  has  to 
reach  a  decision  about  three  things:  (1)  how  much  electrolyte  should 
be  incorporated  in  this  particular  model,  (2)  what  weight  of  heat  paper 
should  be  used  in  the  construction  of  the  batteries,  and  (3)  what  effect 
will  the  temperature  at  which  the  batteries  are  activated  have  on  the 
activated  life  of  the  batteries? 

Denoting  electrolyte  by  a,  heat  paper  by  6,  and  temperature  by  c, 
and  assuming  that  two  levels  of  each  factor  are  to  be  investigated,  the 
eight  treatment  combinations  might  be  as  shown  in  Table  10.6. 

It  is  decided  that  16  batteries  will  be  built  to  each  of  the  four  "elec 
trolyte-heat  paper'7  specifications,  providing  a  total  of  64  batteries  for 
testing.  As  a  precaution  against  bias  being  introduced  because  the  last 
batteries  built  might  be  better  than  the  first  batteries  built,  the  64  bat 
teries  will  be  built  in  a  random  order.  Next,  in  each  set  of  16  batteries, 
8  will  be  randomly  selected  for  testing  at  low  temperature,  and  the 
remaining  8  will  be  reserved  for  testing  at  high  temperature. 

When  this  stage  is  reached,  that  is,  once  each  battery  has  been  built 
and  assigned  a  test  temperature,  the  64  batteries  will  be  arranged  in 
random  order  for  individual  testing.  As  you  can  probably  anticipate, 


27O 


CHAPTER    TO,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 


TABLE   10.6-Factors,  Factor  Levels,  and  Treatment  Combinations  for 
the  Experiment  Described  in  Example  10,17 


Fa 

ctors  and  Factor  Le 

vels 

Treatment 
Combination 

Amount  of 
Electrolyte 
(gm/cell) 

Weight  of  Heat 
Paper  (gm/cell*) 

Test 
Temperature  (°F) 

(1) 

1 

4 

—  50 

<z  

2 

4 

—  50 

b  

1 

6 

—  50 

ab.              .... 

2 

6 

—  50 

c  

1 

4 

100 

ac  

2 

4 

100 

be  

1 

6 

100 

abc  

2 

6 

100 

this  last  restriction  frequently  proves  to  be  unpopular,  especially  if  only 
one  temperature  chamber  is  available.  (NOTE:  The  design  that  has 
been  formulated  is  a  completely  randomized  design  involving  a  23  fac 
torial  with  8  experimental  units  per  treatment.) 

Example  10.18 

Suppose  that  we  now  consider  a  slightly  different  problem.  Like  many 
of  us  engaged  in  research,  the  development  engineer  is  often  hard  pressed 
for  funds.  If  this  were  the  case  in  the  situation  described  in  Example 
10.17,  the  development  engineer  might  place  a  preliminary  order  for  8 
batteries.  Of  the  8,  2  would  be  assembled  to  each  of  the  4  "electrolyte- 
heat  paper"  combinations.  His  plan,  of  course,  would  be  to  test  1  bat 
tery  in  each  pair  at  low  temperature  and  1  at  high  temperature.  This 
testing  would,  naturally,  take  place  in  a  random  order. 

Next,  assume  that,  after  the  first  8  batteries  are  built  and  tested, 
funds  are  made  available  for  the  building  and  testing  of  8  additional 
batteries.  These  would  be  ordered  without  delay  in  order  to  provide 
some  replication  of  the  experiment.  However,  due  to  the  way  in  which 
the  batteries  were  produced  and  tested,  that  is,  first  8  and  then  8  more, 
it  is  clear  that  the  combined  analysis  of  all  16  batteries  must  take  into 
account  the  blocking  which  is  implicit  in  the  data.  (NOTE:  In  this 
example  we  have  a  randomized  complete  block  design  consisting  of  two 
blocks  and  involving  a  23  factorial  set  of  treatment  combinations.) 

Example  10.19 

Referring  again  to  the  problem  described  in  Example  10.17,  suppose 
that  two  additional  complications  arise:  (1)  only  8  batteries  can  be 
tested  in  a  normal  work  day,  and  (2)  in  the  interests  of  economy,  the 
test  engineer  wishes  to  place  4  batteries  in  the  temperature  chamber  at 
the  same  time.  Under  these  restrictions,  we  have  a  natural  set  of  blocks, 
namely,  days.  Further,  within  each  block  it  would  be  desirable  to  test 
1  battery  corresponding  to  each  of  the  8  treatment  combinations. 
Because  of  the  temperature  chamber  restriction,  we  would  decide,  ran- 


TO. 18       ADVANTAGES    AND    DISADVANTAGES    OF    DESIGNED    EXPERIMENTS  271 

domly  for  each  day,  whether  to  first  test  batteries  at  high  temperature 
and  then  test  batteries  at  low  temperature,  or  vice  versa.  Once  this 
decision  is  made,  the  random  order  of  testing  batteries  within  tempera 
tures  must  be  specified.  As  might  be  expected,  the  eventual  analysis  of 
the  data  will  take  due  cognizance  of  all  restrictions  placed  on  the  test 
program.  (NOTE:  The  type  of  design  illustrated  in  this  example  is 
known  as  a  split  plot  design*) 

10.18     ADVANTAGES  AND   DISADVANTAGES  OF  STATIS 
TICALLY   DESIGNED   EXPERIMENTS 

Having  spent  considerable  time  discussing  various  aspects  of,  and 
techniques  in,  experimental  design,  It  is  appropriate  that  the  advan 
tages  and  disadvantages  of  statistically  designed  experiments  be  con 
sidered.  These  will,  of  course,  be  expressed  In  different  ways  by  differ 
ent  people.  However,  as  "was  true  for  the  steps  involved  in  designing 
experiments,  an  examination  of  various  lists  of  advantages  and  disad 
vantages  will  show  that  all  the  lists  cover  essentially  the  same  points. 

Advantages  of  Statistically  Designed  Experiments 

Bicking  (3)  has  listed  the  advantages  of  statistical  designs  over  old 
kinds  of  designs  (nonstatistlcal)  as  follows : 

(1)  Close  teamwork  is  required  between  the  statisticians  and  the  re 
search  or  development  scientists  with  consequent  advantages  in  the 
analysis  and  interpretation  stages  of  the  program 

(2)  Emphasis  Is  placed  on  anticipating  alternatives  and  on  systematic 
pre-planning,  yet  permitting  step-wise  performance  and  producing 
only  data  useful  for  analysis  in  later  combinations 

(3)  Attention  is  focused  on  inter-relationships  and  on  identifying  and 
measuring  sources  of  variability  in  results 

(4)  The  required  number  of  tests  is  determined  reliably  and  often  may 
be  reduced 

(5)  Comparison  of  effects  of  changes  is  more  precise  because  of  group 
ing  of  results 

(6)  The  correctness  of  conclusions  is  known  with  definite  mathematical 
preciseness9 

If  these  advantages  truly  exist,  and  I  believe  they  do,  the  value  of 
statistical  aid  in  planning  experiments  is  evident  and  should  always  be 
sought. 

Disadvantages  of  Statistically  Designed  Experiments 

Happily,  there  are  more  advantages  than  disadvantages  associated 
with  statistically  designed  experiments.  In  fact,  I  found  it  somewhat 
difficult  to  formulate  a  list  of  disadvantages.  However,  a  careful  read 
ing  of  Mandelson.  (30),  together  with  a  realistic  appraisal  of  the  imple- 

8  See  Chapter  13  for  further  discussion  of  split  plot  designs. 

9  Charles  A.  Bicking,  "Some  uses  of  statistics  in  the  planning  of  experiments," 
Industrial  Quality  Control,  Vol.  10,  No.  4,  Jan.,  1954,  p.  22. 


272  CHAPTER    TO,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

mentation  of  certain  statistically  designed  experiments,  did  yield  the 
following  possible  disadvantages: 

(1)  Such  designs  and  their  analyses  are  usually  accompanied  by 
statements  couched  in  the  technical  language  of  statistics.  It 
would  be  much  better  if  the  statistician  would  translate  such 
statements  into  terms  that  are  meaningful  to  the  nonstatis- 
tician.  In  addition,  the  statistician  should  not  overlook  the 
value  of  presenting  the  results  in  graphical  form.  As  a  matter 
of  fact,  he  should  always  consider  plotting  the  data  as  a  pre 
liminary  step  to  a  more  analytical  approach. 

(2)  Many  statistical  designs,  especially  when  first  formulated,  are 
criticized  as  being  too  expensive,  complicated,  or  time-con 
suming.  Such  criticisms,  when  valid,  must  be  accepted  in  good 
grace  and  an  honest  attempt  made  to  improve  the  situation, 
provided  that  the  solution  of  the  problem  is  not  compromised. 

Before  terminating  our  discussion  of  the  advantages  and  disadvan 
tages  of  statistically  designed  experiments,  some  mention  should  be 
made  of  particular  advantages  and  disadvantages  associated  with  fac 
torials.  This  is  deemed  necessary  because  of  the  important  role  that 
factorials  play  in  the  design  and  analysis  of  experiments.  (NOTE: 
There  will  undoubtedly  be  some  overlap  between  the  advantages  and 
disadvantages  given  for  statistically  designed  experiments  in  general 
and  those  about  to  be  given  for  factorials.  However,  a  small  amount 
of  repetition  will  not  be  harmful.) 

Advantages  of  Factorials 

(1)  Greater  efficiency  in  the  use   of  available  experimental  re 
sources  is  achieved. 

(2)  Information  is  obtained  about  the  various  interactions. 

(3)  The  experimental  results  are  applicable  over  a  wider  range  of 
conditions;  that  is7  due  to  the  combining  of  the  various  fac 
tors  in  one  experiment,  the  results  are  of  a  more  comprehen 
sive  nature. 

(4)  There  is  a  gain  due  to  the  hidden  replication  arising  from  the 
factorial  arrangement. 

Disadvantages  of  Factorials 

(1)  The  experimental  setup  and  the  resulting  statistical  analysis 
are  more  complex. 

(2)  With  a  large  number  of  treatment  combinations  the  selection 
of  homogeneous  experimental  units  becomes  more  difficult. 

(3)  Certain  of  the  treatment  combinations  may  be  of  little  or  no 
interest;   consequently,   some  of  the  experimental  resources 
may  be  wasted. 


PROBLEMS  273 

10.19     Summary 

In  this  chapter  several  extremely  important  topics  have  been  dis 
cussed.  It  is  recommended  that  they  be  re-examined  from  time  to  time 
as  the  reader  progresses  through  the  succeeding  chapters.  Such  a  peri 
odic  reappraisal  "will  prove  beneficial  for  a  number  of  reasons,  for  ex 
ample:  (1)  a  thorough  understanding  of  the  concepts,  principles,  and 
techniques  involved  is  essential  for  a  fruitful  study  of  the  experi 
mental  designs  which  are  the  subjects  of  the  next  three  chapters,  and 
(2)  an  appreciation  of  these  important  principles  should  manifest  itself 
in  improved  experimentation. 

Problems 

10.1  Choosing   practical   situations   from   your   special  field   of  interest, 
describe  three  problems  whose  solutions  must  be  determined  experi 
mentally. 

10.2  With  reference  to  Problem  10.1,  discuss  the  need  for  an  experimental 
design  in  each  of  the  three  illustrations. 

10.3  It  is  sometimes  said  that  experimental  design  is  a  subject  which  con 
sists  of  two   (almost  distinct)   parts:  (a)  the  choice  of  treatments, 
experimental  units,  and  characteristics  to  be  observed;  (b)  the  choice 
of  the  number  of  experimental  units  and  the  method  of  assigning  the 
treatments  to  the  experimental  units.  Discuss  this  classification  from 
the  points  of  view  of  the  researcher  and  the  statistician. 

10.4  Define  "systematic  error"  and  discuss  the  relationship  between  this 
factor  and  the  statistical  design  of  experiments. 

10.5  Some    terms    that    occur   rather    frequently    in    the    literature    are: 
(a)  accuracy,  (b)  precision,  (c)  validity,  (d)  reliability,  and  (e)  bias. 
Restricting  your  remarks  to  the  theory  of  statistics  or  to  applica 
tions  of  statistical  methods,  define  and  discuss  each  of  these  terms. 

10.6  Cox  (14)  uses  "Designs  for  the  Reduction  of  Error"  as  the  title  of 
one  of  his  chapters.  What  does  this  title  suggest  to  you? 

10.7  With  reference  to  factorials,  what  is  meant  by  the  phrase  "hidden 
replication"? 

10.8  Discuss  the  use  of  concomitant  information  in  experimental  design. 

10.9  Choosing  practical  situations  from  your  own  special  field  of  interest, 
illustrate  the  concept  of  confounding.  Give  examples  of:  (a)  unavoid 
able  confounding,  (b)  unintentional  confounding,  and  (c)  intentional 
confounding. 

10.10  Choosing  practical  situations  from  your  own  special  field  of  interest, 
illustrate  the  concept  of  randomization, 

10.11  With  reference  to  Problem   10.10,    discuss  the  difficulties   (if  any) 
associated  with  the  randomization  process. 

10.12  Give  your  interpretation  of  the  phrase  "restricted  randomization." 

10.13  What  would  you  do  if,  in  the  planning  of  a  randomized  complete 
block  design,  the  same  order  of  treatments  occurred  (randomly)  in 
each  block? 

10.14  Discuss  the  following  ways  in  which  treatments  can  be  assigned  to 
experimental  units:  (a)  randomly,  (b)  subjectively,  and  (c)  system 
atically.   Give  illustrations  which   show  the  benefits,   dangers,  and 
difficulties  involved  in  each  of  the  three  approaches. 


274  CHAPTER    TO,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

10.15  Cox  (14)  uses  the  phrase  "Randomization  as  a  Device  for  Conceal 
ment"  as  the  heading  of  one  of  the  sections  in  his  book.  Without 
referring  to  his  discussion,  what  do  you  believe  he  has  in  mind? 

10.16  Cox  (14)  makes  a  distinction  between  factors  that  represent  a  treat 
ment  applied  to  the  experimental  units  (treatment  factors)  and  fac 
tors  that  correspond  to  a  classification  of  the  experimental  units  into 
two  or  more  types  (classification  factors).  Give  illustrations  of  each 
of  these  from  situations  in  your  own  special  field  of  interest. 

10.17  The  statement  has  been  made  that  an  uncontrolled  and  unmeasured 
variable  may  be  of  sufficient  importance  to  lead  to  the  conclusion 
that  two  controlled  factors  interact  to  a  significant  degree.  Discuss 
this  idea,  including  all  possible  implications.   What  safeguards  do 
we  have  against  such  a  result  occurring? 

10.18  Cox  (14)  also  states  that  it  is  sometimes  convenient  to  classify  factors 
as  follows:  (a)  specific  qualitative  factors,   (b)   quantitative  factors, 

(c)  ranked  qualitative  factors,  and  (d)  sampled  qualitative  factors. 
How  would  you  define  each  of  these?  Compare  your  ideas  with  those 
expressed  by  Cox. 

10.19  Show  graphically  what  is  meant  by  an  interaction.  Illustrate  your 
ideas  using  the  data  of  Examples  10.8  and  10.9. 

10.20  Explain  the  relationship,  if  any,  between  regression  functions  (i.e., 
response  functions)  and  the  concepts  of  effects  and  interactions. 

10.21  How  would  you  go  about  selecting  the  factors  to  be  investigated  in 
an  experiment?  Illustrate  with  examples  from  your  own  specific  field 
of  interest. 

10.22  Assuming  the  factors  have  been  decided  upon,  how  would  you  go 
about  selecting  the  factor  levels?  Illustrate  with  examples  from  your 
own  specific  field  of  interest. 

10.23  Choosing  practical  situations  from  your  own  special  field  of  interest, 
illustrate  completely  randomized,  randomized  complete  block,  and 
split  plot  designs. 

10.24  Indicate  how  the  examples  provided  in  answer  to   Problem  10.23 
attempted  to  "control  error/' 

10.25  What  is  meant  by  the  precision  of  an  experiment?  of  a  contrast? 

10.26  What  is  meant  by  a  sequential  experiment?  Is  there  any  other  kind? 
Please  discuss. 

10.27  In  Problem  10.3,  reference  was  made  to  £i  .  .  .  the  choice  of  treat 
ments,  experimental  units,  and  characteristics  to  be  observed."  Illus 
trate  each  of  these  with  examples  from  your  own  special  field  of 
interest. 

10.28  Discuss  the  following  items  relative  to  the  selection  of  experimental 
units:   (a)   number  of  units,    (b)   size   of  units,    (c)    shape   of  units, 

(d)  independence  of  units. 

10.29  What  is  meant  by  a  "control"  treatment? 

10.30  Cox  (14)  classifies  observations  into  six  groups:  (a)  primary  observa 
tions,  (b)  substitute  primary  observations,  (c)  explanatory  observa 
tions,     (d)     supplementary    observations    for    increasing    precision, 

(e)  supplementary    observations    for    detecting    interactions,     and 

(f)  observations   for   checking   the   application   of  the   treatments. 
Please  try  to  define  and  illustrate  each  of  these.  Then  compare  your 
ideas  with  those  expressed  by  Cox. 


REFERENCES    AND    FURTHER    READING  275 

10.31  Building  on  the  samples  given  in  Section  10.16,  construct  your  own 
list  of  " steps  in  designing  an  experiment." 

10.32  Contrast  the  one-factor-at-a-time  method  of  experimentation  with 
the  factorial  approach.  Construct  a  table  which  shows  and  compares 
the  advantages  and  disadvantages  of  each. 

10.33  Define:  (a)  absolute  experiments  and  (b)  comparative  experiments. 
Give  examples  of  each.  With  which  type  is  this  book  mainly  con 
cerned? 

10.34  Consider  the  following  ''elements"  of  experimental  method: 

(a)   control,  or  the  elimination  of  the  effects  of  extraneous  variables 
(6)    accuracy  of  instruments  and  data  acquisition 

(c)  reduction  of  the  number  of  variables  to  be  investigated 

(d)  planning  of  the  test  sequence  in  advance  of  the  start  of  experi 
mentation 

(e)  detection  of  malfunctions 

(/)    testing  for  reasonableness  of  results 

{g}    analysis  and  interpretation  of  results 

Evaluate  the  foregoing  list  by  comparing  it  with  the  ideas  expressed 

in  this  chapter. 

10.35  Choosing  practical  situations  from  your  own  special  field  of  interest, 
give  three  examples  of  statistically  designed  experiments.  For  each 
of  these,  point  out  where  and  how  the  concepts  of  this  chapter  were 
employed. 

10.36  Choose  a  practical  problem  in  your  own  area  of  specialization.  Fol 
lowing,  where  practicable,  the  philosophy  expressed  in  this  chapter, 
design  an  experiment  to  provide  data  relevant  to  the  problem.  Justify 
all  of  your  decisions  and  relate  them  to  the  discussion  in  the  text.  If 
feasible,  perform  the  experiment  and  then  analyze  and  interpret  the 
results. 

References  and  Further  Reading 

1.  Anscombe,  F.  J.  Quick  analysis  methods  for  random  balance  experimenta 
tion.  Technometrics,  1  (No.  2)  :195-209,  May,  1959. 

2.  Barbacki,  S.,  and  Fisher,  R.  A.  A  test  of  the  supposed  precision  of  systematic 
arrangements.  Ann.  Eugen.,  7:189,  1936. 

3.  Bicking,  C.  A.  Some  uses  of  statistics  in  the  planning  of  experiments.  In 
dustrial  Quality  Control,  10  (No.  4)  :20-24,  Jan.,  1954. 

4.  Bingham,    R.    S.,   Jr.    Design   of  experiments  from  a   statistical  viewpoint, 
Parts  I  and  II.  Industrial  Quality  Control,  15  (No's.  11  and  12)  :29-34  and 
12-15,  May  and  June,  1959. 

5.  Bross,  I.  D.  J.  Design  for  Decision.  The  Macmillan  Company,  New  York, 
1953. 

6.  Brownlee,   K.   A.   The  principles  of  experimental  design.  Industrial  Quality 
Control,  13  (No.  8):12-20,  Feb.,  1957. 

7    _  Statistical    Theory    and    Methodology    in    Science    and    Engineering, 

John  Wiley  and  Sons,  Inc.,  New  York,  1960. 
8.   Budne,  T.  A.  Random  balance:  Part  I — The  missing  statistical  link  in  fact 

finding  techniques,    Part   II — The   techniques   of  analysis,    Part  III — Case 

histories.  Industrial  Quality  Control,  15  (No's.  10-11-12)  :5-10,  11-16,  16-19, 

April,  May,  and  June,  1959. 
ga  _  The  application  of  random   balance  designs.  Technometrics}  1  (No. 

2):  139-55,  May,  1959. 


276  CHAPTER    10,    DESIGN    OF    EXPERIMENTAL    INVESTIGATIONS 

10.  Caplan,  F.  Statistical  design  in  electronics  production-line  experimentation. 
Industrial  Quality  Control,  12  (No.  5):12-13,  Nov.,   1955. 

11.  Chapin,    F.   S.   Experimental  Designs  in  Sociological  Research.   Harper  and 
Brothers  Publishers,  New  York,  1947. 

12.  Chew,  V.,  (editor)  Experimental  Designs  in  Industry.  John  Wiley  and  Sons, 
Inc.,  New  York,  1958. 

13.  Cochran,  W.  G.,  and  Cox,   G,  M.  Experimental  Designs.  Second  Ed.  John 
Wiley  and  Sons,  Inc.,  New  York,  1957. 

14.  Cox,  D.  R.  Planning  of  Experiments.  John  Wiley  and  Sons,  Inc.,  New  York, 
1958. 

15.  Crump,  S.  L.  Some  aspects  of  experimental  design.  Industrial  Quality  Con 
trol,   10  (No.  4):14-16,  Jan.,  1954. 

16.  Davies,   O.  !L.   (editor)    The  Design  and  Analysis  of  Industrial  Experiments. 
Second  Ed.  Oliver  and  Boyd,  Edinburgh,  1956. 

17.  DeLury,   33.   B.   On  the  design  of  experiments.  Industrial  Quality   Control, 
10  (No.  4):24-29,  Jan.,  1954. 

18.  .  Designing   experiments   to   isolate    sources    of  variation.  Industrial 

Quality  Control,  11  (No.  2)  :22-24,  Sept.,  1954. 

19.  Duffett,  J.  R.  Some  experience  with  the  design  of  experiments.  Industrial 
Quality  Control,  11  (No.  3):  36-40,  Nov.,  1954. 

20.  Federer,  W.  T.  Experimental  Design.  Macmillan  Co.,  New  York,  1955. 

21.  Finney,  D.  J.  Experimental  Design  and  Its  Statistical  Basis.  The  University 
of  Chicago  Press,  Chicago,  1955. 

22.  .  An  Introduction  to  the  Theory  of  Experimental  Design.  The  Univer 
sity  of  Chicago  Press,  Chicago,  1960. 

23.  Fisher,  R.  A.  Statistical  Methods  for  Research  Workers.  Tenth  Ed.  Oliver  and 
Boyd,  Edinburgh,  1946. 

24.  .  The   Design   of  Experiments.   Fourth   Ed.    Oliver   and   Boyd,    Edin 
burgh,  1947. 

25.  Gilbert,  S.  Statistical  design  of  experiments  in  metallurgical  research.  In 
dustrial  Quality  Control,  12  (No.  5):13-18,  Nov.,  1955. 

26.  Hunter,  J.   S.   Determination  of  optimum  operating  conditions  by  experi 
mental   methods,    Part   II— 1-2-3,    Models   and   methods.  Industrial   Quality 
Control,  15  (No's.  6-7-8)  :16-24,  7-15,  and  6-14,  Dec.,  1958,  Jan.  and  Feb., 
1959. 

27.  Jeffreys,  H.  Random  and  systematic  arrangements.  Biometrika,  31:1,  1939. 

28.  Kempthorne,   O.   The  Design  and  Analysis  of  Experiments.  John  Wiley  and 
Sons,  Inc.,  New  York,  1952. 

29.  Leone,  F.  C.,  Nottingham,  R.  B.,  and  Zucker,  J.  Significance  tests  and  the 
dollar  sign.  Industrial  Quality  Control,  13  (No.  12)  :5-20,  June,  1957. 

30.  Mandelson,  J.  The  relation  between  the  engineer  and  the  statistician.  In 
dustrial  Quality  Control,  13  (No.  ll):31-34,  May,  1957. 

31.  Mood,    A.    M.,    The    heart   of   a   reliability   program.    IRE    Transaction    on 
Reliability  and  Quality  Control,  PGRQC-16:16-23,  June,  1959. 

32.  National  Bureau   of   Standards.    Projects   and   Publications   of  the   National 
Applied  Mathematics  Laboratories.  April  through  June,  1949. 

33^  9  Economy  in  the  planning  of  experiments.  Industrial  Quality  Control, 

14  (No.  7):5-6,  Jan.,  1958. 

34.  Peach,   P.   The   use   of  statistics   in   the  design   of   experiments.   Industrial 
Quality  Control,  3  (No.  3):15-17,  Nov.,  1946. 

35.  Pearson,  E.  S.  Some  aspects  of  the  problem  of  randomization.  Biometrika, 
29:53,  1938. 

35    _  ^n  illustration  of  "Student's"  inquiry  into  the  effect  of  balancing 

in  agricultural  experiments.  Biometrika,  30:159,  1938. 

37.  ,  and  Wishart,  J.  (editors).  "Student's"  Collected  Papers.  Biometrika 

Office,  University  College,  London,  1942. 

38.  Purcell,    W.    R.    Balancing    and    randomizing    in    experiments.    Industrial 
Quality  Control,  7  (No.  4):7-14,  Jan.,  1951. 


REFERENCES    AND    FURTHER    READING  277 

39.  Quenouille,   M.  H.   The  Design  and  Analysis  of  Experiment.   Charles  Griffin 
and  Co.,  Ltd.,,  London,  1953. 

40.  Ratner,   R.  A.  Effect  of  variations  in  weight  upon  move  times.   Master  of 
Science  Thesis,  Iowa  State  University,  Ames,  1951. 

41.  Satterthwaite,    F.    E.    Random   balance    experimentation.    T echnometrics  y    1 
(No.  2)  :1 11-37,  May,  1959. 

42.  Shainin,   D.   The  statistically  designed  experiment.   Harvard  Business  Rev., 
July-Aug.,  1957. 

43.  Snedecor,  G.  W.  Statistical  Methods.  Fifth  Ed.   The  Iowa  State  University 
Press,  Ames,  1956. 

44.  "Student"    (W.    S.    Gosset).    Comparison    between    balanced    and    random 
arrangements  of  field  plots.  Biometrika,  29:363,  1938. 

45.  Wilson,  E.  B.,  Jr,  An  Introduction  to  Scientific  Research.  McGraw-Hill  Book 
Company,  Inc.,  New  York,  1952, 

46.  Yates,   F.  The  design  and  analysis  of  factorial  experiments.   Techn.  Comm. 
No.  85,  Imperial  Bureau  of  Soil  Science,  95  pp.,  1937. 

47.  Youden,  W.  J.  Statistical  design.  A  collection   by  the  editors   of  Industrial 
and  Engineering  Chemistry  of  a  series  of  bimonthly  articles  by  Dr.   W.  J. 
Youden,    National  Bureau  of  Standards,   during  his  six  years   (1954—1959) 
as  a  Contributing  Editor.  American  Chemical  Society,  Washington,  D.C. 

48.   .     Problems  of  the  experimenter.   National   Convention   Transactions, 

American  Society  for  Quality  Control,  pp.  41-47,  1959. 

49.   ,  Kempthorne,   O.,   Tukey,  J.  W.,   Box,   G.  E.  P,  and  Hunter,  J.  S. 

Discussion  of  the  papers  of  Messrs,  Satterthwaite  and  Budne  (including  au 
thors'  responses  to  discussion).  T  echnometrics,  1  (No.  2):157— 93,  May,  1959. 

50.  Zelen,  M.,  and  Connor,  W.  S.  Multi-factor  experiments.  Industrial  Quality 
Control,  15  (No.  9):14-17,  Mar.,  1959. 


CH  APTE  R    11 

COMPLETELY  RANDOMIZED  DESIGN 


CHAPTER  10,  several  experimental  designs  were  illustrated.  In  the 
present  chapter,  we  propose  to  discuss  the  simplest  of  these  designs, 
namely,  the  completely  randomized  design,  in  considerable  detail. 
Much  attention  will,  of  course,  be  given  to  methods  of  analyzing  data 
arising  from  such  a  design,  and  it  will  be  observed  that  analysis  of  vari 
ance  (frequently  abbreviated  as  AOV  or  ANOVA)  is  the  method  most 
widely  used. 

11.1      DEFINITION     OF     A     COMPLETELY     RANDOMIZED 
DESIGN 

A  completely  randomized  (CR)  design  is  a  design  in  which  the  treat 
ments  are  assigned  completely  at  random  to  the  experimental  units,  or 
vice  versa.  That  is,  it  is  a  design  that  imposes  no  restrictions,  such  as 
blocking,  on  the  allocation  of  the  treatments  to  the  experimental  units. 
Of  course,  as  in  Examples  10.13  and  10.14,  some  degree  of  balance  may 
be  sought. 

Because  of  its  simplicity,  the  completely  randomized  design  is  widely 
used.  However,  the  researcher  is  cautioned  that  its  use  should  be 
restricted  to  those  cases  in  which  homogeneous  experimental  units  are 
available*  If  such  units  cannot  be  obtained,  some  blocking  should  be 
utilized  to  increase  the  efficiency  of  the  design. 

Example  11.1 

Given  four  fertilizers,  we  wish  to  test  the  null  hypothesis  that  there 
are  no  differences  among  the  effects  of  these  fertilizers  on  the  yield  of 
corn,  We  shall  assume  there  are  20  experimental  plots  available  to  the 
research  worker*  A  sound  procedure  would  be  to  place  each  fertilizer  on 
an  equal  number  of  experimental  plots  so  that  our  estimates  of  the 
mean  effect  of  each  fertilizer  will  have  equal  weight.  Then,  we  insist 
that  the  fertilizers  be  assigned  to  the  plots  at  random.  This  may  be 
accomplished  by  numbering  our  plots  from  1  through  20  and  then 
drawing  tickets  at  random  from  a  hat,  5  tickets  being  identified  by 
coloring  or  code  mark  with  each  of  the  4  fertilizers.  The  first  one  drawn 
specifies  the  treatment  for  plot  No.  1,  the  second  for  plot  No.  2,  and 
so  on, 

Example  11.2 

If,  in  the  preceding  example,  only  17  plots  were  available,  some  lack 
of  balance  would  be  inevitable.  Assuming  that  more  precise  information 
is  desired  on  fertilizer  No.  1,  the  randomization  procedure  could  be 
modified  so  that,  for  example,  8  plots  would  be  treated  with  fertilizer 
No.  1,  3  plots  with  No.  2,  3  with  No.  3,  and  3  with  No.  4. 

t2781 


11.2       ONE    OBSERVATION    PER    EXPERIMENTAL    UNIT 


279 


11.2      COMPLETELY     RANDOMIZED    DESIGN    WITH     ONE 
OBSERVATION    PER    EXPERIMENTAL   UNIT 

If,  in  a  completely  randomized  design,  rit  experimental  units  were 
subjected  to  the  £th  treatment  (^=1,  •  -  -  }  t)  and  only  one  observation 
per  experimental  unit  was  obtained,  the  data  would  appear  as  in  Table 
11.1. 

TABLE   11.1— Symbolic    Representation  of  Data  in  a  Completely  Random 
ized  Design  (Unequal  Numbers  of  Observations  for  Each  Treatment) 


Treatment 

Total 

1 

2 

t 

Observations 

F» 

F21 

?: 

Totals 

r, 

r, 

r* 

t 
i—  i 

Numbers  of  observations 
Means 

T. 

* 

n* 
7, 

1-1 
F  -  T  /  iZ  nt 

I         £-1 

Using  the  equations: 

\  Y2  =  total  sum  of  squares 

=  sum  of  the  squares  of  all  the  observations 

=  i:  y:  F« 

S     j     /    j      •*-    131 


=  sum  of  squares  due  to  the  mean 

=  T2  /  i:  nt, 

'        1=1 

=  among  treatments  sum  of  squares 


(11.1) 


(11-2) 


(11-3) 


and 


=  experimental  error  sum  of  squares 


28O 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

TABLE    11.2-ANOVA  for  Data  of  Table  11.1 


Source  of  Variation 

Degrees  of  Freedom 

Sum  of  Squares 

Mean  Square* 

IVEean  

1 

J\ftsu 

M 

Among  treatments     .    . 

t  —  1 

T 

T 

Experimental     error     (within 
treatments)  

i:  («*-« 

i*«i 

EVV 

E 

Total 

t 

~y  ]  m 

y;  r* 

i—  1 

*  The  mean  squares  are  found  by  dividing  each  sum  of  squares  by  the  corresponding 
degrees  of  freedom.  To  avoid  confusion  with  symbols  for  effects  and  interactions  (see 
Chapter  10),  the  symbols  for  mean  squares  will  always  be  set  in  boldface  type.  This  pro 
cedure  will  be  adhered  to  throughout  the  remainder  of  this  book. 


(11.4) 


TUT         

-LYLyjf 


the  ANOVA  shown  in  Table  11.2  is  obtained.  If  each  n»  =  r&,  that  is, 
if  the  number  of  experimental  units  per  treatment  is  the  same  for  all 
treatments,  Table  11.1  would  be  modified  as  shown  in  Table  11.3. 
Equations  (11.1)  through  (11.4)  would  be  rewritten  as: 

t  n 

YU,  (11-5) 

(11.6) 

TABLE    11.3— Symbolic  Representation  of  Data  in  a  Completely  Random 
ized  Design  (Equal  Numbers  of  Observations  for  Each  Treatment) 


Treatment 

1 

2 

.  .  . 

t 

Total 

F" 

F2i 

Ytl 

Observations 

F12 

f22 

f'2 

h. 

h. 

Ytn 

t 

Totals 

TI 

Tz 

Tt 

—     ^>  *   T- 

—      /  -  •*•  * 

i—  i 

Numbers  of  observations 

n 

n 

n 

tn 

Means 

^ 

F2 

Yt 

7--T/*» 

11.2      ONE   OBSERVATION    PER    EXPERIMENTAL    UNIT 


281 


and 

The  resulting  ANOVA  is  shown  in  Table  11.4. 

TABLE  11.4-ANOVA  for  Data  of  Table  11.3 


(11-7) 
(11-8) 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

JMean              

1 

Mint 

M 

Among  treatments    .        

t—  1 

T 

T 

Experimental  error  (within  treat 
ments)    

t(n—  1) 

J&tnj 

E 

Total 

tn 

T!  Y2 

Up  to  this  point,  our  discussion  of  a  completely  randomized  design 
with  one  observation  per  experimental  unit  has  concentrated  on  the 
calculation  of  the  various  sums  of  squares  and  mean  squares,  and  on 
the  specification  of  the  associated  degrees  of  freedom.  While  the  calcu 
lation  of  the  sums  of  squares  and  mean  squares  has  been  explained  in 
detail,  no  explanation  has  been  given  as  to  why  the  degrees  of  freedom 
are  as  stated.  However,  the  way  in  which  the  degrees  of  freedom  are 
found  seems  reasonably  clear.  Since  the  procedure  will  be  illustrated 
many  times  in  this  and  succeeding  chapters,  no  attempt  will  be  made 
to  formulate  and  state  general  rules. 

Before  the  preceding  analyses  of  variance  can  be  used  for  purposes 
of  statistical  inference,  certain  assumptions  must  be  made  about  the 
observations.  The  nature  of  these  assumptions  will  now  be  examined. 
(NOTE:  In  general,  the  assumptions  underlying  analyses  of  variance 
are  the  same  as  those  usually  associated  with  regression  analyses. 
These  are  additivity,  linearity,  normality,  independence,  and  homo 
geneous  variances.  That  is,  the  statistical  model  most  frequently  assumed 
in  analysis  of  variance  applications  is  a  linear  model  to  which  has  "been 
appended  certain  restrictions  about  independent  observations  from  normal 
distributions.) 

The  basic  assumption  for  a  completely  randomized  design  with  one 
observation  per  experimental  unit  is  that  the  observations  may  be 
represented  by  the  linear  statistical  model 


=   JUL 


y  =  1,  •  •  •  ,  n*  (unequal  numbers) 


or 


(11.9) 


j  =  1,  -  -  •  ,  n    (equal  numbers) 


282 


CHAPTER    11 ,    COMPLETELY    RANDOMIZED    DESIGN 


where  /x  is  the  true  mean  effect,  n  is  the  true  effect  of  the  ith.  treatment, 
and  €tj  is  the  true  effect  of  the  yth  experimental  unit  subjected  to  the 
ith  treatment.  (NOTE :  e^  also  includes  the  effects  of  all  other  extra 
neous  factors.  However,  we  rely  on  the  process  of  randomization  to 
prevent  these  effects  from  contaminating  our  results.)  In  addition,  it 
is  customarily  assumed  that  M  is  a  constant  while  the  e»/  are  NID  (0,,  00. 

However,  the  specification  of  the  model  is  still  incomplete,  for 
nothing  has  been  said  about  the  T,-.  The  researcher  has  two  choices  as 
to  what  he  can  say  about  the  rt-,  namely:  (1)  2^i=i  rt  =  0,  which  reflects 
the  researcher's  decision  that  he  is  concerned  only  with  the  t  treat 
ments  present  in  his  experiment,  or  (2)  the  T*  are  NID  (0,  o-T),  which 
reflects  the  researcher's  decision  that  he  is  concerned  with  a  population 
of  treatments  of  which  only  a  random  sample  (the  t  treatments)  are 
present  in  his  experiment.  These  two  choices  lead  to  what  the  statis 
tician  refers  to  as  Model  I  and  Model  II,  respectively.  Incidentally, 
Model  I  is  sometimes  referred  to  as  the  analysis  of  variance  (fixed 
effects)  model,  while  Model  II  is  known  as  the  component  of  variance 
(random  effects)  model. 

Once  the  foregoing  assumptions  have  been  made,  the  theory  out 
lined  in  Chapter  3  may  be  invoked  to  obtain  "expected  mean  squares." 
These  expected  mean  squares  can  be  of  valuable  assistance  to  the  re 
searcher,  for  they  indicate  the  proper  procedure  to  be  followed  in  esti 
mating  parameters  and/or  testing  hypotheses  about  parameters  within 
the  framework  of  the  assumed  model.  It  is  customary  to  exhibit  these 
expected  mean  squares  in  an  additional  column  in  ANOVA  tables.  So, 
without  further  discussion  at  this  time,  we  re-exhibit  Table  11.2  as 
Table  11.5  (Model  I)  and  Table  11.6  (Model  II)  with  certain  expected 
mean  squares  included.  Similarly,  Table  11.4  is  re-exhibited  as  Table 
11.7  (Model  I)  and  Table  11.8  (Model  II).  (NOTE:  While  formal  deri- 


TABLE   11.5-ANOVA  for  Data  of  Table  11.1  Showing  Certain  Expected 
Mean  Squares  (Unequal  Number  of  Observations  per  Treatment:  Model  I) 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean  Square 

Mean  

1 

M 

M 

Among  treatments  .... 

t  —  1 

1  yy 

T 

-+£vV(«-« 

Experimental  error.  .  . 

Z  frt  -  1) 

i—  1 

Eyy 

E 

"' 

Total 

t 

ZF2 

i—  1 

11.2       ONE   OBSERVATION    PER    EXPERIMENTAL    UNIT 


283 


TABLE   11.6-ANOVA  for  Data  of  Table  11.1  Showing  Certain  Expected  Mean 
Squares  (Unequal  Numbers  of  Observations  per  Treatment:  Model  II) 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square* 

IVIean 

1 

JWi/t/ 

M 

Among  treatments  

t  —  1 

T 

T 

0.2  _J_  nQ03 

Experimental  error  

]£  (ni  -.  i) 

J~L*f~j 

E 

<r* 

T«l 

Total 

t 

23  **»* 

S  F2 

»—  1 

*  The  constant  no  is  a  sort  of  an  average  nt,  and  it  Is  defined  by 
wo  =  f  X>*  -   i  »?/  Z^l/Cf  -  1). 

l_    x_l  i«l  t=l  -J 


TABLE   11.7-ANOVA  for  Data  of  Table  11.3  Showing  Certain  Expected  Mean 
Squares  (Equal  Numbers  of  Observations  per  Treatment:  Model  I) 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square 

Mean 

1 

M 

M 

Among  treatments.  .  .  . 
Experimental  error.  .  . 

t  —  1 
t(n  -  1) 

Ryy 

T 
E 

t 

»—  i 
<r2 

Total 

in 

S  F2 

TABLE   11.8-ANOVA  for  Data  of  Table  11.3  Showing  Certain  Expected  Mean 
Squares  (Equal  Numbers  of  Observations  per  Treatment:  Model  II) 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean    Square 

IS/Iean 

1 

Mm 

M 

Among  treatments              .     .  . 

t—l 

T 

w 

T 

<r2-t-^cr? 

Experimental  error      

t(n—V) 

Eiyy 

E 

o-2 

Total 

tw> 

TZ  F2 

284  CHAPTER    11,    COMPLETELY   RANDOMIZED    DESIGN 

vations  of  expected  mean  squares  will  not  be  given  in  this  book,  certain 
rules  will  be  given  to  aid  the  researcher  in  finding  these  valuable  quan- 
titles.  Until  these  rules  are  expounded,  the  reader  is  asked  to  accept 
the  results  as  given.) 

Having  performed  the  previously  indicated  calculations  and  having 
determined  certain  expected  mean  squares  (based  on  the  specified  as 
sumptions),  we  are  now  ready  to  proceed  to  the  making  of  statistical 
inferences.  Just  what  types  of  inference  will  be  made  will,  of  course, 
depend  on  the  purpose  for  which  the  experiment  was  conducted.  Three 
common  inferences  concern  themselves  with  the  following  problems: 
(1)  hypotheses  about  the  relative  effects  of  treatments,  (2)  estimation 
of  the  magnitude  of  components  of  variance,  and  (3)  estimation  of  the 
mean  effects  of  individual  treatments.  Each  of  these  will  now  be  con 
sidered. 

Consider  first  the  hypothesis  of  "no  differences  among  the  effects  of 
the  t  treatments  in  the  experiment.77  The  way  in  w-hich  this  hypothesis 
was  phrased  indicates  that  Model  I  has  been  assumed.  Thus,  the  hy 
pothesis  may  be  expressed  as  H:T^  =  O  (i=l,  •  •  •  ,  f).  Examination  of 
the  expected  mean  squares  in  Tables  11.5  and  11.7  indicates  (in  each 
case)  that,  if  H  is  true,  both  the  experimental  error  mean  square  and  the 
among  treatments  mean  square  are  estimates  of  <r2.  Thus,  if  H  is  true, 
the  ratio 

mean  square  for  treatments 

T/E  =  -  -  --  (11.10) 

experimental  error  mean  square 

is  distributed  as  F  with 

and     i>2 


degrees  of  freedom  because  of  the  assumption  that  the  e»/  are 
NID  (0,  <r).  If  the  value  of  F  specified  by  Equation  (11.10)  exceeds 
jFa-aOOiJ  K2>,  where  lOOo:  per  cent  is  the  chosen  significance  level,  H  will 
be  rejected  and  the  conclusion  reached  that  there  are  significant  dif 
ferences  among  the  t  treatments. 

Had  Model  II  been  assumed,  that  is,  had  the  hypothesis  been 
phrased  as  follows:  "There  are  no  differences  among  the  effects  of  all 
the  treatments  in  the  population  from  which  the  t  treatments  in  the 
experiment  are  a  sample,"  the  same  test  procedure  would  have  evolved. 
That  is,  under  Model  II,  the  hypothesis  H:o*  =  Q  would  also  be  tested 
by  forming  the  ratio  F  =  T/E.  Why,  then,  have  we  been  so  concerned 
over  the  distinction  between  the  two  models?  There  are  two  reasons 
for  our  concern  over  the  differences  in  assumptions  between  Models  I 
and  II.  These  are:  (1)  the  inferences  in  the  two  cases  are  about  entirely 
different  populations;  and  (2)  in  more  complex  analyses,  quite  dif 
ferent  test  procedures  may  be  indicated.  Many  illustrations  of  these 


11.2       ONE    OBSERVATION    PER    EXPERIMENTAL    UNIT  285 

differences  will  be  forthcoming  in  later  sections  of  this  and  succeeding 
chapters. 

Consider  next  the  estimation  of  components  of  variance.  Regardless 
of  which  model  is  assumed  (that  is,  Model  I  or  Model  II),  it  is  clear 
that 

s2  —  the  experimental  error  mean  square  ==  E  (11.11) 

is  an  estimate  of  o-2.  However,  if  Model  II  is  assumed,  it  is  also  possible 
to  estimate  ov  by  calculating 

2        (mean  square  for  treatments)  —  (experimental  error  mean  square) 


coefficient  of  <r*  in  the  expected  mean  square  for  treatments 
(T  —  E)/no,      for  unequal  numbers 
(T  —  E)/n,       for  equal  numbers. 


Finally,  let  us  consider  the  estimation  of  the  mean  effects  of  indi 
vidual  treatments.  It  should  be  obvious  that  a  point  estimate  of  the 
true  mean  effect  of  the  ith  treatment  (Mt  =  M+ri)  is  given  by  T\-.  How 
ever,  since  confidence  interval  estimates  are  desired,  it  is  necessary  that 
we  determine  the  standard  error  of  the  treatment  mean.  In  Section  6.5, 
the  estimated  variance  of  a  sample  mean, 


was  used  to  define  the  standard  error  of  the  mean 

.  (11  .  14) 


Consequently,  the  estimated  variance  of  the  mean  of  the  ith  treatment  in 
a  completely  randomized  design  with  one  observation  per  experi 
mental  unit  is  given  by 

^   _  experimental  error  mean  square 


number  of  observations  in  ith.  treatment         (11. 15) 
=  E/m  =  s*/m 
and  the  standard  error  of  the  mean  of  the  ith  treatment  is  given  by 

(11.16) 


Of  course,  if  each  ?iv  =  n,  the  same  standard  error  would  be  attached  to 
each  sample  mean.  A  lOOy  per  cent  confidence  interval  estimate  of  M* 
would  then  be  determined  by  calculating 


_ 

=    Y*   =F 

where  v  is  the  number  of  degrees  of  freedom  associated  with  E. 


286 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


Considerable  time  has  been  spent  in  discussing  the  analysis  of  a 
completely  randomized  design  with  one  observation  per  experimental 
unit.  For  example,  calculations  were  explained  in  detail,  assumptions 
were  carefully  stated,  expected  mean  squares  were  introduced,  and 
test  and  estimation  methods  were  developed  with  care.  Such  attention 
to  detail  was  deemed  appropriate  for  an  orderly  development  of  the 
methods  involved.  Further,  the  discussion  given  here  will  greatly  expe 
dite  the  future  presentation  of  similar  methods  for  more  complex  situa 
tions.  Some  examples  will  now  be  given. 

Example  11 .3 

Consider  that  an  experiment  similar  to  that  described  in  Example 
10.13  has  been  performed.  However,  only  four  treatments  were  investi 
gated  and  only  20  batteries  were  available  for  testing.  The  data  in 
Table  11.9  resulted.  Following  Equations  (11.5)  through  (11.8),  the 
appropriate  calculations  are: 

]T  y*  =  104,352 
Myv  =  (1444)  2/20  =  104,256,8 
Tyy  =-  [(369)2  4-  (371)2  +  (345)2  4-  (359)2]/5  —  104,256.8 

=  84.8 
Eyy  =  104,352  —  104,256.8  —  84.8  =  10.4. 

These  lead  to  the  ANOVA  shown  in  Table  11.10.  The  expected  mean 
square  for  treatments  has  been  given  for  both  Model  I  and  Model  II. 
Since  F  =  43.49  >-PT.9»cstie>  =  5,29,  the  hypothesis  ffrr^O  (i=l,  2,  3,  4) 
or  Hia^  —  Q,  whichever  applies,  is  rejected.  Since  the  number  of  obser 
vations  per  treatment  is  the  same  for  each  treatment,  the  standard  error 
of  a  treatment  mean  is  -\/Q. 65/5  =  -\/0- 13  =  0.36  second. 


TABLE  1 1 .9-Activated  Lives  of  Twenty  Thermal  Batteries  Resulting  From 
Experiment  Described  in  Example  11.3 


Treatment 

1 

2 

3 

4 

Total 

73 

74 

68 

71 

73 

74 

69 

71 

Observations  (in  seconds) 

73 

74 

69 

72 

75 

74 

69 

72 

75 

75 

70 

73 

Totals 

369 

371 

345 

359 

1,444 

Numbers  of  observations 

5 

5 

5 

5 

20 

Means 

73.8 

74.2 

69.0 

71.8 

72.2 

1  1 .2       ONE    OBSERVATION    PER    EXPERIMENTAL    UNIT 

TABLE   11.10-ANOVA  for  Data  of  Table  11.9 


287 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square 

F- 

Ratio 

Mean    

1 

104,256.8 

104,256.8 

Treatments   . 

3 

84.8 

28.27 

U  +  (s/3)  i:  A 

i—  1 

43.49 

Experimental  error 

16 

10.4 

0.65 

or 
«*  +  Sa-r 
cr2 

Total 

20 

104,352.0 

Example  11 .4 

Consider  an  experiment  to  study  the  effect  of  storage  condition  on 
the  moisture  content  of  white  pine  lumber.  Five  storage  methods  were 
investigated,  with  varying  numbers  of  experimental  units  (sample 
boards)  being  stored  under  each  condition.  The  data  in  Table  11.11 
were  obtained.  Following  Equations  (11.1)  through  (11.4),  the  appropri 
ate  calculations  are : 


=  863.36 

=  (108.8)2/14 


845.53 


•  + 


M*        L      5         '         3         '         2         l         3 
Eyy  =  863.36  -  845.53  —  10.66  =  7.17, 


t    (27.4)'    t    (7.1)' 


—  845.53  =  10.66 


TABLE   11.11-Moisture  Contents  of  Fourteen  White  Pine  Boards  Stored 

Under  Different  Conditions 


Storage  Conditions 

Total 

1 

2 

3 

4 

5 

Observations  (in  per 
cent)         

7.3 
8.3 
7.6 
8.4 
8.3 

5.4 
7.4 
7.1 

8.1 
6.4 

7.9 
9.5 
10.0 

7.1 

Totals 
Number  of  observa 
tions         

39.9 

5 
8.0 
0.4 

19.9 

3 
6.6 
0.5 

14,5 

2 
7,3 
0.6 

27.4 

3 
9.1 
0.5 

7.1 

1 
7.1 
0.9 

108.8 

14 
7.8 

Means  

Standard  errors 

288 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


This  leads  to  the  ANOVA  shown  in  Table  11.12.  Again,  the  expected 
mean  square  for  treatments  is  given  for  both  Models  I  and  II.  Since 
^  =  3.34  <jP.95C4,9)  =  3.63,  we  are  unable  to  reject  the  hypothesis 
HtTi  —  0(£  =  1,  -  -  -  ,  5)  or  flr:cr?  =  0.  The  standard  errors  of  the  treat 
ment  means,  presented  in  Table  11.11  for  convenience,  were  calculated 
using  Equation  (11.16)  where  a2  =  0.80. 

TABLE   11.12-ANOVA  for  Data  of  Table  11.11 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square 

F- 
Ratio 

Mean    ...           ... 

1 

845.53 

845.53 

Storage  conditions  .  . 
Experimental  error  . 

4 
9 

10.66 
7.17 

2.67 
0.80 

o 

5 

er2  -f-     ^  ^  WiTi/4 

i—  1 

i?r 

3.34 

Total 

14 

863.36 

THE    RELATION    BETWEEN    A    COMPLETELY    RAN 
DOMIZED    DESIGN    AND    "STUDENT'S"    *-TEST   OF 
H:jui  =  M2  VERSUS 


11.3 


In  Section  7.20  it  was  mentioned  that  the  analysis  of  variance  tech 
nique  could  be  used  as  an  alternative  to  "Student's"  £-test  when 
examining  the  hypothesis  flr:^i  =  /x2.  Clearly,  this  same  relationship 
exists  when  we  have  a  completely  randomized  design  involving  only 
two  treatments.  In  this  instance,  the  hypothesis  (under  Model  I)  of 
H  :ri  =  T3  =  0  is  equivalent  to  £T:^i  =  ^2  where  MI 


11.4     SUBSAMPLING    IN    A   COMPLETELY    RANDOMIZED 
DESIGN 

In  many  experimental  situations,  several  observations  may  be  ob 
tained  on  each  experimental  unit.  If  these  observations  are  all  on  the 
same  characteristic  (i.e..,  on  the  same  variable),  the  process  of  obtain 
ing  the  observations  is  often  referred  to  as  subsampling.  Some  examples 
of  subsampling  are: 

(1)  In  the  battery  experiment  of  Example  11.3,  several  observa 
tions  per  battery  might  have  been  obtained  by  connecting 
several  clocks  to  each  battery.  These  several  observations  per 
battery  would  be  referred  to  as  "samples  within  experimental 
units." 

(2)  In  a  field  experiment,  the  researcher  may  not  have  time  to 
harvest  (totally)  each  experimental  plot.  Thus,  he  might  ran- 


11.4      SUBSAMPLING  289 

domly  select  several  quadrats  per  plot  and  harvest  tlie  grain 
in  each  selected  quadrat.  Again,  we  would  describe  these  ob 
servations  as  "samples  within  experimental  units. " 
(3)  In  a  food  technology  experiment  involving  the  storage  of 
frozen  strawberries,  10  pints  (experimental  units)  were  stored 
at  each  of  5  lengths  of  storage  time  (treatments).  When 
ascorbic  acid  determinations  were  made  after  storage,  two 
determinations  were  made  on  each  pint  (samples  within  exper 
imental  units). 

As  you  can  well  imagine,  the  addition  of  subsampling  to  the  experi 
mental  program  will  have  an  effect  on  the  eventual  analysis.  First,  let 
us  see  what  changes  are  required  in  the  assumed  statistical  model. 
Under  conditions  such  as  have  been  described  above,  the  appropriate 
model  is 

Ytjk  ~  p.  +  Tt  +  €ij  +  17  *y*;  i   =    1,    -    *    •   ,  t 

/=!,••-,»<  (11.18) 


where  /z  is  the  true  mean  effect,  T%-  is  the  true  effect  of  the  ith  treat 
ment,  tij  is  the  true  effect  of  the  jth  experimental  unit  subjected  to  the 
ith  treatment,  and  rj^Jk  is  the  true  effect  of  the  &th  sample  taken  from 
the  jth  experimental  unit  subjected  to  the  zth  treatment.  Proceeding 
as  before,  we  assume  that  /z  is  a  constant,  that  the  «</  are  NID  (0,  <r), 
and  that  the  17^  are  NID  (0,  o-,,).  Of  course,  this  still  leaves  the  nature 
of  the  r<  unspecified.  That  is,  do  we  assume  Model  I  or  II?  This  deci 
sion  will  depend  on  the  manner  in  which  the  treatments  involved  in  the 
experiment  were  selected.  (NOTE:  In  most  experimental  situations, 
Model  I  is  appropriate  because  the  researcher  generally  selects  his 
treatments  in  a  nonrandom  fashion  and  is  only  interested  in  making 
inferences  about  the  treatments  actually  present  in  the  experiment. 
However,  since  Model  II  better  fits  some  situations,  we  will  con 
tinue  to  give  it  consideration.) 

In  order  to  simplify  as  much  as  possible  the  presentation  of  the 
method  of  calculating  the  various  sums  of  squares,  let  us  adopt  the 
following  notation: 

jy  =  total  number  of  observations  in  the  whole  experiment 

t       m 


JE»y  =  total  of  the  ntj  observations  on  the^'th  experimental  unit  sub 
jected  to  the  ith  treatment 


-2 

fc=*l 


290  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

nj 

Ti  =  total  of  all  the   ^2/  ni5  observations  on  the  ith  treatment 


">          * 

=  z  2:  YW  = 


=  total  of  all  N  observations 

t         nj      njj  t         nj 


«  =  z:  r*. 

i=i  y«i  jfc=i  t^i  j=i  i=i 

Using  the  preceding  notation,  the  various  sums  of  squares  are  found  as 
shown  below. 

=  total  sum  of  squares 


-34"i/j/  =  sum  of  squares  due  to  the  mean 

H  (11.20) 

=  T*/N9 
Tyy  =  among  treatments  sum  of  squares 

-^ 

(11.21; 


Eyy  =  experimental  error  sum  of  squares 


g       /      o        /       n* 

^)  -  z:  (  r?  /  2:  ^ 

*=i  V      '      y=i 

rw,  (11.22) 


and 

^y  =  pooled  sum  of  squares  among  samples  on  the  same 

experimental  unit  (11.23) 

-     -2WW     -      TW     -     Eyy 


11.4       SUBSAMPLING 


291 


"where 
Y  '  ij 


average  of  the  n^  observations  on  the  /th  experimental  unit 
subjected  to  the  ith  treatment 


Y  i  =  average  of  all  observations  on  the  fth  treatment 


and 

7  —  average  of  all  observations  in  the  whole  experiment 
=   T/N. 

TABLE   11.13-Generalized    ANOVA    for    a    Completely    Randomized    Design 
With  Subsampling  (Unequal  Numbers:  Model  I) 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean  Square 

JVtean 

1 

J^flJ-H 

M 

Treatments  .  .  . 

Experimental 
error 

*—  1 

i:  (**-D 

t««»i 

Tyy 

E-vv 

T 
R 

'l  +  C.<r*+  I^Ci:    nyVrV(*-l) 
i—  1    V  j-1              / 

«vK«* 

Sampling  error 

£  ib  (»»—D 

»—  i    y—  i 

Syy 

S 

2 

0-,, 

Total 

2V 

J2  Y* 

These  results  would  then  be  presented  as  in  Table  11.13  in  which  the 
constants  Ci  and  c2  are  defined  by 


(11.24) 


z:  c»*  - 


and 


t        /     rti  f       n<£  \  t         nf 

23  (  S)  ^?j  /  2D  ^*y )  —  X)  2D 

1=1  \  y=i     '     /— i     /     i=»i  j— i 


N 


(11.25) 


292 


CHAPTER    II,    COMPLETELY    RANDOMIZED    DESIGN 


Had  Model  II  been  assumed,  the  ANOVA  presented  in  Table  11.13 
would  be  exactly  the  same  except  for  the  "expected  mean  square  for 
treatments/'  which  would  appear  as  0^+c2<r2+c3ov  where 


N  — 


N 


(11.26) 


Example  11.5 

Consider  an  experiment  to  investigate  the  fermentative  conversion  of 
sugar  to  lactic  acid.  We  wish  to  compare  the  abilities  of  two  micro 
organisms  to  carry  out  this  conversion.  A  quantity  of  substrate  is  pre 
pared  and  divided  into  two  unequal  portions.  Each  portion  is  then 
divided  into  a  number  of  100  ml.  subportions  (experimental  units)  as 
follows:  No.  1,  4  units;  No.  2,  3  units.  Each  of  the  100  ml.  units  is 
Inoculated  with  one  or  the  other  of  the  two  microorganisms,  the  4  units 
being  Inoculated  with  microorganism  No.  1  and  the  3  units  with 
microorganism  No.  2.  The  fermentation  is  allowed  to  proceed  for  24 
hours,  and  then  each  experimental  unit  (100  ml.  subportion)  is  ex 
amined  for  the  amount  of  residual  sugar,  expressed  as  mg.  per  5  cc., 
to  determine  the  amount  of  change  produced  by  each  microorganism, 
the  converted  sugar  having  been  shown  previously  to  occur  as  lactic 
acid.  Varying  numbers  of  determinations  are  made  on  each  sample. 
The  data  are  recorded  in  Table  11.14. 


TABLE   11. 14- Amount  of   Unconverted   Sugar  in  the   Substrate  Following  a 

24-Hour  Fermentation  Due  to  Two  Different  Microorganisms 

(Coded  data  for  easy  calculation) 


Determi 
nations 

Microorganism  No.  1 

Microorganism  No.  2 

Sample  number 

Sample  number 

1 

2 

3 

4 

1 

2 

3 

1 

5.6 
5.7 

5.0 
5.0 
5.1 

5.4 
5.4 

5.4 
5.5 
5.4 

5.3 
5.5 

7.6 
7.6 

7.8 

7.4 
7.0 

7.2 

7.5 
7.6 

7.5 
7.4 

2  

3..    .     . 

4  

5  

Sums 
»«>• 

11.3 

2 

15.1 
3 

27.1 

5 

10.8 
2 

23.0 
3 

21.6 
3 

30.0 

4 

Following  the  calculational  procedure  outlined,  we  obtain: 


—  3 

=  5 


i=      4 

2=      3 

=    22 

=      2 


11.4       SUBSAMPLING 


293 


1=  3 
===  3 


.#22=21.6 


rx=    64.3 

r2  =    74.6 

T=138.9 


and  hence 
]T  F2  =  902.07 
Myy  =  (138.9)2/22  =  876.9641 
,     (74.6)2 


—  876.9641 

—  24.0927 

(27.1)2    t    (10.8)2        (23.0)2        (21.6)3 

.        -          j         —  j _  j _ 


(30.0) 


yy      L    12      '      10 

=  901.0568  -  876.9641 
=  r(11.3)2        (15. 1)2 
yy  ~~  L      2  3 

—  901.0568 

=  901.9036  —  901.0568  =  0.8468 

Svv  =  902.07  —  876.9641  —  24.0927  —  0.8468  =  0.1664. 
These  results  are  presented  in  ANOVA  form  in  Table  11.15. 


TABLE   11.15-ANOVA  for  Fermentation  Data  of  Table  11.14 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean  Square 

Mean  

1 

876.9641 

876.9641 

Microorganisms  .... 

Experimental  error. 
Sampling  error 

1 

5 
15 

24.0927 

0.8468 
0.1664 

24.0927 

0.1694 
0.0111 

2         j-     ™i.              N  2 

17 

Total 

22 

902.0700 

On  examination  of  the  expected  mean  squares  in  Tables  11.13 
and  11.15,  it  is  seen  that  an  exact  test  of  the  hypothesis  ff:rl-==0 
(i=l,  ••-,£)  is  impossible.  This  unfortunate  circumstance  results 
from  the  fact  that  Ci^c2,  and  this  is  so  because  of  the  unequal  num 
bers  of  samples  per  experimental  unit  and  the  unequal  numbers  of  ex 
perimental  units  per  treatment.  This  result  clearly  attests  to  the  desir 
ability  of  equal  numbers  of  observations  in  the  various  subclasses,  and 
for  this  reason  the  statistician  always  recommends  "equal  frequencies" 
when  he  is  consulted  at  the  design  stage  of  any  research  project. 

What,  then,  can  be  done  in  a  situation  such  as  described  above? 
That  is,  since  unequal  frequencies  are  sometimes  inevitable,  is  there 
any  approximate  test  procedure  that  can  be  used?  There  is.  However, 
discussion  of  this  approximation  will  be  deferred  until  Section  11.7. 

Before  terminating  the  present  discussion  of  subsampling  in  a  com 
pletely  randomized  design,  the  simplifications  associated  with  equal 


294  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

frequencies  will  be  demonstrated.  If,  in  a  completely  randomized 
design,  there  are  t  treatments,  n  experimental  units  per  treatment,  and 
m  samples  per  experimental  unit,  the  appropriate  statistical  model  is 

y  ijk  =  p>  +  r±  +  etj  +  -rjijk}          i  =  1,  •  •  •  ,  t  (11  .  27) 

j  =   1,  .  .  .  ,  n 
k  =   1,  •  -  -  ,  m 

where  all  terms  are  defined  as  before.  The  calculations  are  now  speci 
fied  by 

F2  =  total  sum  of  squares 


nm 


*^~^    rr1   /    *.  KJ- 

y         /    -  /  wwi    -^-~     /I// 

X    y       -^     t/     A/'AAZ'  JKt  yyy 

i=l 

experimental  error  sum  of  squares 

t          n 


t—  1    J=l 

f  13  23 

L  i=i  j=i 


sum  of  squares  due  to  the  mean 

(11.29) 
T*/tnm, 

treatment  sum  of  squares 

—  7)2 

(11.30) 


*-     F;)2 

i/m  -  23  r! 

-  Tm,  (11.31) 


and 

51      —  y^F2  Af      T     E  (11*32) 

where 

771 
f?  X"^        V  /-*    1          00\ 

-CSij    — '      X   ^     *  ijki  \LL.3O) 


23  23  Y^  =  23  Js«,  (H.34) 

J-l   *=-!  3=1 


t          n         rn 


11.4       SUBSAMPLING 

t  n 


and 


i  =  Ti/nm, 


Y  =  T/tnm. 


295 

(11.35) 

(11.36) 
(11.37) 

(11.38) 


TABLE   11.16-Generalized  ANOVA  for  a  Completely  Randomized 

Design  with  Subsamplmg  (Equal  Numbers:  Model  I 

and  Model  II) 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean  Square 

IVIean 

1 

Mvv 

M 

Treatments  

t  —  1 

T 

T 

for*  +in<r*  +  nmj*f  i*/(f  -  1) 

Experimenal  error  .  .  . 
Sampling  error 

t(n  -  1) 
tn(m  —  1) 

Eyy 
g 

E 

s 

or 
[a^  +  m<r2  +  nmer^ 
a*  +  mcrz 
<? 

TJ 

Total 

tnm 

y~L  YZ 

These  sums  of  squares  would  then  be  presented  in  ANOVA  form  as  in 
Table  11.16.  Examination  of  the  expected  mean  squares  in  Table  11.16 
indicates  that,  because  of  the  equal  frequencies,  there  will  be  no  diffi 
culty  in  testing  H:n  =  0  (i=l,  -  -  •  ,  t)  or  H:<r*  =  Q.  In  addition,  the 
components  of  variance  are  easily  estimated  by 

s*  =  S  (11.39) 


and 


=  CE  —  S)/m. 


(11.40) 


And,  finally,  the  standard  error  of  a  treatment  mean  is  given  by 


(11.41) 

(NOTE :  Although  not  explicitly  stated,  it  should  be  clear  that  a-2  and 
<rL  could  also  have  been  estimated  when  unequal   frequencies  occur. 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


This  can  be  seen  by  studying  the  expected  mean  squares  in  Table 
11.13.) 

Example  11.6 

An  agronomist  conducted  a  field  trial  to  compare  the  relative  effects 
of  5  particular  fertilizers  on  the  yield  of  Trebi  barley.  Thirty  homo 
geneous  experimental  plots  were  available  and  6  were  assigned  at 
random  to  each  fertilizer  treatment.  At  harvest  time,  3  sample  quadrats 

TABLE   11.17-Coded  Values  of  Yields  from  Ninety  Sample  Quadrats 

Fertilizer  Treatments 


1 

2 

3 

4 

5 

57 

67 

95 

102 

123 

46 

72 

90 

88 

101 

28 

66 

89 

109 

113 

26 

44 

92 

96 

93 

38 

68 

89 

89 

110 

20 

64 

106 

106 

115 

39 

57 

91 

102 

112 

39 

61 

82 

93 

104 

43 

61 

98 

98 

112 

23 

74 

105 

103 

120 

36 

47 

85 

90 

101 

18 

69 

85 

105 

111 

48 

61 

78 

99 

113 

35 

60 

89 

87 

109 

48 

75 

95 

113 

111 

50 

68 

85 

117 

124 

37 

65 

74 

93 

102 

19 

61 

80 

107 

118 

Under  each  treatment,  the  18  observations  are  arranged  in  six  groups  of  three.  Each 
group  consists  of  the  observed  yields  on  the  three  quadrats  taken  from  a  single  experi 
mental  plot. 

were  taken  (at  random)  from  each  experimental  plot  and  the  yield 
was  obtained  for  each  of  the  90  quadrats.  The  data,  in  coded  form,  are 
given  in  Table  11.17.  Using  Equations  (11.28)  through  (11.32),  we 
obtain : 

J^,Y*  =  646,285 
Myy  =  (7187)V°0  =  573,921.88 

TM  =  [(650)2  -f-  (1140)*  +  (1608)2  +  (1797)2  +  (1992)*]/18  -  573,921.88 
=  639,168.72  —  573,921.88  =  65,246,84 


11.4       SUBSAMPLING 


297 


Evv  =  L(131)2  +  -  .  -  +  (344)«]/3  -  639,168.72 

=  641,001.67  -  639,168.72  =  1,832.95 
Suv  =  5,283.33  (by  subtraction). 

These  are  summarized  in  Table  11 .18.  It  is  easily  verified  that  F  =  222.47, 
with  z/i  =  4  and  v^  =  25  degrees  of  freedom,  is  highly  significant,  and  thus  the 
hypothesis  H:r^  =  Q(i  =  l,  -  •  -  ,  5)  is  rejected.  (NOTE:  An  experienced 
analyst  could  probably  have  predicted  this  result  on  examination  of  the 
data,  but  the  analysis  and  the  statistical  test  make  the  conclusion  an 
objective  one  rather  than  a  subjective  one.)  In  case  a  confidence  interval 
estimate  of  a  treatment  mean  is  desired,  the  standard  error  of  a  treat- 


TABLE   11.18-ANOVA  for  data  of  Table  11.17 


Source  of 
Variation 

Degrees 
of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean  Square 

F- 
Ratio 

Mean 

1 

573,921.88 

573,921.88 

Fertilizers.  .  . 

Experimental 
error 

4 

25 

65,246.84 
1,832.95 

16,311.71 
73.32 

4  +  3<r*  +   (18/4)  i:   rf 
I—  1 

c?  _!_  3^-2 

222  .47 

SamplinfiT  error 

60 

5  283  33 

88  O6 

2 

O~-n 

vn 

Total 

90 

646,285.00 

ment  mean  is  calculated.  Its  value  is  -\/E/nm  —  -\/(73.32)/18  = 
=  2.02.  It  is  also  clear  that  components  of  variance  may  be  estimated  in 
a  simple  manner.  For  example,  s%  =  88.06.  However,  when  an  estimate 
of  <r2  is  sought,  the  calculations  yield  s2=  (73.32-88.06) /3  <0.  Since 
cr2,  by  definition,  is  positive,  it  is  unreasonable  to  quote  a  negative 
estimate.  Thus,  in  the  present  situation,  the  "best"  estimate  of  cr2  will 
be  taken  to  be  zero,  even  though  this  is  a  biased  estimate.  More  will 
be  said  about  the  implications  of  this  in  a  later  section.  For  the  moment, 
we  shall  be  content  with  observing  that  apparently  the  variation  among 
the  true  effects  of  different  experimental  units  is  small,  and  thus  the 
researcher  might  consider  less  replication  (fewer  experimental  units  per 
treatment)  in  a  future  experiment  of  this  type. 

The  reader  will,  no  doubt,  have  realized  that  the  concept  of  sub- 
sampling  may  be  extended  to  many  stages.  That  is,  we  can  have 
"samples  within  samples  within  samples  .  .  .  ,  "  and  the  resulting 
ANO  VA  would  reflect  such  multi-stage  subsampling  by  partitioning  the 
total  sum  of  squares  into  many  more  parts.  Rather  than  continue  the 
discussion  in  general  terms,  we  shall  rely  on  problems  at  the  end  of  the 


298  CHAPTER    IT,    COMPLETELY    RANDOMIZED    DESIGN 

chapter  to   illustrate  not   only  the  principles  involved,   but  also  the 
mechanics  of  the  appropriate  calculations. 

11.5  EXPECTED  MEAN  SQUARES,  COMPONENTS  OF 
VARIANCE,  VARIANCES  OF  TREATMENT  MEANS, 
AND  RELATIVE  EFFICIENCV 

In  Sections  11.2  and  11.4,  the  reader  was  introduced  to  the  concepts 
of  components  of  variance,  expected  mean  squares,  and  variances  of 
treatment  means.  In  those  sections,  no  reasons  were  given  as  to  why 
the  expected  mean  squares  contained  the  indicated  components  of 
variance  nor  why  the  coefficients  of  the  components  of  variance  were 
as  given.  We  now  propose  to  remove  this  deficiency.  In  addition,  a 
scheme  will  be  proposed  that  permits  the  estimation  of  the  relative 
efficiency  of  different  proposed  designs  involving  various  degrees  of 
subsampling.  The  discussion  will  be  conducted  with  reference  to  Tables 
11.16  and  11.18. 

Reference  to  Tables  11.16  and  11.18  shows  that  the  expected  mean 
square  for  sampling  error  contains  only  one  component  of  variance. 
This  is  so  because  the  only  factor  which  affects  (or  causes  or  produces) 
the  variation  "among  samples  within  experimental  units"  is  the  77^ 
factor.  However,  the  expected  mean  square  for  experimental  error  con 
tains  two  components  of  variance  since  this  source  of  variation  reflects 
the  variation  among  the  means  of  the  samples  taken  from  each  experi 
mental  unit,  and  these  means  will  vary  not  only  because  of  the  varia 
tion  from  experimental  unit  to  experimental  unit,  but  also  because  of 
the  variation  among  the  samples  taken  from  each  experimental  unit. 
To  discuss  the  expected  mean  square  for  treatments,  it  is  appropriate 
to  consider  first  the  sum  of  squares.  The  treatment  sum  of  squares 
reflects  the  variation  among  the  means  of  all  the  observations  (on 
samples)  recorded  for  each  treatment.  Now,  these  means  will  vary 
because  of  three  contributing  factors:  (1)  variation  among  treatments 
(fertilizers),  (2)  variation  among  experimental  units  (plots)  within 
treatments,  and  (3)  variation  among  samples  (quadrats)  within  experi 
mental  units.  Thus,  the  expected  mean  square  involves  three  compo 
nents  of  variance  if  Model  II  is  assumed,  or  two  components  of  vari 
ance  and  one  sum  of  squares  if  Model  I  is  assumed.  (NOTE:  The 
reader  may  verify  the  reasonableness  of  the  foregoing  remarks  by  sub 
stituting  the  assumed  linear  statistical  model  for  Yi3^  in  the  expres 
sions  for  the  various  sums  of  squares.) 

How  were  the  various  coefficients  in  the  expected  mean  squares  de 
termined?  The  coefficient  of  <r*  is  1  (and  thus  not  shown)  because  this 
reflects  the  variation  among  individual  samples.  The  coefficient  of  <rz 
is  m  (m  =  3  in  Table  11.18)  because  there  were  m  observations  (samples) 
per  experimental  unit.  The  coefficient  of  o>  when  Model  II  is  assumed, 
or  °f  s^t-i  T^/(t — 1)  when  Model  I  is  assumed,  is  nm  because  there 
were  nm  observations  (m  samples  on  each  of  n  experimental  units)  per 
treatment.  In  Table  11.18,  n  —  &  and  m  =  3.  We  might  note  that 
another  way  of  expressing  the  justification  of  the  coefficients  described 


11 ,5       EXPECTED    MEAN    SQUARES  299 

above  is  to  say  that  each  treatment  mean  Is  the  average  of  nm  observa 
tions,  while  each  experimental  unit  mean  is  the  average  of  m  observa 
tions. 

The  estimation  of  the  various  components  of  variance  has  been  well 
illustrated  in  the  preceding  sections.  However,  a  recapitulation  will  be 
made  to  summarize  the  procedure.  Since  S  is  an  unbiased  estimator 
of  o\j,  it  is  reasonable  to  write 

si  =  S.  (11.42) 

Similarly,  E  is  an  unbiased  estimator  of  cr^+mo-2,  and  thus  we  write 

s*  +  ms*  =  E.  (11.43) 

If,  then,  we  combine  Equations  (11.42)  and  (11.43)  as  shown  in  Equa 
tion  (11.44),  an  unbiased  estimator  of  cr*  is  obtained: 


m  in 

Now  that  the  preceding  estimates  are  available,  it  is  possible  to 
determine  (subject  to  sampling  variation,  of  course)  which  factor  is 
contributing  the  most  to  the  observed  variation.  Then,  perhaps,  an 
improvement  can  be  made  in  experimental  technique,  or  the  design 
layout  (configuration)  can  be  changed,  to  better  control  the  variation 
in  future  experiments  of  the  same  type.  To  pursue  this  aspect  of  anal 
ysis,  the  concept  of  "relative  efficiency7'  of  one  design  compared  to 
another  design  of  the  same  type  but  involving  different  numbers  of 
experimental  units  and/or  samples  will  be  investigated. 

Before  such  a  comparison  can  be  made,  a  criterion  for  measuring 
efficiency  must  be  established.  The  criterion  adopted  in  this  book  will 
involve  the  estimated  variance  of  a  treatment  mean.  We  will  say  that 
a  design  which  provides  a  smaller  estimated  variance  of  a  treatment 
mean  than  does  some  other  design  is  the  more  efficient  of  the  two. 

With  reference  to  Table  11.16,  and  in  agreement  with  the  definition 
given  earlier,  the  estimated  variance  of  a  treatment  mean  is 

estimated  variance  of  the  individual  items  contributing 
to  the  mean 

number  of  items  (observations)  averaged  to  get  the  mean 

(11.45) 


nm 

4 


n 


Examination  of  Equation  (11.45)  leads  to  the  following  conclusions: 

(1)   If  the  estimates  of  the  components  of  variance,    s2   and   &*, 
remairt  relatively  constant,  an  increase  in  n  or  m  (or  both) 


3OO  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

will  result  in  a  smaller  estimated  variance  of  a  treatment  mean. 

(2)  An  increase  in  n  (the  number  of  experimental  units  per  treat 
ment)  will  have  more  of  an  effect  than  an  increase  in  m  (the 
number  of  samples  per  experimental  unit)  in  reducing  F(FZ-). 
This  supports  the  statement  made  in  Section   10.16  to  the 
effect  that  "It  (replication)  enables  us  to  obtain  a  more  pre 
cise  estimate  of  the  mean  effect  of  any  factor.  .  ^ .  ^_ 

(3)  If  either  s2  or  s*  (or  both)  can  be  made  smaller,  F(  F*-)  can  be 
made  smaller.  This  could  be  accomplished  by  choosing  more 
homogeneous  experimental  units  or  by  improving  the  experi 
mental  technique. 

Let  us  now  return  to  the  problem  of  estimating  the  efficiency  of  a 
proposed  design  relative  to  the  design  used.  To  do  this,  we  must  first 
estimate  what  the  variance  of  a  treatment  mean  would  be  if  the  pro 
posed  design  were  used.  Assuming  that:  (1)  the  proposed  design  would 
involve  n'  experimental  units  per  treatment  and  m'  samples  per  experi 
mental  unit  and  (2)  the  estimates  of  <r2  and  a*  would  remain  un 
changed,  the  new  estimated  variance  of  a  treatment  mean  would  be 

2      I  /    2 

y-'(F,)  =      *        m/      •  (11.46) 

nm 

If  F'(FT)  <  F(F»),  the  proposed^design  is  said  to  be  more  efficient  than 
the  present  design;  if  ^7(F»)  >  F(Ft-),  the  proposed  design  is  said  to  be 
less  efficient  than  the  present  design.  Thus,  as  a  measure  of  relative 
efficiency i  we  use  the  ratio  of  F(F»)  and  F'(FV).  If  the  efficiency  of  the 
proposed  (new}  design  relative  to  the  present  (old)  design  is  desired,  one 
calculates  (in  per  cent) 


R.E.  of  new  to  old  =  100[F(  7<)/?'CF*)],  (11.47) 

while  if  the  efficiency  of  the  present  (old")  design  relative  to  the  proposed 
(new}  design  is  desired,  one  calculates  (in  per  cent) 

R.E.  of  old  to  new  =  100  [F^F^/t^F,)].  (11.48) 

Some  texts  use  the  concept  of  "relative  information"  and  it  would  be 
wise  for  us  to  see  what  relationship  this  bears  to  relative  efficiency.  If 
information  is  defined  as  the  reciprocal  of  the  variance,  then  it  is  only  a 
matter  of  simple  algebra  to  show  that  relative  information  is  the  same 
as  relative  efficiency.  For  example, 


R.I.  of  old  to  new  =  -y  —  ^  _        X  100  =  R.E.  of  old  to  new.     (11  .  49) 

Li/  ^  \  Y  i)  j 

Similarly, 

R.I.  of  new  to  old  =  R.E!.  of  new  to  old.  (11.50) 


11 .6       F-RAT1OS    THAT    ARE    LESS    THAN    UNITY  3O1 

It  should  be  noted  that  there  are  other  definitions  of  relative  informa 
tion  to  be  found  in  the  literature  (e.g.,  Yates:  Design  and  Analysis  of 
Factonal  Experiments)  which  differ  from  relative  efficiency.  However, 
if  we  define  our  terms  as  above,  the  two  concepts  may  be  used  inter 
changeably. 

Example  11.7 

The  experiment  on  frozen  strawberries  discussed  in  (3)  in  the  first 
paragraph  of  Section  11.4  was  performed.  However,  all  that  is  available 
is  the  abbreviated  AISTOVA  of  Table  11.19.  The  estimates  of  the  com 
ponents  of  variance  are  4  =  5  and  s2  =  (20  —  5) /2  =  7.5  where  the  symbol 
8  is  used  to  denote  determinations  (rather  than  77  to  denote  samples). 
The  estimated  variance  of  a  treatment  mean  is 


10(2) 


5  -f-  2(7.5) 
20 


=  1. 


The  question  is  then  asked,  "Is  the  present  design  more  or  less  efficient 
than  a  similar  design  employing  6  pints  per  storage  time  and  3  determi 
nations  per  pint?"  Calculating 


5-1-3(7.5) 
6(3) 


1.53, 


the  answer  is,  "The  present  design  is  more  efficient  than  the  proposed 
design."  In  fact,  the  efficiency  of  the  present  design  relative  to  the 
proposed  design  is:  R.E.  of  old  to  new  =  100(1.53/1)  =  153  per  cent. 

TABLE    11.19-Abbreviated  ANOVA  of  Ascorbic  Acid 
Content  of  Frozen  Strawberries 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square 

Among  storage  times  

4 

4OO 

100 

<r*  +  2^-*-2°y-     • 

Among  pints  treated  alike  .  . 
Between  determinations  on 
pints  treated  alike.     . 

45 
5O 

9OO 
250 

20 

s                        4  £j    *' 

°* 

F-RATIOS     THAT 


11.6      SOME     REMARKS     CONCERNING 
ARE    LESS  THAN    UNITY 

In  all  the  examples  considered  so  far,  tlie  calculated  F-  values  have 
been  greater  than  unity.  Thus,  in  each  of  these  cases,  the  only  decision 
to  be  made  by  the  analyst  was  whether  the  calculated  value  should  be 
termed  statistically  significant  or  nonsignificant.  If  significant,  the 
hypothesis  H:<n  =  Q  (i=l,  -  •  -  ,  Q  or  H:o*  =  0  was  rejected;  if  not  sig 
nificant,  the  appropriate  hypothesis  was  not  rejected  (perhaps  even 


3O2  CHAPTER    I  1 ,    COMPLETELY    RANDOMIZED    DESIGN 

accepted).  However,  it  is  possible  (and  quite  probable)  that  a  calcu 
lated  F-value  will  turn  out  to  be  less  than  unity.  What  should  our  con 
clusion  be  in  such  a  situation? 

We  can,  of  course,  simply  say  that  F  was  not  significant  and  thus  the 
hypothesis  cannot  be  rejected.  However,  such  an  easy  dismissal  of  the 
question  is  not  wise,  for  it  could  cause  us  to  ignore  a  valuable  warning 
sign.  Suppose,  as  might  happen,  that  F,  with  v\  and  f2  degrees  of 
freedom,  is  so  small  that  Fr  =  l/^P,  with  z>2  and  v±  degrees  of  freedom,  is 
significant.  What  should  our  conclusion  be  in  this  case?  It  appears  as 
though  something  should  be  rejected;  but  what  is  it?  In  this  situation, 
it  seems  reasonable  to  reject  the  postulated  statistical  model. 

If  the  statistical  model  is  rejected  because  of  a  significant  F'  value, 
what  are  the  steps  that  should  then  be  taken?  Some  of  these  are : 

(1)  The  experimental  procedure  should  be  reviewed  to  see  if  the 
various  assumptions  are  satisfied.  For  example,  if  the  proper 
randomization  was  not  employed,  the  validity  of  the  inde 
pendence  assumption  is  doubtful. 

(2)  If   sufficient   observations   are   available,   the   assumption   of 
normality  could  be  checked  by  plotting  the  data  either  on 
regular  graph  paper  or  on  normal  probability  paper. 

(3)  The  assumption  of  homogeneous  variances  might  be  checked, 
but  this  would  require  a  large  number  of  observations  within 
subclasses. 

(4)  The  underlying  phenomenon  should  be  restudied  to  see  if  the 
assumed  linear  model  is  a  good  approximation  to  the  true  state 
of  affairs.  If,  as  a  result,  the  assumed  model  is  rejected,  a 
search  should  be  made  for  a  new  model  which  better  describes 
the  observed  data  and  the  phenomenon  under  investigation. 

11.7   SATTERTHWAITE'S   APPROXIMATE   TEST   PRO- 
CEDURE 

When  discussing  the  analysis  of  a  completely  randomized  design 
involving  subsampling,  it  was  noted  that  no  exact  test  of 
J2r:ri  =  0(i=5=l,  •  *  •  ,  f)  was  possible  when  the  experiment  involved 
unequal  frequencies  at  the  various  stages  of  subsampling.  At  that  time, 
it  was  promised  that  an  approximate  test  procedure  would  be  explained 
later.  We  are  now  ready  to  fulfill  that  promise. 

The  proposed  approximation,  due  to  Satterthwaite  (29),  proceeds  as 
follows :  Using  estimates  of  the  components  of  variance,  mean  squares  will 
be  synthesized  which  will  have  the  same  expected  value  if  the  hypothesis 
to  be  tested  is  true.  These  synthetic  mean  squares  will  then  be  used  to  form 
a  ratio  which  is  approximately  distributed  as  F. 

How  are  the  synthetic  mean  squares  formed?  If  we  denote  the  actual 
mean  squares  existing  in  an  ANOVA  by  MSi,  MS%,  •  •  •  ,  MSk,  then  a 
synthetic  mean  square  may  be  obtained  by  forming  a  linear  combina 
tion  such  as 

L  =  aiMSi  +  a2MS2  +  -  -  -  4-  akMSk  (11.51) 


11.  8       SELECTED    TREATMENT    COMPARISONS  3O3 

where  the  at-  are  constants.  The  degrees  of  freedom  associated  with  L 
are  then  estimated  by 

.  .  .  * 

*/vh 

where,  of  course,  v±  represents  the  degrees  of  freedom  associated  with 
MSi(i—~L,  •  -  *  ,  fc).  Sometimes  both  the  numerator  and  denominator 
mean  squares  (in  the  approximate  /^-ratio)  will  be  synthesized.  How 
ever,  it  is  more  likely  that  only  one  synthetic  mean  square  will  be  used 
in  any  given  situation. 

Because  of  the  lack  of  uniqueness  of  the  approximate  .F-ratio  (dif 
ferent  ^-ratios  could  result  from  the  use  of  different  synthetic  mean 
squares)  and  because  of  the  necessity  of  approximating  the  degrees  of 
freedom,  the  procedure  is  of  limited  usefulness.  However,  if  used  with 
care,  it  can  be  of  value  to  the  researcher  and/or  statistician.  The  reader 
is  referred  to  Cochran  (10)  for  a  further  discussion  of  this  problem. 

Example  11.8 

Referring  to  Example  11.5,  we  recall  that  an  exact  test  of  H":ri  =  r2  =  0 
was  impossible.  This  was  so  because  c\  ?^C2  in  Table  11.15.  It  is  decided 
to  form  a  "synthetic  experimental  error  mean  square"  that  will  have  an 
expected  value  of  cr^  +  czo-*.  This  could  be  done  by  calculating 

-j—  #2«S> 

[i  - 


O  (0.1694)  +  [1  - 

The  approximate  jFVratio  would  then  be  F  =  24.0927/Z/  with  degrees  of 
freedom  v±  =  1  and  v<2.  =  v,  where 


[ai(0.1694)]V5  +  [a2(0.0111)]2/15 

The  details  of  the  numerical  calculations  are  left  as  an  exercise  for  the 
reader, 

11.8      SELECTED  TREATMENT  COMPARISONS:  GENERAL 
DISCUSSION 

In  Section  10.15,  the  idea  of  making  specific  comparisons  among 
treatment  means  was  introduced.  At  that  time,  also,  the  concept  of  an 
orthogonal  contrast  was  presented,  and  it  was  suggested  that  orthogo 
nal  contrasts  were  to  be  preferred  over  nonorthogonal  contrasts.  How 
ever,  the  researcher  was  warned  not  to  let  the  statistician's  desire  for 
orthogonality  override  his  (the  researcher's)  needs. 

In  this  section  some  general  comparisons  among  treatments  will  be 
examined,  not  to  illustrate  the  concept  of  a  contrast,  but  to  demon 
strate  the  manner  in  which  the  ANOVA  is  modified  to  provide  the 
proper  analysis.  Because  of  the  infinitely  many  possibilities,  this  will 
best  bejione  by  discussing  a  few  illustrative  cases. 


3O4 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


For  example,  consider  an  experiment  involving  t  treatments  and  n 
experimental  units  per  treatment  in  which  no  subsampling  occurred. 
If  treatment  No.  1  were  a  "control"  treatment,  it  would  be  of  interest 
to  make  tlie  following  specific  comparisons  among  the  treatments: 
(1)  treatment  No.  1  versus  the  rest  and  (2)  among  the  rest.  The  sums  of 
squares  for  these  two  comparisons  would  be  determined  as  follows: 


-   T*/tn 


SS(l  versus  rest) 

\Tlfn  +  (r2  +  -  - 
xSVSXamong  the  rest) 


These  results,  when  coupled  with  the  basic  ANOVA,  "would  be  pre 
sented  as  in  Table  1 1 .20,  where  the  sums  of  squares  (degrees  of  freedom) 
for  the  selected  comparisons  are  offset  to  indicate  that  they  are  portions 
of  the  treatment  sum  of  squares  (degrees  of  freedom).  (NOTE:  In  this 
example,  the  sums  of  squares  for  the  two  comparisons  add  up  to  the 
treatment  sum  of  squares.  The  reader  is  "warned  that  this  will  not 
always  be  the  case.) 

TABLE   11.20-Generalized  ANOVA  Showing  Two  Selected  Treatment 

Comparisons 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

Mean  

1 

Jl4T,_., 

M 

Treatments  

t—  1 

T 

T 

T/E 

1  vs.    rest  

1 

(CV)™ 

Ci 

Ci/JB 

Among  the  rest.  .,.,,... 

t  —  2 

(  O2  lint 

C2 

Co/E 

Experimental  error  

t(n—l) 

Ew 

E 

Total 

in 

51  ^2 

A  second  illustration  based  on  the  same  type  of  design  would  be  the 
case  in  which  the  t  treatments  segregate  into  k  groups  containing 
t\,  t%,  -  •  *  ,  tk  treatments,  respectively,  where 


In  such  a  case,  the  natural  comparisons  would  be:  (1)  among  groups 
and  (2)  among  treatments  within  the  ith  group;  i=  1,  -  -  •,  fc.  The  sum 
of  squares  for  the  first  of  these  fc  +  1  comparisons  would  be  calctdated 
as  f  olio  ws  : 


Gyy  = 


groups)   = 


—  T*/tn. 


11.8      SELECTED   TREATMENT   COMPARISONS  3O5 

The  sum  of  squares  among  treatments  in  the  first  group  is  given  by 


The  sums  of  squares  among  treatments  in  each  of  the  remaining  /b—  1 
groups  would  be  found  in  a  similar  manner.  The  results  would  then  be 
presented  in  ANOVA  form  as  in  Table  11.21.  (NOTE:  Once  again  the 

TABLE  11.21-Generalized  ANOVA  Showing  k+1 
Selected  Treatment  Comparisons 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

1 

-tkt  r/7/ 

AT 

^lean  
Treatments            

t—1 

-t  j/j/ 

r 

T/E 

Among  groups      

k  —  1 

Gyy 

G 

G/E 

'Within  group  1       .... 

h—l 

(WOw 

Wi 

W,/E 

^^ithin  group  2    

22—  1 

(TiT2)w 

W* 

Wz/E 

\jyithin  group  k       .... 

te—  1 

(TF*)w 

wk 

Wk/E 

j  f  —      i  > 

77 

E 

Experimental  error  

t(n—i) 

-&yv 

»~n_  •*-«  1 

•f/tsr 

y^  y2 

lotal 

tn 

^1^  ^ 

sums  of  squares  for  the  various  comparisons  add  up  to  the  treatment 

sum  of  squares.)  ,  .    . 

One  more  general  illustration  will  be  given.  In  this  instance,  assume 
(again)  that  one  treatment  is  a  "control."  However,  the  researcher 
wishes  to  do  more  than  compare:  (1)  control  versus  rest  and  (2)  among 
the  rest  He  also  wishes  to  compare,  separately,  each  noncontrol  treat 
ment  versus  the  control.  Thus,  in  addition  to  the  sums  of  squares  indi 
cated  in  the  first  illustration,  he  would  also  compute: 

(C,)w  =  -55(1  vs  2)  =  (Tl 
s  3)  =  (r* 


vs  0  =  (Tl  +  Tt}/n  -  (Tx  +  Tt)*/2n. 
These  results  would  then  be  presented  as  in  Table  11.22.  (NOTE  :  This 


306 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


TABLE   11.22-Generalized  ANOVA  Showing  t+1 
Selected  Treatment  Comparisons 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

Mean  

1 

Mm/ 

M 

Treatments  

t—1 

T 

T 

T/E 

Control  vs.  rest    .     .  . 

1 

(C^yy 

Ci 

Ci/jE 

Among  rest      

t  —  2 

\\-si&)yy 

C2 

C2/£ 

Control  vs.  2  

1 

\\^%)yy 

C3 

CS/JE 

Control  vs.  3  

1 

(CO,*, 

C4 

c4/£ 

Control  vs    t   . 

1 

\\-"t    I  .1/7/rr 

v£  1  1 

Cn-i/E 

T^XTDCr  intent"  ?1  prror 

t(n^\\ 

Total 

tn 

T!  F2 

time  neither  the  degrees  of  freedom  nor  the  sums  of  squares  for  the 
comparisons  will  add  up  to  the  treatment  sum  of  squares.) 

Example  11  .9 

The  experiment  described  in  Example  10.10  was  performed  and  we 
wish  to  investigate  the  specified  comparisons.  Assuming  that  the  data 
given  in  Table  7.20  were  the  results  of  this  experiment,  it  iib  seen  that: 

(C^Vy  =  [(184  +  68)2/6  +  (170  -h  378)  V14]  -  (800)  2/20 

=  34.3 

(C^yy  -  [(184)V4  +  (68)  V2]  -  (252)  V6  -  192.0 
(C£yy  =  [(170)2/5  +  (378)2/9]  -  (548)2/14  =  205.7. 

Combining  these  figures  with  those  of  Table  7.21,  we  get  Table  11.23. 
Examination  of  the  F-  values  in  Table  11.23  indicates  that  all  the  treat 
ments  (electrolytes)  differ  significantly  in  their  effects  on  the  charac 
teristic  (of  the  batteries)  being  studied.  (NOTE:  See  Problein  11.30  for 
the  expected  mean  squares.) 

11.9      SELECTED  TREATMENT  COMPARISONS:  ORTHOG 
ONAL  AND   NONORTHOGONAL  CONTRASTS 

Having  spent  considerable  time  discussing  treatment  comparisons  in 
general,  let  us  now  concentrate  on  the  subject  of  contrasts,  and  par 
ticularly  on  orthogonal  contrasts. 

It  may  be  verified  that  the  sum  of  squares  associated  with  a  particu 
lar  contrast  is  given  by 


=  c 

^j 


1=1 


-( 


(11.53) 


11.9       ORTHOGONAL   AND    NONORTHOGONAL    CONTRASTS 


307 


where  all  symbols  except  t  are  defined  as  in  Section  10.15.  The  symbol 
t  is  used  here,  rather  than  k  as  in  Section  10.15,  to  conform  to  the  nota 
tion  being  used  in  the  present  chapter.  If  each  treatment  total 

TABLE   11.23-ANOVA  for  Experiment  of  Example  11.9 

(Data  in  Table  7.20)  Showing  the  Analysis  of  a 

Specified  Set  of  Treatment  Comparisons 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

IVIean 

1 

32,000 

32,000 

Treatments    (electro 
lytes)      

3 

432 

144 

72 

Ci   (1    and   2   vs.   3 
and  4)  

1 

34.3 

34.3 

17.15 

C2  (1  vs.  2) 

1 

192.0 

192.0 

96 

C3  (3  vs.  4)        

1 

205.7 

205.7 

102.85 

1  ft 

32 

2 

Experimental  error  .  .  . 

nrv-vfoi 

20 

32  464 

is  the  sum  of  the  same  number  of  observations  (that  is,  if 
i=l,   -  -  -  ,  0>  Equation  (11.53)  simplifies  to 


n 


=  n  for 


(11.54) 


n 


The  results  would  then  be  presented  in  an  AN  OVA  in  agreement  with 
the  format  adopted  in  the  preceding  section.  [NOTE:  If  a  set  pf^— 1 
orthogonal  contrasts  among  t  treatments  is  investigated,  the  individual 
sums  of  squares  (one  for  each  contrast)  will  add  up  to  the  treatment 
sum  of  squares.  ] 

Example  11.10 

Consider  again  the  experiment  described  in  Example  10.10  and 
analyzed  in  Example  11.9.  The  sums  of  squares  associated  with  the 
three  contrasts  could  also  have  been  calculated  as  follows: 

[(7)  (184)  +  (7)  (68)  +  (-3)  (170)  +  (-3)(378)]' 
[4(7)2  +  2(7)2  +  5(-3)2  +  9(-3)»] 

[(1)(184)  +  (-2)  (68)  +  (0X170)  +  (0)(378)]2 
+  2(-2)»  +  5(0)2  +  9(0)2] 
(0)(68)  +  (9)  (170)  +  (-5)(378)]2 


(CO* 


(COw  j-4(0)2  +  2(Q)2  +  5(9)2  +  9(_5)*] 

The  ANOVA  will,  of  course,  be  the  same  as  in  Table  11.23. 


308 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


Example  11.11 

The  experiment  described  in  Example  10.11  was  performed,  and^the 
data  shown  in  Table  11.24  were  recorded.  The  appropriate  calculations 


are: 


^F2  =  32,378 

Myv  =  (800)2/20  =  32,000 

Tyy  =  [(ISO)2  4-  (160)2  4-  (160)2  4-  (164)2  4-  (136)2]/4  -  32,000 

EW  =  32,378  —  32,000  —  248  =  130 

[(-1)(180)  +  (4)  (160)  +  (-1X160)  +  (-1X164)  +  (-1)(136)]2 


248 


(CO 


4[(-l)2  +  (4)2  +  (-1)2  +  (-1)2  +  (-1)2] 
(Q)(16Q)  +  (1)(160)  +  (-1X164)  +  (- 


4[(1)2  +  (O)2  +  (I)2  +  (-1)2  •+•  (-1)2] 
+  (0)(16Q)  +  (-1)(16Q)  +  (0)(164)  +  (0)(136)]2 


4[(1)24-  (0)24-  (~1)24-  (O)2 
i-  (0)(160)  4-  (0)(160)  4 


(O)2] 


=  100 


50 


98 


^    *)y"  4[(0)2  +  (O)2  +  (O)2  4-  (I)2  +  (~1)2] 

These  results  are  then  summarized  as  in  Table  11. 25.  Using  a:  =  0.05, 
all  contrasts  except  Ci  are  judged  to  be  statistically  significant.  (NOTE: 
See  Problem  11.31  for  the  expected  mean  squares.) 

TABLE   11.24r-Data  From  Experiment  Described  in  Example  10.11 
and  Discussed  in  Example  11.11 

Electrolytes 


1 

2 

3 

4 

5 

40 

38 

44 

41 

34 

45 

40 

42 

43 

35 

46 

38 

40 

40 

34 

49 

44 

34 

40 

33 

TABLE   11. 25- AN  OVA  for  Experiment  Described  in  Example  10.11 
(Data  in  Table  11,24;  Discussion  in  Example  11.11) 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

TV/T^ko-n 

1 
4 
1 
1 
1 
1 
15 

32,000 
248 
0 
100 
50 
98 
130 

32,000 
62 
0 
100 
50 
98 
8.67 

Elcr^T^lytes         .  ^    . 

7.15 
0 
11.53 
5.77 
11.30 

Ci                                

Co                     

Cs                

CA                                     

L*  •>  f\<a.T*1  TY1  d^T"!  "f~  £t  1     f^TTVIT 

Total 

20 

32,378 

IK9       ORTHOGONAL   AND    NONORTHOGONAL    CONTRASTS  309 

Up  to  this  point,  ttie  discussion  of  contrasts  has  centered  on:  (1) 
ANOVA  techniques  for  isolating  the  sums  of  squares  associated  with 
each  contrast  and  (2)  the  use  of  the  corresponding  mean  squares  to  test 
the  hypothesis  that  the  true  effects  estimated  by  the  contrasts  are  0. 
However,  the  problem  of  estimation  should  not  be  overlooked. 

If  the  true  effect  estimated  by  a  contrast  C3-  is  denoted  by  the 
symbol  <f>j,  it  is  desirable  to  construct  a  confidence  interval  estimate  of 
0y.  That  is,  two  numbers,  L  and  U,  are  sought  such  that  we  can  be 
100-y  per  cent  confident  that  <f>  will  be  between  L  and  U.  To  determine 
Lf  and  U,  the  standard  error  of  a  contrast  is  needed.  Defining  the  esti 
mated  variance  of  a  contrast  by 

=  v (  2:  CV 


the  standard  error^of  a  contrast  is  given  by  VF(C/). 

The  nature  of  V(Tf)  will,  of  course,  depend  on  whatever  assumptions 
are  made  concerning  the  observations.  If  we  are  dealing  with  a  com 
pletely  randomized  design  involving  t  treatments  and  n  experimental 
units  per  treatment  in  which  no  subsampling  has  been  performed  and 
if  the  usual  assumptions  (see  Section  11.2)  have  been  made,  then 


;?y  (11.56) 

and 

r\  

l    s~~t    ~p-    ,  \'\S "&(("*  \  ^1  1     ^7^ 

U) 

where  v  stands  for  the  number  of  degrees  of  freedom  associated  with  s2 
in  Equation  (11.56). 

Example  11.12 

Consider  the  experiment  discussed  in  Examples  10.11  and  11.11.  The 
data  were  presented  in  Table  11.24  and  the  ANOVA  in  Table  11.25. 
For  this  case,  we  have: 

F(Ci)  =  4(8.67)[(-l)2  +  (4)2  -h  (-1)2  +  (-1)2  +  (-1)2] 
F(C2)  =  4(8.67)  [(I)2  +  (O)2  +  (I)2  +  (-1)2  +  C-l)2] 
F(C3)  =  4(8.67)[(1)2  +  (O)2  +  (-1)2  +  (O)2  +  (O)2] 
F(C4)  =  4(8.67)[(0)2  +  (O)2  +  (O)2  +  (I)2  +  (-1)']. 


310  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

Since  s2==8.67  had  15  degrees  of  freedom,  confidence  intervals  for 
4>i(z  =  l,  2,  3,  4)  may  easily  be  constructed  using  Equations  (10.23) 
and  (11.57). 

11.10     ALL    POSSIBLE    COMPARISONS    AMONG    TREAT 
MENT  MEANS 

In  Sections  11.8  and  11.9,  the  usual  method  of  analyzing  comparisons 
among  treatment  means  was  discussed  in  considerable  detail.  How 
ever,  one  very  important  (statistical)  restriction  on  the  use  of  the  de 
scribed  method  was  not  mentioned.  This  restriction  is  as  folio  TVS:  The 
comparisons  to  be  studied  should  be  selected  in  advance  of  any  analysis  of 
the  data.  That  is,  the  method  of  analyzing  contrasts  described  in  the 
preceding  sections  would  not,  in  general,  be  valid  if  the  comparisons 
were  decided  upon  after  a  perusal  of  the  data  and  (perhaps)  a  pre 
liminary  ANOVA.  In  other  words,  the  comparisons  should  have  been 
decided  upon  during  the  planning  stage. 

The  restriction  stated  in  the  preceding  paragraph  can,  however,  work 
a  hardship  on  the  researcher.  Much  experimentation  is  of  a  purely 
exploratory  nature  and  little,  if  any,  idea  of  which  comparisons  might 
be  of  interest  is  available  prior  to  the  collection  and  analysis  of  the 
data.  In  such  cases,  the  researcher  would  like  to  gain  more  from  the 
analysis  than  a  simple  statement  that  the  treatment  means  are,  or  are 
not,  statistically  significant.  He  would  also  like  to  know,  for  example, 
if  some  of  the  treatments  might  be  considered  equivalent  and  which 
treatment  is  "best." 

How  can  the  researcher  attain  the  goals  stated  in  the  preceding 
paragraph?  This  problem  has  received  much  attention  from  statisti 
cians  in  recent  years,  and  some  of  those  who  have  made  contributions 
in  the  area  are:  Bechhofer  (4),  Duncan  (15,  16),  Dunnett  (17),  Hartley 
(22),  Keuls  (24),  Kramer  (25,  26),  Newman  (27),  Scheffe  (30),  and 
Tukey  (33,  34,  35,  36).  Incidentally,  the  methods  of  Duncan,  Scheff6, 
and  Tukey  (the  major  protagonists)  are  discussed  in  detail  in  Federer 
(19),  and  numerical  illustrations  are  given  for  each  method. 

Before  proceeding  to  discuss  the  method  which  I  favor,  time  will  be 
taken  to  mention  an  associated  technique  which  has  been  widely  used 
by  researchers  for  many  years.  This  technique  involves  what  is  known 
as  a  least  significant  difference  or  LSD,  which  is  defined  by 


LSD  = 


where  v  represents^the  degrees  of  freedom  associated  with  the  variance 
estimate  used  in  F(Fi—  Fy).  The  LSD  technique  operates  as  follows: 
If  the  absolute  value  of  the  difference  between  any  two  treatment 
means  exceeds  the  LSD,  the  effects  of  the  two  treatments  are  judged 
to  be  significantly  different  ;  if  the  absolute  value  of  the  difference  does 


11. TO       COMPARISONS    AMONG    TREATMENT   MEANS  311 

not  exceed  the  LSD,  no  such  conclusion  is  reached.  The  reader  is 
warned  that  indiscriminate  use  of  the  LSD  technique  is  dangerous  for, 
if  we  have  enough  treatments,  the  probability  is  high  that  at  least  one 
of  the  t(t— 1)/2  differences  will,  due  to  chance  alone,  be  judged  sig 
nificantly  different.  Thus,  the  use  of  the  LSD  is  to  be  discouraged. 
(NOTE:  When  Z  =  2,  the  LSD  is  a  legitimate,  but  redundant,  device.) 

The  method  to  be  used  in  this  book  for  making  (when  desirable)  all 
possible  comparisons  among  treatment  means  is  that  proposed  by 
Scheff6  (30).  While  this  method  has  not  been  the  one  most  widely 
adopted,  it  does  have  certain  advantages.  These  advantages  are: 
(1)  it  is  closely  related  to  the  concept  of  a  contrast,  (2)  it  uses  tables 
that  are  widely  available  (viz.,  .F-tables),  and  (3)  it  is  easy  to  use.  Let 
us  now  see  how  the  technique  works. 

Recalling  that  a  contrast  is  defined  by 


(11.59) 
the  procedure  is  to  calculate 

'2  (11.60) 


where 

A*  =  (t  -  l^ci^oc,,.,,),  (11.61) 


^(Ti),  (11.62) 

vi  =  t  —  1,  (11.63) 

and  ?2  stands  for  the  degrees  of  freedom  associated  with  the  denomi 
nator  mean  square  used  in  the  7^-test  of  H\T±~T<L—  •  •  *  =T*.  Then,  if 
\Cj\  >A[F(C/)]1/2,  the  hypothesis  H:<pj  =  0  will  be  rejected.  (See 
Section  11.9  for  the  definition  of  <£/.)  That  is,  if  the  absolute  value  of 
Cj  exceeds  A[F(C/)]1/2,  the  contrast  <7/  will  be  said  to  differ  signifi 
cantly  from  0.  [NOTE:  The  original  F-test  rejects  -ff:r;  =  0 
(i=  1,  -  -  -  ,  f)  if  and  only  if  at  least  one  <7/  is  significantly  different,  by 
Scheff^'s  techjiique,  from  0.  The  application  of  Scheffe's  procedure  per 
mits  us,  then,  to  determine  which  of  the  C/  are  significant.  ] 

Example  11.13 

Consider  the  experiment  described  in  Examples  10.10  and  11.9.  The 
data  were  presented  in  Table  7.20  and  analyses  in  Tables  7.21  and  11.23. 
The  value  of  A  to  be  used  in  making  any  desired  comparison  is  found 
to  be  3.98  since 

A*  =  0  -  l)/^!-*)^.^)  -  (3)F.99ca.i6)  =  (3) (5.29)  -  15.87. 


312  CHAPTER    11  ,    COMPLETELY    RANDOMIZED    DESIGN 

Let  us  examine  contrast  C%  described  in  Example  10.10,  namely, 

c2  =  (I)T!  +  (-2)ra  +  (0)T3  +  (O)r4  =  (i)(i84)  +  (-2)  (as)  -  48. 

The  estimated  variance  of  C%  is  given  by 

•p-(C2)  =  (l)2(4s2)  +  (-2)*(2*2)  =  12^2=  12(2)  =24. 
Therefore, 

2  =  (3.98)  V24  =  (3.98)  (4,899)  «  19.498. 


Since  j  Cy|  =  48  >  19.498,  we  conclude  that  the  difference  between  the 
effects  of  treatment  No.  1  and  treatment  No.  2  is  statistically  signifi 
cant.  Incidentally,  this  agrees  with  the  conclusion  reached  in  Example 
11.9.  Other  comparisons  among  the  treatment  effects  could  be  made  in 
a  like  manner. 

Example  11.14 

Consider  the  experiment  described  in  Examples  10.11  and  11.11.  The 
data  were  presented  in.  Table  11.24  and  the  analysis  in  Table  11.25.  In 
this  illustration,  A2  =  4(4.89)  =  19.56  and  thus  A  =4.42.  If  we  are  inter 
ested  in  Cz  as  defined  in  Table  10.3,  it  may  be  verified  that  (72  =  40, 
F(C2)  =  16s2  -16(8.67)  =138.72,  [F(Cy  ]1/2  =  1  1.78,  and  -A[F(C2)]1/2 
=  52.07.  Since  |  C*\  =40  <A  [F(C2)  ]1/2  =  52.07,  we  conclude  that  C2  is 
not  significantly  different  from  0. 

It  is  noted,  however,  that  this  conclusion  is  the  opposite  of  that 
reached  in  Example  11.11.  Why  is  this?  The  reason  may  be  explained 
as  follows:  Scheff^'s  method  will  not  lead  to  significant  results  (if  the 
appropriate  null  hypothesis  is  true)  as  frequently  as  will  the  classical 
approach  of  orthogonal  comparisons  because  we  have  been  permitted 
to  examine  the  data  before  deciding  on  the  analysis.  This,  obviously, 
should  lead  to  fewer  cases  of  claiming  significance  when  no  real  dif 
ferences  exist.  This  is  as  it  should  be,  for,  if  we  can  look  at  the  data 
before  deciding  on  the  comparisons  to  be  investigated,  we  should  be 
able  to  lessen  our  chances  of  making  errors.  From  the  point  of  view  of 
estimation,  this  decrease  in  the  "frequency  of  errors"  takes  the  form  of 
longer  confidence  intervals  (i.e.,  our  estimates  are  less  precise)  than 
those  provided  by  the  classical  approach. 

11.11  RESPONSE  CURVES:  A  REGRESSION  ANALYSIS  OF 
TREATMENT  MEANS  WHEN  THE  VARIOUS  TREAT 
MENTS  ARE  DIFFERENT  LEVELS  OF  ONE  QUANTI 
TATIVE  FACTOR 

The  reader  may  be  wondering  why  the  subject  of  this  section  is 
under  discussion  at  this  time.  Did  we  not  discuss  regression  analyses 
completely  enough  in  Chapter  8?  Of  course  we  did,  but  now  we  wish 
to  utilize  the  techniques  of  regression  to  make  more  complete  and  in 
formative  analyses  of  data  arising  from  completely  randomized  designs 
in  which  the  treatments  are  different  levels  of  a  single  quantitative 
factor. 


11.11       RESPONSE    CURVES  313 

How  is  this  possible?  Let  us  suppose  th^at  the  treatments  being 
examined  are:  (1)  different  levels  (or  rates)  of  application  of  the  same 
fertilizer,  (2)  different  weights  of  an  object  being  moved  in  a  time-and- 
motion  study  project,  or  (3)  different  intensities  of  a  given  stimulus  in 
a  psychological  experiment.  If  situations  such  as  these  arise,  it  seems 
reasonable  to  investigate  how  the  measured  characteristic  varies  with 
changes  in  the  level  of  the  treatment.  That  is,  we  would  like  to  know 
if  the  change  in  the  measured  characteristic  takes  place  in  a  linear, 
quadratic,  ,  .  .  fashion  as  the  level  of  the  treatment  is  increased  or 
decreased.  In  other  words,  we  wish  to  gain  some  idea  of  the  shape  of 
the  response  curve  so  that  an  estimate  may  be  made  of  the  optimum 
level  of  the  treatment. 

Just  how  will  the  type  of  analysis  indicated  above  be  carried  out? 
The  first  step  is  to  plot  the  treatment  means,  thus  gaining  some  idea 
as  to  the  general  shape  of  the  response  curve.  Once  this  has  been  done, 
the  researcher  will  be  ready  to  undertake  a  more  rigorous  analysis  of 
his  sample  data. 

Equations  for  various  possible  response  curves  could,  of  course,  be 
determined  using  the  techniques  of  Chapter  8.  However,  the  deter 
mination  of  the  equation  of  the  response  curve  is  not  the  immediate 
aim  of  our  analysis.  The  immediate  aim  is  to  reach  an  objective  de 
cision  (based  on  more  than  a  simple  plotting  of  the  means)  as  to  the 
nature  of  the  regression  function  that  will  best  describe  the  effect  of 
the  treatment  on  the  response  variable. 

Perhaps  the  most  convenient  way  of  reaching  the  goal  stated  in  the 
preceding  paragraph  is  to  determine  how  much  of  the  treatment  sum 
of  squares  would  be  associated  with  each  of  the  terms  (linear,  quad 
ratic,  .  .  .  )  in  a  polynomial  regression.  If  the  various  levels  of  the 
treatment  being  studied  are  equally  spaced,  this  analysis  can  best  be 
carried  out  using  the  method  of  orthogonal  polynomials  introduced  in 
Section  8.20.  (NOTE:  The  assumption  of  equal  spacing  will,  in  general, 
present  no  problem,  for  both  the  researcher  and  the  statistician  will 
ordinarily  plan  the  experiment  in  such  a  way  as  to  insure  that  the 
assumption  will  be  satisfied.  That  is,  in  most  applications  equal  spacing 
is  the  usual  state  of  affairs.) 

If  each  treatment  total  (T*)  is  the  sum  of  n  observations,  the  desired 
sums  of  squares  are  found  using 


(t 
g^ 

due  to  the  kth  degree  term  =  -  ;  (11  .  64) 


n 
,-••,£  —  1; 


where  the  £**  are  orthogonal  polynomial  coefficients.  Extensive  tables 
of    orthogonal    polynomial    coefficients    are    given    in    Anderson    and 


314  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

TABLE   11.26-Partial  Table  of  Orthogonal  Polynomial  Coefficients 


*-2 

t  =  3 

*-4 

t  =  5 

i 

k=l 

k=l  k  =  2 

*  =  1  ^  =  2 

k  =  3 

k=l  k  =  2  k  =  3 

£  =  4 

1  

—  1 

—  1   +1 

—  3   +1 

_1 

—  2   +2   —1 

+  1 

2, 

0   —2 

•£   -^ 

+3 

—1   —  1   +2 

—  4 

3  

+  1   +1 

+  1   —1 

—  3 

0   —2    0 

+  6 

4 

+3   +1 

+  1 

+1   —1   —2 

—  4 

5  

_|_2   +2   +1 

Houseman  (1) ;  for  your  convenience  an  abbreviated  tabulation  is 
provided  in  Table  11.26.  In  agreement  with  the  notation  previously 
adopted,  the  sums  of  squares  associated  with  the  linear,  quadratic, 
cubic,  .  ,  .  terms  will  be  denoted  by  (T  £)yy,  (TQ}yy,  (Tc)vy,  -  -  -  .  In 
addition,  since  it  is  unlikely  that  the  researcher  will  wish  to  isolate 
more  than  a  few  terms  when  studying  the  treatment  sum  of  squares, 
the  balance  (if  any)  will  be  represented  by  (T^^yy.  For  example,  if  the 
linear,  quadratic,  and  cubic  effects  were  isolated,  the  sum  of  the  squares 
of  the  deviations  from  regression  would  be  given  by 


—   Tyy 


(11.65) 


The  results  of  the  foregoing  calculations  may  then  be  summarized  as  in 
Table  11.27. 

Example  11 .15 

Consider  the  data  in  Table  11.28.  Although  an  examination  of  the 
treatment  totals  suggests  that  a  linear  response  function  may  be  ap 
propriate,  the  quadratic  effect  will  also  be  isolated  for  illustrative  pur- 

TABLE   11.27-Generalized  ANOVA  For  a  Completely  Randomized 

Design  Showing  the  Isolation  of  the  Linear,  Quadratic,  and  Cubic 

Components  of  the  Treatment  Sum  of  Squares 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

Mean  

1 

MW 

M 

Treatments     

t  —  1 

T 

T 

T/E 

TL  

1 

(  T  T  Iini 

TL 

TL/E 

To  

1 

(To}w 

TQ 

To/E 

Tc    

1 

(Tc)™ 

Tc 

Tc/E 

jTx>etv  

t  —  4 

Tritr-n 

Tr>ev/E 

Experimental  error.  . 

t(n—  1) 

J—JtrtJ 

E 

Total 

tn 

T:  Y* 

11.11       RESPONSE    CURVES 


315 


TABLE   11. 28- Yields  (Converted  to  Bushels/Acre)  of  a  Certain  Grain 

Crop  in  a  Fertilizer  Trial 


Level  of  Fertilizer 

No 

Treatment 

10  Ibs. 
per  Plot 

20  Ibs. 
per  Plot 

30  Ibs. 
per  Plot 

40  Ibs. 
per  Plot 

20 
25 
23 
27 
19 

25 
29 
31 
30 
27 

36 
37 
29 
40 
33 

35 
39 
31 
42 
44 

43 
40 
36 

48 

47 

Totals 
Means 

114 
22.8 

142 
28.4 

175 
35 

191 
38.2 

214 

42.8 

poses.  The  following  sums  of  squares  were  obtained: 
52  F2  =  29,560 
Myy  -  (836)  »/25  =  27,955.84 
TW  =  [(H4)2  +  (142)2  +  (175)2  +  (191)2  +  (214)2]/5  -  27,955.84 

=  1256.56 
Eyy  -  29,560  -  27,955.84  -  1256.56  «  347.60 

h  (0)(175)  +  (1)(191)  +  (2)(214)]2 


5[(-2)«  +  (-1)2  +  (O)2  +  (I)2  +  (2)2] 
(249) 2 


50 


1240.02 


[(2)  (114)  +  (-1X142)  +  (-2)(175)  +  (-1X191)  +  (2)(214)]2 


(-27)2 
70 


5[(2) 


10.41 


(-2) 


(2)2] 


w  =  1256.56  -  1240.02  -  10.41  =  6.13. 
These  are  summarized  in  Table   11.29.  Examination  of  the  F-ratios 

TABLE  11.29-ANOVA  for  Data  of  Table  11.28 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 

Squares 

Mean 
Square 

.F-Ratio 

!MLean                      .  .  . 

1 

27,955.84 

27,955.84 

Fertilizer  levels  .... 
TL    

4 
1 

1,256.56 

1,240.02 

314.14 
1,240.02 

18.07 
71.35 

TQ  

1 

10.41 

10.41 

0.60 

TD  e-o  

2 

'  6.13 

3.07 

0.18 

Experimental  error 

20 

347.60 

17.38 

Total 

25 

29,560.00 

316  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

confirms  our  subjective  judgment  that  the  response  of  yield  to  rate  of 
application  of  the  fertilizer  is  linear  within  the  range  of  the  levels  of 
fertilizer  applied.  This  suggests  that  the  rate  of  application  of  the 
fertilizer  might  be  increased  even  more,  with  an  accompanying  increase 
in  the  yield.  However,  the  reader  is  warned  that  extrapolation  of  the 
linear  relationship  much  beyond  40  Ibs./plot  could  (possibly)  lead  to 
erroneous  conclusions.  Another  way  of  putting  this  is  to  say  that  the 
optimum  level  of  fertilizer  application  has  probably  not  yet  been 
reached,  and  further  experimentation  should  be  carried  out  along  these 
lines. 

11.12  ANALYSIS  OF  A  COMPLETELY  RANDOMIZED  DE 
SIGN  INVOLVING  FACTORIAL  TREATMENT  COM 
BINATIONS 

By  now  the  reader  should  be  gaining  some  facility  in  the  calculation 
of  sums  of  squares  associated  with  various  sources  of  variation.  Thus, 
the  advent  of  another  special  situation,  namely,  factorial  treatment 
combinations,  should  present  no  new  problems.  In  fact,  once  the  reader 
realizes  that  the  factorial  analysis  is  simply  another  way  of  partitioning 
the  treatment  sum  of  squares,  he  is  well  on  the  way  to  a  solution. 
Let  us  now  examine  the  details. 

It  has  previously  been  noted  that  the  usual  statistical  model  asso 
ciated  with  a  completely  randomized  design  involving  t  treatments  and 
n  experimental  units  per  treatment  is 

Ya  =  M  +  T*  +  e,y;  i  =   1,  -  •   -  ,  t  (11.66) 

j  =  ly  .  .   .  y  n. 

If  we  are  now  informed  that  the  t  treatments  are  actually  all  combina 
tions  of  a  levels  of  factor  a  and  b  levels  of  factor  &  (that  is,  t  =  a&) ,  the 
statistical  model  may  be  rewritten  as 

Ftf*  =  fji  +  oLi  +  fr  +  (<*#)#  +  €#*;          i  =   1,  •  •  •  ,  a  (11 .  67) 

J  =   1,  •  •  '  ,  ft 
k  =   1,  •  -  -  ,  n 

where j^is^  the  true  mean  effect,  ca  is  the  true  effect  of  the  ith  leveLof 
factor  a,  £/  is  the  true  effect  of  the  jth  level  of  factor  &,  (<*£)*/  is  the 
true  effect  "of  the  interaction  of  the  ith  level  of  factor  a  with  the  jth 
level  of  factor  6,  and  e-^k  is  the  true  effect  of  the  &th  experimental  unit 
subjected  to  the  (i?)th  treatment  combination.  As  usual,  it  is  assumed 
that  M  is  a  constant  and  that  the  e*/*  are  NID  (0?  a-} .  Rather  than  discuss 
assumptions  concerning  oa}  /3j,  and  (cqS)*/  at  this  time,  our  attention 
will  be  directed  towards  the  calculation  of  the  various  sums  of  squares. 
When  tlie  method  of  calculation  has  been  explained,  we  shall  return 
to  the  assumptions  and,  as  a  consequence,  to  the  expected  mean 
squares,  estimation  and  test  procedures,  and  other  related  topics. 

A  moment's  reflection  will  confirm  that  the  basic  calculations  are 
unchanged.  That  is,  ^,Y2,  Myy,  Tvy,  and  Eyy  will  all  be  calculated  as 


11.12       FACTORIAL   TREATMENT    COMBINATIONS 

before.  However,  if  we  adopt  the  following  notation: 

Ai  =   total  of  all  observations  associated  with  the  ith  level 

of  factor  a 
b      n  (11.68) 

=  z;  z;  Y«» 

y=i  £=1 
BJ  =  total  of  all  observations  associated  with  the  jth  level 

of  factor  6 

(11.69) 

-  ib  i:  Y«*> 

1=1  A^I 

and 

TV  =  total  of  all  observations  associated  with  both  the  ith 

level  of  factor  a  and  thejth  level  of  factor  b 
=  entry  in  the  (i/)th  cell  of  the  a  X  6  table  (11.70) 

=  i: 


it  may  be  shown  that 

AyV  =  sum  of  squares  associated  with  the  different  levels  of  a 

(F<  -    F)2 

(11.71) 


=  sum  of  squares  associated  with  the  different  levels  of  b 
=  an 


(11.72) 


and 

•S^ab  =  among  subclasses  (cells)  sum  of  squares  for  the  #X6  table1 


(U.73) 

a          b 


1  The  reader  \vill  recognize  that,  in  this  particular  situation,  Sab  =  Tyy.  How 
ever,  the  new  notation  and  terminology  were  introduced  at  this  time  to  acquaint 
the  reader  with  a  system  (of  notation,  terminology,  and  calculation)  that  will 
prove  most  valuable  when  factorials  involving  more  than  two  factors  are  analyzed. 


318  CHAPTER    T1r    COMPLETELY    RANDOMIZED    DESIGN 

Using  the  preceding  results,  it  may  be  verified  that 

=  sum  of  squares  associated  with  the  interaction  of 
factors  a  and  b 


=  n  :  ; 

*=i  y—  i 


,,  -  F,  -  Y,  + 


(11.74) 


These  results  are  summarized  in  ANOVA  form  in  Table  11.30. 

TABLE  11.30-ANOVA  for  a  Two-Factor  Factorial  in  a  Completely 

Randomized  Design 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 

Square 

Mean  

1 

MW 

M 

Treatments 
A  

a—I 

•&-int 

A 

B  

b  —  1 

Bin, 

B 

AB   

(a—  1)(6—  1) 

(A&)yy 

AB 

Experimental  error 

ab(n  —  1) 

JCLim 

E 

Total 

abn 

T\  F2 

Having  explained  the  calculation  of  the  various  sums  of  squares, 
your  attention  Is  now  directed  to  the  assumptions  associated  with  the 
or*,  £j,  and  (a/8)v-.  There  are  four  possible  sets  of  assumptions  that  can 
be  made  with  respect  to  the  true  treatment  effects.  These  are  discussed 
below. 

Model  I:  Analysis  of  Variance  (Fixed  Effects)  Model 

This  model  is  assumed  when  the  researcher  is  concerned  only  with  the 
a  levels  of  factor  a  and  the  b  levels  of  factor  b  present  in  the  experiment. 
Mathematically,  these  assumptions  are  summarized  by: 


="  0. 


y— i 


Model  II:  Component  of  Variance  (Random  Effects)  Model 

This  model  is  assumed  when  the  researcher  is  concerned  with:  (1)  a 
population  of  levels  of  factor  a  of  which  only  a  random  sample  (the  a 
levels)  are  present  in  the  experiment  and  (2)  a  population  of  levels  of 
factor  b  of  which  only  a  random  sample  (the  6  levels)  are  present  in 


11.12       FACTORIAL  TREATMENT    COMBINATIONS  319 

the  experiment.  Mathematically,  these  assumptions  are  summarized 
as  follows: 

at  are  NID  (0?  <rj 

ft-  are  NID  (0,  <rj 
(aftis  are  NID  (0, 


Model  III:  Mixed  Model  (a  Fixed,  b  Random) 

This  model  is  assumed  when  the  researcher  is  concerned  with:  (1) 
only  the  a  levels  of  factor  a  present  in  the  experiment  and  (2)  a  popu 
lation  of  levels  of  factor  6  of  which  only  a  random  sample  (the  6  levels) 
are  present  in  the  experiment.  Mathematically,  these  assumptions  are 
summarized  as  follows  : 

i>;=   2b(«/3)*  =  0 

twi  t-=i 

ft-  are  NID  (0,  ^). 
Please  note  that   231=1  («£)#  was  not  assumed  to  be  0. 

Model  III:  Mixed  Model  (a  Random,  b  Fixed) 

This  model  is  assumed  when  the  researcher  is  concerned  with:  (1)  a 
population  of  levels  of  factor  a  of  which  only  a  random  sample  (the  a 
levels)  are  present  in  the  experiment  and  (2)  only  the  &  levels  of 
factor  &  present  in  the  experiment.  Mathematically,  these  assumptions 
are  summarized  as  follows: 

oa  are  NID  (07  <r«) 

=  0. 


Please  note  that    X)?-!  G*£)v  was  not  assumed  to  be  0. 

While  the  logic  underlying  the  preceding  mathematical  formulations 
is  beyond  the  scope  of  this  text,  it  is  hoped  that  the  validity  of  the 
expressions  will  be  substantiated  by  the  arguments  which  will  accom 
pany  the  specification  of  the  several  F-tests.  Thus,  it  is  requested  that 
the  reader  accept  the  expressions  in  good  faith  and  concentrate  on 
learning  the  methods  of  analysis.  In  the  long  run,  this  will  prove  most 
beneficial. 

Based  on  the  foregoing  assumptions,  the  expected  mean  squares  may 
now  be  derived.  As  in  the  preceding  examples,  the  derivations  will  be 
omitted  and  only  the  results  tabulated.  The  expected  mean  squares 
for  each  of  the  four  cases  are  shown  in  Table  11.31. 

Examination  of  the  expected  mean  squares  in  Table  11.31  will  indi 
cate  the  proper  F-tests  for  such  hypotheses  as  HI  :  a.*  =  0  (i  =  1  ,  -  •  •  ,  a)  , 
#2:/3y  =  0  (j  =  l,  -  •  -  ,  &),  Jy3:(a/S)iy  =  0  (i=l,  -  -  -  ,  a;  j  =  1,  -  -  •  ,  6), 


•, 

O 

U 


is 


8  g 


Ml 

' 


l 

w  j? 
i  ^ 


? 

i—  i 

» 

J» 

O 

a 

-W! 

1 

S 

»—  i 

M 

"|     "|     "| 

1 

M                        M                        «                 « 

b             b             b         b 

O 
TJ 

7 

-S 

«> 

03 

-W3 

8 

I 

tT 

•8 

N^    T   M| 

03 

"O 

1 

"b           Is            %        *b 

03 
03 

w 

W 
1—  1 

"1  T 

'03 

•§ 

"1  "1  "I 

C*                         C4                          C4                  C4 

b             b             b         b 

hH 

1 

iH 

^ 
I        I       "1 

«w.i  -wi  tS 

%           *b           "b       "b 

| 

4-J 

8 

T3 

S 

s 

03 

Source  of  \ 

!               1 

a                        .| 
S  S     ^       ccj       -^     cl 

S   S                                     H 

11.12       FACTORIAL   TREATMENT    COMBINATIONS 


321 


TABLE   ll.Sa-.P-Ratios  for  Testing  the  Appropriate  Hypotheses  When  Dealing 

With  a  Two-Factor  Factorial  in  a  Completely  Randomized  Design  (See 

Table  11.30  for  the  ANOVA  and  Table  1 1.31  for  the  Expected 

Mean  Squares) 


Source  of  Variation 

F-Ratio 

Model  I 

Model  II 

Model  III 
(a  fixed,  b 
random) 

Model  III 
(a  random, 
b  fixed) 

Mean  

Treatments 
A  

A/E 
B/E 
AB/E 

A/AB 
B/AB 
AB/E 

A/AB 
B/E 
AB/E 

A/E 
B/AB 
AB/E 

B  .     .  .  .            ... 

AB 

Experimental  error.  .  . 

Total 

=0.  For  your  convenience  these  are 


H  4:0^  =  0,  H5:o%  =  Qy  and  H& 
specified  in  Table  11.32. 

Before  attempting  a  discussion  of  the  reasons  why  the  expected  mean 
squares  (and  thus  the  ^-tests)  are  as  indicated,  a  three-factor  factorial 
will  be  considered.  When  this  has  been  done,  a  general  discussion  of 
test  procedures  will  be  undertaken  and  numerical  examples  presented. 

When  a  three-factor  factorial  is  associated  with  a  completely 
randomized  design  involving  n  experimental  units  per  treatment  com 
bination,  the  appropriate  statistical  model  is 


i  =  1,  -  -  •  ,  a    (11.75) 

y  =  i,  •  •  -  ,  » 

k  =  1,  -  -  -  ,  c 
*=!,-••,» 

in  which  all  terms  are  defined  in  a  manner  analogous  to  ttie  definitions 
accompanying  Equation  (11.67).  The  basic  sums  of  squares,  namely, 
]F^F2,  Myy,  Tyy,  B^d  E  yy  &*&  calculated  in  the  usual  way.  Then,  if  one 
forms  an  aX&Xc  table,  an  aX&  table,  an  aXc  table,  and  a  &Xc  table, 
the  remaining  sums  of  squares  may  be  found  as  follows: 


among  cells  sum  of  squares  for  the 


table 


(11.76) 


t=i  y—  i  fc—  i 


322  CHAPTER    II,    COMPLETELY    RANDOMIZED    DESIGN 

X  b  table 

(11.77) 
v  y 


=  among  cells  sum  of  squares  for  the  a  X  b  table 


Sac  —  among  cells  sum  of  squares  for  the  a  X  c  table 

\h    5T*         2     , 

=  2-.,  2^  Tik/bn  —  Myv,  (11.78) 

Sbc  =  among  cells  sum  of  squares  for  the  b  X  c  table 

T\/an-M  (11.79) 

ben  —  Mm,  (11.80) 

=  22  JB,-/acn  —  Myy,  (11.81) 

==  X)  Ck/abn  —  Myy,  (11.82) 

A=I 

=      Sab      —       Ayy     —      Byy,  (11.83) 

~=     ^ac      -       Ayy     ~      CVU,  (11.84) 

=     Sic     —      BVV     —      Cyy,  (11    .    gS) 

In  the  above  expressions,  the  various  totals  are  denned  as  shown 
below : 

T%jk  —  total  of  all  observations  associated  with  the  ith  level 
of  factor  a,  the  yth  level  of  factor  b,  and  the  kth  level 
of  factor  c 
==  entry  in  the  (ijtyth  cell  of  the  a  X  b  X  c  table  (11.  87) 


and 


total  of  all  observations  associated  with  the  ith  level 

of  factor  a  and  the  yth  level  of  factor  6 

entry  in  the  (ij)th  cell  of  the  a  X  b  table  (11.88) 


*  1    Z=l 


11.12       FACTORIAL    TREATMENT    COMBINATIONS  323 

total  of  all  observations  associated  with  the  iih  level 

of  factor  a  and  the  kth  level  of  factor  c 

entry  in  the  (ijfe)th  cell  of  the  a  X  c  table  (11.89) 

b          n  b 

/   /      s     j     *    ijkl    ===:       S  _^     -t    iyky 

y-1    Z=i  y_i 

total  of  all  observations  associated  with  the/th  level 

of  factor  b  and  the  kih  level  of  factor  c 

entry  in  the  (jfyih  cell  of  the  b  X  c  table  (11.90) 


i=l    Z=l  1=1 

total  of  all  observations  associated  with  the  ith  level 
of  factor  a 

«*  (ii  .91) 


total  of  all  observations  associated  with  jth  level  of 
factor  b 

</«  =  ib  s  rw  (n.92) 


and 

Ck  =  total  of  all  observations  associated  with  the  £th  level 
of  factor  c 

=  z  i:  i:  YM  = 


The  pertinent  sums  of  squares  are  summarized  in  ANOVA  form  in 
Table  11.33. 

As  was  the  case  with  a  two-factor  factorial,  the  assumptions  concern 
ing  the  true  treatment  effects  can  take  several  forms.  In  fact,,  for  a 
three-factor  factorial,  there  are  eight  different  situations.  Rather  than 
discuss  all  of  these,  only  four  representative  cases  will  be  exhibited. 
These  are  described  below. 


324  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

TABLE   11.33-ANOVA  for  a  Three-Factor  Factorial  in  a  Completely 

Randomized  Design 


Source  of  Variation 

Degrees  of  Freedom 

Sum  of 
Squares 

Mean 
Square 

IVLean          

1 

-Ww 

M 

Treatments 
A    

a—  I 

Ayy 

A 

B               

b  —  1 

Byy 

B 

C    

c—\ 

r* 

^w 

C 

AB   

(a—  1)(6-1) 

(AB~)m 

A# 

AC            

(a—  l)(e-l) 

(AC)m 

AC 

BC            

(6—  l)(e—  1) 

(.BQyy 

BC 

ABC  

(o—  l)(6-l)(c-l) 

(ABQW 

ABC 

Experimental  error.  . 

abc(n—  1) 

Eyy 

E 

TVktcil 

/T  hfyt 

y  FZ 

-^^    •*• 

Model  I :  Analysis  of  Variance  (Fixed  Effects)  Model 

This  model  is  assumed  when  the  researcher  is  concerned  only  with 
the  a  levels  of  factor  a,  the  b  levels  of  factor  6,  and  the  c  levels  of  factor 
c  present  in  the  experiment.  Mathematically,  these  assumptions  are 
summarized  by: 


i=»i 


y=i 
& 


=  0. 


Model  II  :  Component  of  Variance  (Random  Effects)  Model 

This  model  is  assumed  when  the  researcher  is  concerned  with:  (1)  a 
population  of  levels  of  factor  a  of  which  only  a  random  sample  (the  a 
levels)  are  present  in  the  experiment,  (2)  a  population  of  levels  of  factor 
b  of  which  only  a  random  sample  (the  b  levels)  are  present  in  the  experi 
ment,  and  (3)  a  population  of  levels  of  factor  c  of  which  only  a  random 
sample  (the  c  levels)  are  present  in  the  experiment.  Mathematically, 
these  assumptions  are  summarized  as  follows: 

<xz  are  NTD  (0,  <ra) 
/3j  are  NID  (0, 


fc  are  NID  (0,  <rT) 

tf  are  NID  (0,  cra/3) 

*  are  NID  (0,  <r«Y) 

&  are  NID  (0, 

ik  are  NID  (0, 


11.12      FACTORIAL  TREATMENT    COMBINATIONS  325 

Model  III:  Mixed  Model  (a  and  b  Fixed,  c  Random) 

This  model  is  assumed  when  the  researcher  is  concerned  with:  (1) 
only  the  a  levels  of  factor  a  present  in  the  experiment,  (2)  only  the  & 
levels  of  factor  6  present  in  the  experiment,  and  (3)  a  population  of 
levels  of  factor  c  of  which,  only  a  random  sample  (the  c  levels)  are 
present  in  the  experiment.  Mathematically,  these  assumptions  are 
summarized  as  follows: 


**  =  0 
i—  i  y—  i 

Y*  are  NID  (0,  <JT). 
Please  note  that 

c  c 

]C  Or)**,     ]C  (^y)y*,   and 
&=1  A=l 

were  ?zoi  assumed  to  be  0. 


Model  III:  Mixed  Model  (a  Fixed,  b  and  c  Random) 

This  model  is  assumed  when  the  researcher  is  concerned  with:  (1) 
only  the  a  levels  of  factor  a  present  in  the  experiment,  (2)  a  population 
of  levels  of  factor  6  of  which  only  a  random  sample  (the  &  levels)  are 
present  in  the  experiment,  and  (3)  a  population  of  levels  of  factor  c  of 
which  only  a  random  sample  (the  c  levels)  are  present  in  the  experi 
ment.  Mathematically,  these  assumptions  are  summarized  as  follows: 


f  are  NID  (0, 

k  are  NID  (0,  <rr) 

are  NID  (0, 
Please  note  that 

5  c  & 


k9    and 

j—  1 

were  ?^o^  assumed  to  be  0. 

Based  on  the  foregoing  assumptions,  the  expected  mean  squares  are 
derived  and  the  results  presented  in  Table  11.34.  The  proper  7^-tests  for 
various  hypotheses  are  shown  in  Table  11.35. 

Having  outlined  the  methods  of  calculation  associated  with  two-  and 
three-factor  factorials  in  a  completely  randomized  design,  we  are  now 


£  fi 


fl 

I   «' 

fr  9 ; 


UJ   o 

s -• 

cu    o3 


ted 
ial 
e  1 


to 
ab 


Ex 
Fa 


pec 


4. 


11 

a  ~, 

I 
s 


3 


•S    ^ 

o    .^ 


1 


•g 


S  -n 

$  £ 


" 


- 


+ 


t 


b 
a 
5i 


Wl 


«W3 

+ 

°b 


I 
^ 


§ 

» 

4- 


-1 
-3 


+        + 


t 


t 
"I 


I 


OS     ^ 
+ 

"I 


t 


Wl 


O 

H 


11.12       FACTORIAL    TREATMENT    COMBINATIONS 


327 


ready  to  discuss  the  expected  mean  squares  exhibited  in  Tables  11.31 
and  11.34  and  the  ^-ratios  exhibited  in  Tables  11.32  and  11.35.  Perhaps 
the  best  way  to  approach  this  topic  is  to  talk  about  the  types  of  in 
ferences  that  the  researcher  wishes  to  make.  You  will  recall  that  the 
various  Models  (I,  II,  and  III)  reflect  the  researcher's  desire  to  make 
inferences  about :  (1)  only  the  levels  of  the  factors  present  in  the  experi 
ment,  (2)  populations  of  levels  of  factors  of  which  only  a  random 

TABLE    11.35-F-Ratios  for  Testing  the  Appropriate  Hypotheses  When 

Dealing  With  a  Three-Factor  Factorial  in  a  Completely  Randomized 

Design  (See  Table  11.33  for  the  ANOVA  and  Table  11.34 

for  the  Expected  Mean  Squares) 


Source  of  Variation 

.F-Ratio 

Model  I 

Model  II 

Model  III 
(a  and  b 
Fixed,  c 
Random) 

Model  III 
(a  Fixed, 
b  and  c 
Random) 

Mean  

Treatments  

A/E 
B/E 
C/E 
AB/E 
AC/E 
BC/E 
ABC/E 

no  exact  test 
no  exact  test 
no  exact  test 
AB/ABC 
AC/ABC 
EC/ABC 
ABC/E 

A/AC 
B/BC 
C/E 
AB/ABC 
AC/E 
BC/E 
ABC/E 

no  exact  test 
B/BC 
C/BC 
AB/ABC 
AC/ABC 
BC/E 
ABC/E 

A  

B  

c  

AB   .      .  .     . 

AC  

BC   

ABC 

T^yp^Tirnprital  error 

Total 

sample  (of  levels  from  each  population)  is  present  in  the  experiment* 
and  (3)  a  mixture  of  the  two  preceding  situations,  respectively.  In  each 
of  these  situations,  the  researcher  may  reason  as  follows : 

(1)  When  dealing  with  a  situation  in  which  Model  I  applies,  the 
conclusions  reached  about  any  particular  effect  will  be  un- 
contaminated  by  any  other  effect  since,  by  proper  definition 
of  the  terms  in  the  statistical  model,  the  average  contribution 
of  every  other  effect  can  be  made  equal  to  zero.  Consequently, 
all  ^-values  will  be  calculated  by  forming  the  ratio  of  the 
mean  square  for  the  effect  under  scrutiny  and  the   experi 
mental  error  mean  square.  That  is,  all  effects  are  tested  against 
experimental  error. 

(2)  When  dealing  with  a  situation  in  which  Model  II  applies,  the 
conclusions  reached  about  any  particular  effect  will  be  con 
taminated  by  all  those  effects  which  represent  interactions 


328  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

between  the  effect  under  scrutiny  and  other  effects  present  in 
the  experiment.  This  reflects  the  researcher's  realization  that 
his  conclusions  (inferences)  about  the  effect  under  scrutiny  are 
uncertain  not  only  because  of  the  e's  but  also  because  of  the 
chance  contributions  of  the  randomly  selected  levels  of  any 
factor.  That  is,  a  different  random  sample  (of  levels  of  any 
factor)  might  lead  to  different  conclusions,  and  the  researcher 
attempts  to  incorporate  this  uncertainty  into  his  conclusions 
by  testing  a  particular  effect  against  an  "error"  which  includes 
an  estimate  of  this  additional  variability.  Thus,  the  expected 
mean  squares  will  be  as  shown  in  Tables  11.31  and  11.34  where 
it  is  observed  that  each  expected  mean  square  contains  all  the 
components  of  variance  whose  subscripts  contain  all  the 
letters  representing  the  effect  under  scrutiny.  This  is  the 
mathematical  way  of  expressing  the  ''contamination"  dis 
cussed  above.  Consequently,  the  -P-tests  are  as  specified  in 
Tables  11.32  and  11.35.  [NOTE:  This  illustrates  the  remark 
made  in  Chapter  10,  namely,  "  .  .  .  the  (proper)  experimental 
error  for  testing  a  particular  effect/7  ] 

(3)  When  dealing  with  a  situation  in  which  Model  III  applies,  the 
conclusions  reached  about  any  particular  effect  may  or  may 
not  be  contaminated  by  other  effects.  That  is,  we  have  a 
mixture  of  cases  (1)  and  (2).  To  summarize  what  could  be  a 
rather  involved  discussion,  let  us  state  the  following  rule: 

The  expected  mean  square  for  any  effect  will  contain,  in 
addition  to  its  own  special  term,  all  components  of  variance 
which  represent  interactions  between  the  effect  under 
scrutiny  and  other  effects  whose  levels  were  randomly 
selected.  It  will  not  contain  components  of  variance  repre 
senting  interactions  between  the  effect  under  scrutiny  and 
other  effects  whose  levels  comprise  the  entire  population 
(of  levels)  to  be  investigated. 

The  expected  mean  squares  specified  in  Tables  11.31  and  11.34 
are,  of  course,  simply  results  of  the  above  reasoning  and,  as  a 
consequence,  the  F-tests  are  as  shown  in  Tables  11.32  and 
11.35.  [NOTE:  Again,  this  illustrates  the  remark  made  in 
Chapter  10,  namely,  "  .  .  .  the  (proper)  experimental  error 
for  testing  a  particular  effect/'] 

To  aid  the  researcher  in  writing  out  expected  mean  squares  for  different 
situations,  the  following  sequence  of  steps  is  recommended: 

(1)  Include,  when  applicable,  a  component  of  variance  for  each 
subsampling  stage. 

(2)  Include  a  component  of  variance  representing  experimental 
error. 

(3)  Include  every  component  of  variance  whose  subscripts  include 


11.12       FACTORIAL   TREATMENT    COMBINATIONS  329 

all  the  letters  specifying  the  effect  Math  which  the  expected 
mean  square  is  associated. 

(4)  Insert  coefficients  in  front  of  each  component  of  variance  in 
accordance  with  the  approach  discussed  in  Section  11.5. 

(5)  Delete  from  the  set  specified  in  step  (3),  all  terms  representing 
interactions  between  the  effect  associated  with  the  expected 
mean  square  and  other  effects  whose  levels  were  not  randomly 
selected. 

(6)  For  a  main  effect,  replace  the  component  of  variance  for  that 
effect  by  a  "sum  of  squares  divided  by  the  appropriate  degrees 
of  freedom"  if  the  effect  is  a  "fixed  effect," 

Before  presenting  illustrations  of  the  methods  discussed  in  this 
section,  some  additional  remarks  need  to  be  made.  In  the  interest  of 
economy,  these  are  presented  here  in  the  briefest  form  possible: 

(1)  It  will  have  been  noted  that  the  degrees  of  freedom  for  inter 
action  effects  were   specified  without  any  explanation.   The 
general    rule    is:     For     an     interaction     effect     denoted    by 
ABCD  -  -  -  ,  the  degrees  of  freedom  are 

v  =  (a  —  !)(&  —  l)(c  —  1)(<2  —  1)  -  -  -  . 

(2)  When,  as  in  Table  11.35,  no  exact  tests  of  certain  hypotheses 
are    available,    approximate    tests    can    be    made    following 
Satterthwaite's  procedure.    (See  Section  11.7.)   For  example, 
when  Model  II  was  assumed  in  Table  11.34,  an  approximate 
test  of  H:  o-l  =  Q  is 

p  £*  A/[AB  +  AC  -  ABC}. 

(3)  Conclusions  (inferences)  about  one  factor  in  a  factorial  must 
take  due  cognizance  of  all  interactions  of  this  factor  with  other 
factors.  That  is,  recommendations  about  one  factor  must  give 
consideration  to  the  way  in  which  its  effect  is  influenced  by 
other  factors. 

Example  11.16 

Consider  a  4X3  factorial  in  a  completely  randomized  design  with 
three  experimental  units  per  treatment  combination.  The  data  are 
given  in  Table  11.36.  Proceeding  as  indicated,  the  following  sums  of 
squares  were  calculated: 

53  F2  =  564,389 
Myy  =  (4023)2/36  =  449,570.2 

Tvy  «  Sat,  =  [(306) 2  +  *  *  •  4-  (268)2]/3  —  449,570.2  =  67,160.8 
Evv  =  564,389  —  449,570.2  —  67,160.8  =  47,658,0 

Ayy  —  [(726)2  +  (991)2  +  (1022)2  +  (1284)2]/9  —  449,570.2  =  17,351.7 
Byy  =  [(1624)2  +  (1500)2  +  (899)2]/12  —  449,570.2  =•  25,061.2 
=  67,160.8  -  17,351.7  -  25,061.2  =  24,747.9. 


33O 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


These  results  are  summarized  in  ANOVA  form  in  Table  11.37,  where 
the  expected  mean  squares  are  shown  for  each  of  the  four  cases  illus 
trated  in  Table  11.31.  Since  the  data  were  hypothetical,  no  F-tests  will 
be  performed-  Such  tests  and  the  resulting  inferences  will  be  illustrated 
in  succeeding  examples  in  which  actual  experimental  data  will  be 
examined. 

TABLE   11.36-Hypothetical  Data  for  Illustrating  the  ANOVA  for  a 
4X3  Factorial  in  a  Completely  Randomized  Design 


ox 

#2 

a3 

a4 

bi         Z>2         63 

b\         bz        bz 

bi         b%        b$ 

&i         b%         &s 

128 

34 

16 

152 

40 

118 

76 

102 

132 

180 

220 

60 

42 

134 

18 

128 

88 

80 

158 

96 

60 

90 

220 

48 

136 

172 

46 

216 

76 

93 

168 

162 

68 

150 

156 

160 

Example  11 .17 

Consider  an  agronomic  experiment  to  assess  the  effects  of  date  of 
planting  (early  or  late)  and  type  of  fertilizer  (none,  Aero,  Na,  or  K)  on 
the  yield  of  soybeans.  Thirty-two  homogeneous  experimental  plots  were 
available.  The  treatments  were  assigned  to  the  plots  at  random,  subject 
only  to  the  restriction  that  4  plots  be  associated  with  each  of  the  8 
treatment  combinations.  The  data  are  given  in  Table  11.38  and  the 
ANOVA  (assuming  Model  I)  in  Table  11.39. 

TABLE   11. 38- Yields  of  Soybeans  at  the  Agronomy  Farm,  Ames,  Iowa,  1949 

(In  bushels  per  acre) 


Date  of 

Planning 

Fertilizer 

Experimental  Units  Within  Treatments 

1 

2 

3 

4 

Early 

Check 
Aero 

Na 
K 

Check 
Aero 

Na 
K 

28.6 
29.1 
28.4 
29.2 

30.3 
32.7 
30.3 
32.7 

36.8 
29.2 
27.4 
28.2 

32.3 
30.8 
32.7 
31.7 

32.7 
30.6 
26.0 

27.7 

31.6 
31.0 
33.0 
31.8 

32.6 
29.1 
29.3 
32.0 

30.9 
33.8 
33.9 
29.4 

Late  

Assuming  oi  =  0.01,  it  is  seen  that  the  hypothesis  "date  of  planting  has 
no  effect' J  must  be  rejected.  Examination  of  the  mean  yields  indicates 
that  the  later  date  of  planting  is  better  (i.e.,  is  associated  with  higher 
yields).  Of  course,  more  information  is  needed  concerning  the  distinc 
tion  between  "early"  and  "late"  before  explicit  recommendations  can 
be  made.  No  statistically  significant  effects  were  noted  for  either 


Q 
M 
o 


*i 

CO 


PQ 


"co: 

•  £H 

-WI 

j—  (       E 

0? 

^       ^ 

cs 

11 

..    A   ^ 

b                b                b 

^ 

?       ?       f 

"*  —  " 

*b             *b             *b              b 

0 

-W.I 

o 

H   £? 

5 

s 

cr 

CO 

T3 

"£««.«. 

cu 

E 

+    +     +     . 

°b            *b             *b              t> 

CJ 

1* 

H-t 
>—  1 

^8           "cf 

1 

S 

c,^              «t              «^ 

o               b             ^ 

*b             *b             °b               b 

HH 

"5       ?V  c°^"^ 

ro> 

**W"  <0^"^  *W3 

o 

i"  1  s" 

•b      -b      "b      "b 

000*0^0 

C^S                 OS                vO                ^O                ^>* 

53     °* 

O                CO*                O                ^                V2 
?.              oo              ro              o>              g 

w 

ON                  »^5                  C*4                  ^                  •»-< 

cq                4>-                CM                O\                C5 

o 

1-8    1 

CO                 o^ 
CO 

o          ^          ^          ^          gg 

S 

CO 

§3            o 

,       ^               CO               CM               0               ^ 

NO 

CO 

Q         £ 

g 

^  ^  "5 
g  ^  -S 

1                                                          *§§ 

-i—  « 
O 

H 

CO               ^ 

g      o     ^3             BQ             X           g* 

^5!       r    i                                                                      ["J 

KH      CT^                                                          *    ' 

332 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


fertilizers  or  for  the  interaction  between  fertilizers  and  date  of  planting. 
(NOTE:  Had  a.  been  chosen  as  0.05,  the  interaction  effect  would  have 
been  significant*  This  illustrates  the  dependence  of  the  inferences  upon 
the  choice  of  significance  level,  a  fact  which  is  sometimes  overlooked, 
or  forgotten,  by  the  analyst.  That  is,  we  must  always  remember  that  a 
statement  about  significance  or  nonsignificance  is  a  direct  function  of 
the  selected  value  of  a..} 

TABLE   11.39-ANOVA  for  Experiment  Described  in  Example  11.17 
(Data  Given  in  Table  11.38) 


Degrees 

Source  of 

of 

Sum  of 

Mean 

Expected  Mean 

JF- 

Variation 

Freedom 

Squares 

Square 

Square 

Ratio 

Mean  

1 

30,368.80 

30,368.80 

Treatments 

Dates  of  planting 

1 

32.00 

32.00 

<r*  +  (16/1)  i  «w 

10.42 

Fertilizers  

3 

16.40 

5.47 

2+rs/,n^*2 

O"     ~j  —    {&/  *J  )    f       M  i 

1.78 

Fertilizers  X  dates 

y-i 

sf              4-v                   2 

of  planting,  .  .  . 

3 

38.40 

12.80 

4.17 

t—  i      y-i 

Experimental  error 

24 

73.74 

3.07 

cr2 

Total 

32 

30,529.34 

Example  11.18 

Consider  a  3X4X3  factorial  in  a  completely  randomized  design  with. 
6  experimental  units  per  treatment  combination.  The  data  are  given 
in  Table  11.40.  Proceeding  as  directed  earlier,  Tables  11.41  through 
11*44  were  obtained  and  the  following  sums  of  squares  calculated: 

23  ^2  =  27,981 
Myy  =  19,703.56 


Tvy 
Eyy 


$bc 

A 

•flyy 

Byy 
Cyy 


Sabc  =  3283.27 
4994.17 
2913.27 
1065.32 

670.83 

941.79 

463.79 

84.93 

1507.69 

38.60 

122.11 

124.36. 


11.12       FACTORIAL    TREATMENT    COMBINATIONS 


333 


TABLE   11.4O-Hypothetical  Data  for  Illustrating  the  ANOVA  for  a 
3X4X3  Factorial  in  a  Completely  Randomized  Design 


ax 

a2 

as 

bl     b%     bs    &4 

61   b*   b*   b4 

6,     62     *3    &4 

3 

10 

9 

8 

24 

8 

9 

3 

2 

8   9 

8 

2 

10 

9 

8 

29 

16 

11 

3 

2 

7   5 

3 

8 

10 

2 

8 

27 

16 

15 

8 

2 

15   7 

14 

Ci 

1 

6 

8 

14 

14 

13 

8 

5 

9 

30   9 

2 

7 

8 

9 

6 

18 

10 

2 

16 

14 

7   6 

11 

8 

1 

10 

12 

3 

8 

8 

4 

11 

2   2 

9 

29 

45 

47 

56 

115 

71 

53 

39 

40 

69  38 

47 

4 

12 

3 

8 

22 

7 

16 

2 

2 

2    7 

2 

7 

10 

5 

8 

28 

18 

10 

6 

6 

6   5 

9 

7 

9 

2 

7 

27 

15 

12 

7 

7 

16    1 

13 

Ci 

14 

5 

7 

15 

34 

11 

9 

5 

13 

11    8 

3 

7 

9 

8 

2 

19 

9 

12 

12 

13 

6   6 

12 

7 

6 

12 

3 

3 

15 

8 

4 

12 

3   2 

10 

46 

51 

37 

43 

133 

75 

67 

36 

53 

44  29 

49 

5 

10 

5 

8 

23 

9 

17 

3 

2 

8   6 

3 

9 

10 

27 

8 

28 

16 

11 

7 

8 

9   8 

15 

15 

7 

6 

15 

30 

14 

12 

5 

11 

18   3 

8 

C3 

8 

6 

4 

18 

16 

12 

13 

15 

17 

8   7 

16 

7 

17 

3 

10 

17 

10 

20 

9 

9 

8   6 

17 

3 

2 

10 

5 

3 

7 

8 

6 

11 

7   3 

14 

47 

52 

55 

64 

117 

68 

81 

45 

58 

58  33 

73 

These  results  are  summarized  in  ANOVA  form  in  Table  11.45.  Since 
the  data  were  hypothetical,  no  expected  mean  squares  are  given. 
Neither  are  any  F-tests  performed.  The  reader  is  referred  to  the  prob 
lems  at  the  end  of  the  chapter  for  illustrations  of  various  tests  and  the 
resulting  inferences. 


TABLE  11.41  — 


Table  Formed  From  the  Data  of  Table  11.40 


a 

i 

a>i 

i 

a 

3 

61 

b* 

*3 

64 

61 

b* 

&3 

64 

61 

bz 

63          64 

Cl 

29 

45 

47 

56 

115 

71 

53 

39 

40 

69 

38     47 

c%       .... 

46 

51 

37 

43 

133 

75 

67 

36 

53 

44 

29     49 

47 

.57, 

.5,5 

64 

117 

68 

81 

45 

58 

58 

33      73 

334 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 
TABLE   HA2-aXb  Table  Formed  From  the  Data  of  Table  11.40 


&! 122  365  151 

148  214  171 

139  201  100 

163  120  169 

TABLE   11.43-aXc  Table  Formed  From  the  Data  of  Table  11.40 

177  278  194 

177  311  175 

c3 218  I           311  222 

TABLE   11.44~6Xc  Table  Formed  From  the  Data  of  Table  11.40 

&i 

184  185  138       142 

232  170  133       128 

222  178  169       182 


TABLE   11.45-ANOVA  for  Data  of  Table  11.40 


Source  of  Variation 


Degrees  of 
Freedom 


Sum  of  Squares 


Mean  Square 


Mean 
Treatments 

A 

B 

C 

AB 

AC 

BC 

ABC 
Experimental  error , 


2 

3 

2 

6 

4 

6 

12 

180 


19,703.56 

941.79 

463 . 79 

84.93 

1,507.69 

38.60 

122.11 

124.36 

4,994.17 


19,703.56 

470.90 

154.60 

42.46 

251.28 

9.65 

20.35 

10.36 

27.75 


Total 


216 


27,981.00 


Even  though  this  section  is  already  quite  long,  there  are  several 
items  which  need  mentioning  before  we  leave  (for  the  time  being)  the 
subject  of  factorials.  These  items  are:  (1)  general  computational  pro 
cedures  for  factorials  involving  four  or  more  factors,  (2)  special  compu 
tational  methods  for  2n  and  3n  factorials,  (3)  subsampling  in  com 
pletely  randomized  designs  involving  factorial  treatment  combinations, 


11.12       FACTORIAL    TREATMENT    COMBINATIONS  335 

and  (4)  analysis  of  response  curves  associated  with  the  various  main 
effects  and  interactions.  A  brief  discussion  of  each  of  these  will  be 
given  in  the  following  paragraphs. 

The  general  computational  procedure  for  factorials  proceeds  as 
follows.  First  compute  ^Y*,  Myy,  Tyy,  Eyy,  and  any  sums  of  squares 
required  because  of  subsampling.  Then,  to  subdivide  Tw  in,  say,  a 
four-factor  factorial,  form  in  succession  the  four-way  table,  all  three- 
way  tables,  and  all  two-way  tables.  As  each  table  is  formed,  compute 
the  border  totals  as  a  check  on  the  entries  you  have  made  in  the  cells 
of  the  tables.  Then,  starting  with  the  two-way  tables,  calculate  the 
sums  of  squares  for  each  of  the  main  effects  and  for  each  of  the  two- 
factor  interactions.  Then,  proceeding  to  the  three-way  tables,  calculate 
the  sums  of  squares  associated  with  each  of  the  three-factor 
interactions.  And,  finally,  utilizing  the  four-way  table,  the  sum  of 
squares  associated  with  the  four-factor  interaction  may  be  obtained. 
The  extension  to  5,  6,  •  •  -  ,  AT  factors  is  easy.  After  obtaining  the 
basic  sums  of  squares,  form  the  AT-way  table,  all  possible  (N  —  l)-way 
tables,  all  possible  (N  —  2)-way  tables,  .  .  .  ,  all  possible  three-way 
tables,  and  all  possible  two-way  tables  in  the  order  mentioned.  Then 
calculate,  in  the  following  order,  all  main  effect  sums  of  squares,  all 
two-factor  interaction  sums  of  squares,  .  .  .  ,  all  (N  —  l)-f actor  inter 
action  sums  of  squares,  and  the  A^-factor  interaction  sum  of  squares. 

Whenever  all  the  factors  are  at  p  levels,  and  there  are  n  factors,  the 
statistician  refers  to  such  an  arrangement  as  a  pn  factorial.  Of  particular 
interest  are  those  cases  where  p  =  2  or  3.  When  such  cases  arise,  there 
are  available  to  the  research  worker  certain  special  computational 
techniques.  These  are  explained  in  considerable  detail  in  such  references 
as  Yates  (39)  and  Kempthorne  (23),  and  may  be  pursued  by  those 
readers  whose  primary  interest  is  in  computation.  Since  the  methods 
outlined  earlier  in  this  section  are  valid  for  all  cases,  there  seems  little 
reason  to  burden  the  reader  with  a  specialized  technique  at  this  time. 
Accordingly,  we  shall  do  no  more  than  has  already  been  done,  that  is, 
point  out  the  existence  of  the  methods  and  give  pertinent  references 
for  the  use  of  interested  persons. 

Wlten  subsampling  occurs  in  a  completely  randomized  design  in 
volving  factorial  treatment  combinations,  the  methods  of  analysis  are 
simply  a  combination  of  those  given  in  this  section  and  Section  11.4. 
Thus,  no  detailed  discussion  of  computational  techniques  will  be  pre 
sented  at  this  time.  However,  to  illustrate  the  nature  of  the  ANOVA's, 
two  cases  will  be  mentioned.  The  first  of  these  will  involve  only  one 
subsampling  stage,  while  the  second  will  involve  two  stages  of  sub- 
sampling.  If  only  one  stage  of  subsampling  is  involved,  the  appropriate 
statistical  model  (for  a  two-f actor  factorial)  is 

Yijkl  =  M  +  on  +  ft-  +  («£)  ^  +  €i/fc  +  IK**;     *=!,-••,«         (11 . 94) 

J  -  1,  -  •  •  ,  b 
£=!,•••,» 
/  =  1,  -  -  •  ,  p, 


336  CHAPTER    11,   COMPLETELY    RANDOMIZED    DESIGN 

TABLE  11  46-Abbreviated  ANOVA  for  a  Two-Factor  Factorial  in  a 
Completely  Randomized  Design  Involving  One  Stage  of  Subsamplmg 

(Model  I) 


Source  of 
Variation 


Mean. 

Treatments 


Experimental  error. 
Sampling  error 


Total 


Degrees  of 
Freedom 


-  1 


b  - 


(a  —!)(&  — 

ab(n  —  1) 
abn(p  —  1 


Expected  Mean  Square 


T2  -4-  pcrz  +  pnb  y^  ca/(a  —  1) 
17  i-i 

b  2 

CT2   +  2?<T2  +   PW*     X   &'/(&    ~    *) 
"  3-1 

^  +     <r*  +  pn±,f:  &ftl-/(a 


and  the  ANOVA  would  appear  (in  abbreviated  form)  as  m  Table  11.46. 
(NOTE :  If  only  one  sample  were  obtained  from  each  experimental  unit ; 
e  K  if  one  small  sample  is  taken  from  a  field  plot  to  estimate  the  yield 
of  the  entire  plot,  p  in  Table  11.46  is  set  equal  to  1  and  the  line  for 
"sampling  error"  is  deleted.  However,  if  the  whole  plot  is  harvested, 
the  sampling  error  is  0  and  the  ANOVA  would  be  as  shown  m  Table 
11  30  )  In  the  second  case  to  be  examined,  that  is,  a  case  involving 
two  stages  of  subsampling,  the  appropriate  statistical  model  (for  a  two- 
factor  factorial)  is 

(11.95) 


on 


i  =  1, 


3 
k 

m  = 


i, 
i, 
i, 


•  ,  a 

•  ,  n 

•  ,  P 

•  ,d, 


and  the  ANOVA  would  appear  (in  abbreviated  form)  as  in  Table  11-47. 
The  extension  to  cases  involving  more  than  two  stages  of  subsampling 

should  be  obvious.  -..-,-,     ^  •       ^.v, 

As  indicated  in  Section  11.11,  it  is  often  advisable  to  examine  the 
response  curve  which  summarizes  the  effects  of  the  various  levels  of  a 
factor  upon  the  characteristic  being  measured.  When  our  data  nt  a 
factorial  arrangement,  we  may  find  it  possible  to  examine  response 


11.12       FACTORIAL   TREATMENT    COMBINATIONS 


337 


TABLE   11.47-Abbreviated  ANOVA  for  a  Two-Factor  Factorial  in  a 
Completely  Randomized  Design  Involving  Two  Stages  of  Subsampling 

(Model  I) 


Source  of 
Variation             i 

Degrees  of 
Freedom 

Expected  Mean  Square 

jMCean.  

1 

Treatments 
A  

a—  1 

0.2_j-jo.2+^p<3.2_j_^pw5  ]T  «VO—  1) 

B     . 

b—l 

b 

2_|_^.2_|_^      2_|_  jp       y-  /3i/(b—  1) 

AB  

(a~  1)(6—  1) 

a      b 
crl+d^-t-dp^-i-dpn^,  ^Z  (ctpYa/(a—  1)(6  —  1) 

Experimental  error  .  . 
First  stage  sampling 
error  

ab(n-l) 
abn(p  —  1) 

17                                      »-l  y-x 

<T*-h<Z<T*+£2><r2 
2_|_^.2 

Second   stage  sam 
pling  error          .  . 

abnv(d  —  1) 

5   l        if 
o-* 

5 

Total 

abnpd 

curves  associated  with  the  levels  of  2  or  more  factors.  For  example,  if 
we  have  2  factors,  a  and  &,  we  may  subdivide  the  2  sums  of  squares, 
Ayy  and  Byy,  into  parts  designated  as  (AL)VV}  (Ao)y3/,  -  -  •  ,  and  (BxJ)yV) 
(Bo)yy,  -  •  •  ,  respectively.  That  is,  we  may  obtain  the  linear,  quad 
ratic,  -  -  •  ,  sums  of  squares  associated  with  each  of  the  factors  a  and  b. 
However,  since  we  are  now  dealing  with  factorials,  it  is  also  possible 
to  subdivide  the  interaction  sum  of  squares,  (AB}yy.  The  parts  into 
which  (AB}yy  may  be  subdivided  will  be  designated  as  (Ax,BrJ)Vy, 
(AxJEtQ^yy,  (AQBL)yi/,  (AgjBq)^,  •  •  •  .  If  a,  third  factor,  c,  were  present, 
we  would  then  have  such  quantities  as  (Ci^)yyy  (Co)yy,  f  A  -n -^ 


(ALBLCz,) 


yv, 


^  etc.  The  number  of 
possible  subdivisions  is,  of  course,  limited  by  the  number  of  levels  of 
the  various  factors  involved.  Because  we  have  already  devoted  so 
much  time  to  the  discussion  of  factorials  in  a  completely  randomized 
design,  the  details  of  this  technique  (i.e.,  response  curve  analyses  for 
the  various  main  effects  and  interactions)  will  not  be  discussed  here. 
However,  the  technique  will  be  discussed  in  the  following  chapter  in 
connection  with  a  randomized  complete  block  design.  Since  the  method 
is  the  same  regardless  of  the  design  (as  long  as  the  completely  random 
ized  design  has  equal  numbers  of  observations  in  each  category),  the 
person  desiring  the  details  now  can  jump  ahead  and  read  Section  12.12. 
The  reader  will  appreciate,  I  am  certain,  that  the  foregoing  discussion 


338  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

has  only  scratched  the  surface  of  the  subject  of  analyzing  factorials. 
However,  I  also  feel  that  sufficient  material  has  been  given  to  enable 
the  researcher  to  handle  the  most  commonly  occurring  situations. 
Should  more  complex  situations  arise,  reference  to  one  or  more  of  the 
books  listed  at  the  end  of  the  chapter  should  prove  helpful.  If  not,  a 
professional  statistician  should  be  consulted. 

11.13     NONCONFORMITY     TO     ASSUMED     STATISTICAL 
MODELS 

By  now  the  reader  is  well  aware  that  the  usual  assumptions  in  analy 
sis  of  variance  involve  the  concepts  of  additivity,  normality,  homo 
geneity  of  variances,  and  independence  of  the  errors.  However,  up  to 
this  point,  little  has  been  said  about:  (1)  tests  to  assess  the  validity 
of  the  assumptions,  (2)  the  consequences  if  the  assumptions  are  not 
satisfied,  and  (3)  transformations  which,  if  applied  to  the  original  data, 
may  justify  the  use  of  the  assumptions  in  connection  with  the  trans 
formed  data  (i.e.,  the  data  as  they  appear  after  the  transformation  has 
been  applied).  In  this  section  each  of  these  topics  will  be  discussed 
briefly.  For  those  who  wish  more  details,  several  references  are  given. 
In  particular,  three  excellent  expository  articles  are  those  by  Bartlett 
(3),  Cochran  (9),  and  Eisenhart  (18). 

First,  let  us  consider  various  statistical  tests  that  have  been  proposed 
to  check  on  the  validity  of  the  several  assumptions. 

Homogeneity  of  Variances 

In  Section  7.21,  Bartlett's  test  was  given  for  testing  the  hypothesis 
fl":cr?  =  cr|=  •  -  -  =<T|  where  a  random  sample  of  n^  observations  had 
been  taken  from  the  ith  normal  population  (i=  1,  •  -  •  ,  fc) .  Clearly,  this 
test  is  appropriate  for  checking  on  the  homogeneity  of  variances.  How- 
eyer5__Bartlett^s^test  has  been  shown  to  be  quite  sensitive  to_non- 
normality.  Thus,  if  nonnormality  is  suspected  or  has  been  demon 
strated,  the  test  should  be  modified  as  suggested  by  Box  and  Anderson 
(5).  For  a  discussion  of  other  tests,  the  reader  is  referred  to  Anscombe 
and  Tukey  (2),  Box  and  Anderson  (5),  David  (12),  and  Dixon  and 
Massey  (14). 

Normality 

To  check  on  the  assumption  of  normality,  one  can  use  the  chi-square 
test  of  goodness  of  fit  given  in  Section  7.15.  An  alternative,  and  perhaps 
preferred,  method  is  the  Kolniogorov-SniiiTLov^^tiesJL  -discussed  in 
Chapter  15.  For  those  who  are  satisfied  with  a  less  objective  approach, 
the  data  (or  the  residuals)  may  be  plotted  on  normal  probability  paper 
and  a  subjective  judgment  rendered. 

Additivity 

When  the  assumption  of  additivity  is  questioned,  the  problem  is 
somewhat  more  involved.  This  is  so  because  there  are  three  major 


11.13      NONCONFORMITY  TO   ASSUMED   STATISTICAL   MODUS 

causes  of  nonadditivity,  namely,  (1)  the  true  effects  may  be  multi 
plicative,  (2)  interactions  may  exist  but  terms  representing  such  effects 
have  not  been  included  in  the  assumed  model,  and  (3)  aberrant  obser 
vations  may  be  present.  If  the  experimental  design  is  such  that  inter 
action  effects  may  be  isolated,  the  methods  of  the  preceding  section 
may  be  used  to  check  on  (2).  However,  if  this  is  not  possible,  the 
researcher  may  use  the  more  general  tests  suggested  by  Tukey  (32,  37) 
and  by  Ward  and  Dick  (38)  .  Rather  than  give  the  details  of  these  tests, 
we  refer  the  reader  to  the  original  publications.  If  access  to  these  pub 
lications  is  not  possible,  perhaps  the  illustrations  in  Snedecor  (31)  and 
Hamaker  (21)  will  suffice. 

I  ndepend  ence 

The  assumption  of  independence  or,  granting  normality,  of  un- 
correlated  errors  is  a  crucial  assumption  and  its  importance  should  not 
be  overlooked.  Of  course,  by  utilization  of  the  device  of  randomization, 
the  researcher  can  do  his  best  to  see  that  the  correlation  between  errors 
will  not  continually  favor  (or  hinder)  any  particular  treatment.  If  one 
wishes  to  test  for  randomness,  methods  are  available.  However,  since 
these  will  be  discussed  in  Chapters  15  and  16,  no  details  will  be  given 
at  this  time.  The  interested  reader  may  jump  ahead  to  the  appropriate 
sections.  (NOTE:  The  procedure  discussed  in  Section  11.6  may  also 
be  helpful  in  this  situation.) 

In  general,  the__consequences  are  not  serious  when  the  assumptions 
madeira  connection,  with  analyses  of  variance  are  not  strictly  satisfied. 
That  is,  moderate  departures  from  the  conditions  specified  by  the 
assumptions  need  not  alarm  us.  For  example,  minor  deviations  from 
normality  and/or  some  degree  of  heteroschedasticity  (lack  of  homo 
geneity  of  variances)  will  have  little  effect  on  the  usual  tests  and  the 
resulting  inferences.  In  summary,  the  analysis  of  variance  technique 
i^jpiite^obustr^|i(i,thus  the  researcher  can  rely  on,  its  doing  a  good  job. 
und^r  JDQSk  JSJECIITQ  qt  *vn  f*.^  However,  since  trouble  can  arise  because  of 
failure  of  the  data  to  conform  to  the  assumptions,  ways  of  handling 
such  situations  must  be  examined. 

When  some  action  is  needed  to  make  tlxe  data  conf  ormjto  the  jusual 

approach  is  to  transform  the  original  data 


inlsuchlTway  that  the  transformed  data  will  meet  the  conditions  specie 
fied  by  the  assumptions.  For  example,  if  the  true  effects  are  multipli 
cative  instead  of  additive,  it  is  customary  to  take  logarithms  and  thus 
change,  for  instance, 


Y  =  jjLcttffrs  (11.96) 

into 

Y'  =  log  Y  =  log  M  +  log  en  +  log  fy  +  log  €„.          (11.97) 

Fortunately,  in  most  cases,  one  transformation  will  suffice.  That  is,  it 
is  usually  not  necessary  to  make  a  series  of  transformations,  each  to 


346 


CHAPTER  11,  COMPLETELY  RANDOMISED  DESIGN 


correct  a  separate  "deficiency "  in  the  original  data.  The  reason  for 
this  fortunate  state  of  affairs  is  that,  in  general,  the  utilization  of  a 
transformation  to  correct  one  particular  deficiency  (say,  nonadditivity) 
will  also  help  with  respect  to  another  deficiency  (say,  nonnormality) . 
With  this  in  mind,  the  more  common  transformations  are  summarized 
in  Table  11.48.  Further  details  may  be  found  in  Bartlett  (3)  and 
Tukey  (32). 

TABLE   11.48-Some  Common  Transformations 


Transformation 

Conditions  Leading  to 
Its  Application 

Name 

Equation 

Logarithmic  .  .  .  . 

F'  =  log  Y 

1.  The  true  effects  are  multiplicative  (or 
proportional)  . 
or 
2.   The  standard  deviation  is  proportion 
al  to  the  mean. 

Square  root  .... 

F'=VF 
or 

The  variance  is  proportional  to   the 
mean  (e.g.,  when  the  original  data  are 
samples  from  a  Poisson  distribution). 

F'-VF+l 

Arcsine           .  .  . 

F'  =  arcslne  Vp 

The    variance    is    proportional    to    p, 
(1  —  ju)  as,  for  example,  when  the  orig 
inal  data  are  samples   (expressed  as 
proportions    or   relative    frequencies) 
from  binomial  populations. 

Reciprocal  

F'=1/F 

The    standard    deviation    is    propor 
tional  to  the  square  of  the  mean. 

Before  leaving  the  subject  matter  of  this  section,  one  other  technique 
for  handling  heterogeneous  variances  should  be  mentioned.  This  tech 
nique  is  as  follows:  Partition  the  experimental  error  sum  of  squares  in 
correspondence  with  any  partitioning  of  the  treatment  sum  of  squares. 
However,  this  technique,  valid  though  it  may  be,  is  seldom  employed 
because;  (1)  it  is  difficult  and  time-consuming  to  perform  and  (2)  each 
portion  of  Evy  will  usually  possess  a  very  small  number  of  degrees  of 
freedom  so  that  the  subsequent  F-tests  will  be  of  little  value  (i.e.,  they 
will  not  be  very  powerful  or  discriminating  tests).  Because  this  tech 
nique  is  used  so  rarely,  no  further  discussion  will  be  given  at  this  time. 
However,  an  example  of  subdividing  the  experimental  error  sum  of 
squares  will  be  presented  in  the  next  chapter. 

11.14     THE    RELATION     BETWEEN    ANALYSIS    OF    VARI 
ANCE  AND   REGRESSION   ANALYSIS 

Perhaps  the  most  concise  statement  that  can  be  made  concerning  the 
relation  between  analysis  of  variance  and  regression  analysis  is  the 


11.15       PRESENTATION    OF    RESULTS  341 

following:  Analysis  of  variance  and  regression  analysis  are  essentially 
the  same.  Why,  then,  have  we  spent  so  much  time  (and  we  are  not 
through  yet)  discussing  analysis  of  variance  as  a  separate  topic?  The 
answer  is :  Because  there  are  many  cases  (based  on  specific  conditions) 
that  are  more  easily  explained  using  the  methods  of  this  and  succeeding 
chapters  than  those  given  in  Chapter  8. 

Because  of  the  complexity  of  the  topic,  the  general  equivalence  of  the 
two  methods  (i.e.,  analysis  of  variance  and  regression  analysis)  will  not 
be  discussed  in  this  book.  The  interested  reader  is  referred  to  Graybill 
(20)  and  Kempt  home  (23)  for  a  general  discussion  of  the  basic  theory, 
and  to  Chew  (8)  for  some  illustrative  examples. 

11.15      PRESENTATION   OF    RESULTS 

Even  though  an  ANOVA  table  is  very  convenient  for  summarizing 
certain  aspects  of  the  analysis  of  a  set  of  data,  it  suffers  from  a  rather 
serious  deficiency,  namely,  that  it  tends  to  overemphasize  tests  of 
hypotheses  and  underemphasize  estimation.  Since  estimation  is  the 
more  important  of  these  two  aspects  of  statistical  inference,  this  could 
be  serious  if  steps  are  not  taken  to  remedy  the  situation.  Two  steps  that 
can  be  taken  to  improve  matters  are:  (1)  always  accompany  an 
ANOVA  table  with  tables  of  means,  together  with  their  standard  er 
rors,  and  (2)  "whenever  possible,  portray  the  results  in  graphical  form. 
If  these  two  steps  are  taken  and  if  a  readable  report  is  prepared,  the 
results  of  your  research  will  be  more  easily  understood  and  appreciated. 

Example  11.19 

Re-examination  of  Example  11.3  will  show  that  the  means  were 
given  in  Table  11.9,  the  ANOVA  in  Table  11.10,  and  the  standard  error 
of  the  mean  in  the  discussion.  Actually,  the  standard  error,  which  was 
the  same  for  each  mean  because  of  the  equal  sample  sizes,  might  better 
have  been  included  in  Table  11.9. 

Example  11.20 

Re-examination  of  Example  11.4  will  show  that  the  suggestion  made 
in  Example  11.19  was  adopted  in  that  case.  That  is,  the  standard  errors 
were  presented  along  with  the  means  to  which  they  applied. 

Example  11.21 

Re-examination  of  Example  11.6  will  show  that  the  ANOVA  was 
given  in  Table  11.18  and  the  standard  error  of  a  treatment  mean  was 
included  in  the  discussion.  However,  the  treatment  means  were  not 
explicitly  exhibited  although  they  could  easily  have  been  obtained.  Had 
a  complete  report  of  the  research  been  prepared,  this  deficiency  would 
have  been  noted  and  removed. 

Example  11.22 

Re-examination  of  Examples  11.11  and  11.12  will  show  that  standard 
errors  were  (implicitly)  found  for  each  of  the  selected  contrasts.  As 
noted  in  the  discussion  of  Example  11.12,  the  point  and  interval  esti- 


342 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


mates  of  the  true  effects  of  the  contrasts  could  then  be  calculated.  These 
would,  of  course,  be  included  in  the  research  report. 

Example  11.23 

Re-examination  of  Example  11.15  will  indicate  that  any  research 
report  concerning  this  experiment  would  have  benefited  by  a  graph 
showing  the  treatment  means  (average  yields)  as  a  function  of  the 
amount  of  fertilizer  applied  to  the  experimental  plots.  It  is  suggested 
that  the  reader  plot  these  means  and  examine  the  graph  in  connection 
with  the  recommendations  made  in  Example  11.15. 

Example  11.24 

Re-examination  of  Example  11.17  will  reveal  that  the  treatment 
means  were  not  given.  Since  they  are  pertinent  to  the  conclusions,  we 
give  them  in  Table  11.49.  The  standard  errors  of  the  treatment  means 

TABLE   11.49— Treatment  Means  for  the  Experiment  Discussed  in  Example 
11.17  (Data  in  Table  11.38;  ANOVA  in  Table  11.39) 


Date  of 

Planting 

Fertilizer 

Early 

Late 

Average 

Check    

32.68  (0.88) 

31.28  (0.88) 

31.98  (0.62) 

Aero  

29.50  (0.88) 

32.08  (0.88) 

30.79  (0.62) 

Na  

27.78  (0.88) 

32  .  48  (0  .  88) 

30.12  (0.62) 

K  

29.28  (0.88) 

31.40  (0.88) 

30.34(0.62) 

Average 

29.81  (0.44) 

31.81  (0.44) 

30.81 

The  figures  in  parentheses  in  tlie  table  are  the  standard  errors  of  the  means  to  which 
they  are  appended. 

shown  in  Table  11.49  were  calculated  by  taking  the  square  roots  of  the 
folio  wing  estimated  variances: 

V(Yi)    =  V  (date  of  planting  mean)  =  3.07/16  =  0.1919 

V(Y^    =-  V  (fertilizer  mean)  =  3.07/8  =  0.3838 

V(Yii)  =  V  (date  of  planting  X  fertilizer  mean)  =*  3.07/4  =  0.7675. 

A  graphical  presentation  of  the  means  is  given  in  Figure  11.1  where, 
of  course,  the  reader  must  realize  that  the  slopes  of  the  lines  are  a  direct 
reflection  of  the  scales  adopted.  However,  since  our  main  use  of  the 
graph  will  be  in  the  interpretation  of  the  interaction,  this  will  not  matter, 
for  we  shall  be  concerned  only  with  the  slopes  of  the  lines  relative  to 
one  another.  A  study  of  Figure  11.1  will  confirm  the  conclusions 
reached  in  Example  11.17,  namely:  (1)  the  late  date  of  planting  is 
apparently  better  than  the  early  date  of  planting,  (2)  there  is  little 
difference  among  the  main  effects  of  the  four  fertilizers,  and  (3)  there 
is  some  indication  of  a  possible  interaction.  (NOTE:  This  last  conclu 
sion  is  suggested  by  the  lack  of  "parallelism"  of  the  plotted  lines.) 

In  addition  to  the  remarks  made  in  the  first  paragraph  of  this  section 
and  illustrated  in  Examples  11.19  through  11.24,  the  reader  should 


11.15       PRESENTATION    OF    RESULTS 


343 


33 

LJ 

32 

01 
O 

31 

r^ 

30 

CD 

*~* 

29 

o 

_J 
UJ 

28 

>~ 

27 

EARLY 


CHECK  AERO  Na 

TYPE    OF    FERTILIZER 


K 


FIG.    1  1 .1—  Graphical    representation    of   the   mean 
yields    given    in    Table    1  1 .49. 


realize  that  many  experiments  are  conducted  and  analyses  of  variance 
performed  only  to  estimate  components  of  variance.  Important  as  this 
topic  is,  it  is  felt  that  the  discussion  given  earlier  in  the  chapter  will 
prove  sufficient  for  most  applications.  Should  further  details  be  desired, 
it  is  suggested  that  a  professional  statistician  be  consulted. 

One  other  topic  should  be  mentioned  in  connection  with  the  presen 
tation  of  results.  This  topic  is  concerned  with  the  general  way  in  which 
ANOVA's  are  commonly  presented.  Two  customs  have  become  quite 
firmly  established  over  the  years  and  they  are  as  follows: 

1.  (a)   If  an  .P-ratio  exceeds  the  95  per  cent  point  but  does  not 

exceed  the  99  per  cent  point,  the  F-ratio  (or  the  mean 
square  for  the  effect  being  tested)  is  tagged  with  a  single 
asterisk  (*). 

(b)  If  an  F-ratio  exceeds  the  99  per  cent  point,  the  F-ratio 
(or  the  mean  square  for  the  effect  being  tested)  is  tagged 
with  a  double  asterisk  (**) . 

2.  If  space  is  at  a  premium,  only  an  abbreviated  A1STOVA  will  be 
presented.  When  this  is  done,  it  is  customary  to  include  only 
the  columns  for:  (1)  sources  of  variation,  (2)  degrees  of  freedom, 
and  (3)  mean  squares. 

Incidentally,  when  the  asterisk  convention  is  used,  it  is  good  practice 
to  define  the  symbols  at  the  bottom  of  every  ANOVA  table  by  use  of 
the  following  footnotes: 

*  Significant  at  a  =  0.05. 
**  Significant  at  a.  =  0.01. 


344 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


The  use  of  these  customs  will  be  illustrated  in  succeeding  chapters. 

If  any  words  can  be  put  together  to  summarize  the  implications  of 
this  section,  they  are  as  follows:  Do  not  forget  the  reader.  Remember,  you 
are  writing  not  for  yourself  but  for  others.  Anything  you  can  do  to  make 
your  assumptions,  procedures,  results,  analyses,  and  conclusions  more 
understandable  will  add  to  the  value  of  your  research. 

Problems 

11.1  What  are  the  proper  objectives  of  analyses  of  variance  (using  experi 
mental  or  survey  data) ;  that  is,  for  what  purposes  may  we  properly 
use  analyses  of  variance? 

11.2  Forty  technicians  were  available  to  investigate  5  methods  of  deter 
mining  the  iron  content  of  a  certain  chemical  mixture.  Eight  of  the 
technicians  used  method  No.  1,  8  used  No.  2,  and  so  on.  The  assign 
ment  of  technicians  to  methods  was  performed  in  a  random  manner. 
Each  technician  made  only  one  determination.  Given  that:  (1)  the 
total  of  the  40  observations  was  80,  (2)  the  among  methods  mean 
square  was  6,  and  (3)  the  pooled  variance  among  technicians  within 
methods  was  8,  fill  in  the  following  ANOVA  table.  (NOTE:  omit  the 
spaces  marked  X.} 


Source  of  Variation 

Degrees 
of 
Freedom 

Sum 
of 
Squares 

Mean 
Square 

Expected 
Mean 
Square 

Mean  

X 

Among  methods  ,      T  ,  .  . 

Among  technicians 
within  methods  

Total 

X 

X 

11.3        Given  the  following  abbreviated  ANOVA: 


Source  of  Variation 

Degrees 
of  Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square 

Among  treatments  

4 

244 

61 

<r2  +  7  23  -£/4 

Among  experimental  units 
within  treatments  

30 

270 

9 

t~i 
cr2 

(ct)   Write  out  the  appropriate  model. 

(&)   State  the  null  hypothesis,  both  in  words  and  symbolically,  that 

the  experiment  was  probably  designed  to  test, 
(c)    Test  the  hypothesis  given  in  the  answer  to  (&)  using  a  probability 

of  Type  I  error  equal  to  .05. 


PROBLEMS  345 

11.4  A  process  is  designed  to  produce  a  fishline  that  will  have  a  "15-lb.- 
test"  rating.  The  braided  line  may  be  treated  with  4  different  water- 
proofings.  The  hypothesis  is  that  the  4  treatments  have  the  same, 
if  any,  effect  on  the  test  rating  of  the  cord.  Twenty  samples  of  each 
type  of  treated  cord  are  tested  for  breaking  strength.  Assuming  that 
analysis  of  variance  is  a  valid  technique  to  use  in  this  case,  set  up  the 
appropriate  table  showing  the  proper  subdivision  of  the  degrees  of 
freedom.  Discuss  any  further  analyses  that  might  be  useful  in  investi 
gating  the  treatments. 

11.5  It  is  desired  to  test  10  different  baking  temperatures  when  we  use  a 
standard  cake  mix.  Fifty  sample  batches  of  mix  are  prepared,  and  5 
are  assigned  at  random  to  each  of  the  10  temperatures.  Six  judges 
score  the  cakes,  and  the  average  score  is  recorded  for  each  cake.  Give 
the  proper  subdivision  of  the  degrees  of  freedom,  and  write  out  the 
mathematical  model  assumed.  State  the  hypothesis  to  be  tested.  Dis 
cuss  and  evaluate  the  method  of  analysis. 

11.6  Four  methods  of  performing  a  certain  operation  have  been  tried  and 
we  have  10  observations  for  each  method.  The  mean  productivities 
under  each  method  are  60,  70,  80,  and  90,  respectively.  Not  having 
the  original  data  from  which  to  calculate  the  sums  of  squares,  we 
assume  that  the  coefficient  of  variation  (square  root  of  the  pooled 
estimate  of  az  divided  by  the  average  of  all  observations)  is  0.1.  On 
this  assumption,  test  the  hypothesis  that  the  "method  population 
means"  are  equal. 

11.7  Given  that  the  means  of  10  individuals  in  each  of  5  groups  are  30,  32, 
34,  36,  and  38,  and  that  the  variance  of  a  group  mean  is  8,  compute 
the  analysis  of  variance. 

11.8  An  investigation  to  study  the  variation  in  average  daily  gains  made 
by  pigs  among  and  within  litters  when  fed  the  same  ration  gave  the 
following  results: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Among  litters  

29 

0.0576 

Among  pigs  in  the  same  litter  

180 

0.0144 

How  would  you  use  this  information  to  design  experiments  to  test  the 
effects  of  different  rations  on  average  daily  gains? 

11.9  Community  X  and  community  Y  are  two  neighboring  small  towns. 
Community  X  is  supplied  with  electricity  by  a  private  power  com 
pany,  while  community  Y  operates  a  municipally  owned  but  ineffi 
cient  high-cost  power  plant.  As  a  result,  cost  of  electricity  to  home- 
users  is  higher  in  community  Y  than  in  community  X]  for  example, 
the  charge  for  the  first  50  watts  is  $3.00  in  X  and  $4.50  in  Y.  A  ran 
dom  sample  of  household  meter  readings  for  the  same  month  was 
taken  in  each  community.  The  following  values  were  obtained: 


346 


CHAPTER   11,   COMPLETELY   RANDOMIZED    DESIGN 


11.10 


X  Sample 
(16  Observations) 
(Kw-hours  used) 

F  Sample 
(21  Observations) 
(Kw-hours  used) 

28                      16 

6                      12 

12                     22 

36                      18 

14                     28 

24                      18 

4                       4 

58                      16 

16                     22 

60                      22 

28                       4 

6                      14 

30                     34 

14                      16 

76                     30 

54                     26 

22                      44 

16                      58 

18 

Analyze  these  data  in  two  ways  : 

(a)  Compare  home  consumption  of  electricity  in  the  two  communities 
by  means  of  the  comparison  of  two  groups  using  "Student's" 


(&)  Prepare  an  analysis  of  variance  of  these  two  samples. 
It  is  suspected  that  five  filling  machines  in  a   certain  plant  are 
filling  cans  to  different  levels.  Random  samples  of  the  production 
from  each  machine  were  taken,  with  the  following  results: 


Machine 

A 

B 

C 

D 

E 

11.95 

12.18 

12.16 

12.25 

12.10 

12.00 

12.11 

12.15 

12.30 

12.04 

12.25 

12.08 

12.10 

12.02 

12.10 

12.02 

Analyze  the  data  and  state  your  conclusions. 

11.11  The  amount  of  carbon  used  in  the  manufacture  of  steel  is  assumed  to 
have  an  effect  on  the  tensile  strength  of  the  steel.  Given  the  following 
data,  perform  the  appropriate  analysis  and  interpret  your  results. 
The  tensile  strengths  of  six  specimens  of  steel  for  each  of  three  dif 
ferent  percentages  of  carbon  are  shown.  (The  data  have  been  coded 
for  easy  calculation.) 


PROBLEMS 


347 


Percentage  of  Carbon 


0.10 


0.20 


0.30 


23 

42 

47 

36 

26 

43 

31 

47 

43 

33 

34 

39 

31 

37 

42 

31 

31 

35 

11.12  A  public  utility  company  has  a  stock  of  voltmeters  which  are  used 
interchangeably  by  the  employees.  The  question  arises  as  to  whether 
all  the  voltmeters  are  homogeneous.  Since  it  would  be  too  expensive 
to  check  all  the  meters,  a  random  sample  of  6  meters  is  obtained  and 
all  6  are  read  three  times  while  being  subjected  to  a  constant  voltage. 
The  following  data,  expressed  as  deviations  from  the  test  voltage, 
were  recorded.  Analyze  and  interpret. 

Meter 


0.95 

0.33 

—  2.15 

—  1.20 

1.80 

—  1.05 

1.06 

—  1.46 

1.70 

0.62 

0.88 

—  0.65 

1.96 

0.20 

0.48 

1.50 

0.20 

0.80 

11.13  An  experiment  had  for  its  objective  the  evaluation  of  variance  com 
ponents  for  the  variation  in  ascorbic  acid  concentration  (mg.  per 
100  g.)  in  turnip  greens.  Two  leaves  were  taken  from  near  the  center 
of  each  of  5  plants.  Ascorbic  acid  concentration  was  determined  for 
each  leaf.  This  was  repeated  on  each  of  6  days,  a  new  selection  of 
plants  being  obtained  each  day.  The  following  data  were  collected: 


Day 

Leaf 

Plant 

1 

2 

3 

4 

5 

1 

A 

9.1 

7.3 

7.3 

10.7 

7.7 

B 

7.3 

9.0 

8.9 

12.7 

9.4 

2 

A 

12.6 

9.1 

10.9 

8.0 

8.9 

B 

14.5 

10.8 

12.8 

9.8 

10.7 

3 

A 

7.3 

6.6 

5.2 

5.3 

6.7 

B 

9.0 

8.4 

6.9 

6.8 

8.3 

4 

A 

6.0 

8.0 

6.8 

9.1 

8.4 

B 

7.4 

9.7 

8.6 

11.2 

10.3 

5 

A 

10.8 

9.3 

7.3 

9.3 

10.4 

B 

12.5 

11.0 

8.9 

11.2 

12.0 

6 

A 

10.6 

10.9 

10.4 

13.1 

7.7 

B 

12.3 

12.8 

12.1 

14.6 

9.4 

348 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


11.14 


Plants,  days,  and  leaves  are  to  be  considered  as  random  variables 
(there  might  be  some  question  about  days).  Calculate  the  analysis  of 
variance  and  evaluate  the  variance  components  for  leaves  of  the  same 
plant,  plants  of  the  same  day,  and  days. 
Suppose  we  have  the  mathematical  model 

Yak  «  M  +  n  +  ei3  +  Si,*         i  =  1,  2,  3,  4 

j  -  1,  2 
k  «  1,  2 

where  n  is  the  true  effect  of  the  fth  treatment,  €#  is  the  effect  of  the 
jth  experimental  unit  subjected  to  the  ith  treatment,  and  5t-y&  is  the 
Ath  determination  on  the  (ij)  th  experimental  unit.  We  wish  to  test  the 
hypothesis  H:ri  =  0  for  all  i.  The  following  values  are  known: 

3Ti  =     8  .En  «  3  £12  =  5 

T2  =     7  JS2i  =  3  £22  =  4 

T3  =  10  £31  =  2  £32  =  8 

2%  =     7  ^41  =  5  ^42  =  2 


11.15 


and  ]Cy2  =  18.  Complete  the  appropriate  analysis  of  variance,  test 
the  hypothesis,  and  interpret  your  results. 

Given  the  following  abbreviated  ANOVA  of  data  collected  from  an 
experiment  involving  6  treatments,  10  experimental  units  per  treat 
ment,  and  3  determinations  per  experimental  unit  : 


Degrees  of 

Mean 

Source  of  Variation 

Freedom 

Square 

Expected 

Mean  Square 

Treatments  ,  ,  ,  . 

5 

12,489 

O"g    —f~    3o" 

"+^±r? 

5  tZ 

Exp.  units  within  treatments.  .  .  . 

54 

3,339 

2      _!_   1 
<T$      -j-   OCT 

2 

Det   per  experimental  unit 

120 

627 

2 
(TK 

11.16 


(a)  Write  out  the  model  assumed,  stating  explicitly  what  each  term 
represents. 

(6)  Test  the  hypothesis  that  the  6  treatments  have  the  same  popula 
tion  mean. 

(c)  Compute  the  variance  of  a  treatment  mean. 

(d)  Given  that  th.6  sample  mean  for  treatment  No.  3  is  193.7,  com 
pute  and  interpret  the  95  per  cent  confidence  interval  for  esti 
mating  the  true  population  mean  of  treatment  No.  3. 

(e)  Assuming  that  the  estimates  of  the  components  of  variance  would 
remain  unchanged,  would  it  be  more  or  less  efficient  to  use  9  ex 
perimental  units  per  treatment  and  4  determinations  per  experi 
mental  unit?  Show  all  calculations  necessary  to  support  your  an 
swer.  What  is  the  gain  or  loss  in  information? 

We  conducted  a  completely  randomized  experiment  to  study  some 
chemical  characteristics  of  5  varieties  of  oats.  We  assigned  each  variety 
at  random  to  6  plots,  making  a  total  of  30  plots.  Instead  of  harvesting 


PROBLEMS 


349 


11.17 


the  entire  plot,  we  selected  at  random  8  3-by-3-foot  samples  from 
each  plot.  For  each  sample  we  made  3  chemical  determinations. 
Indicate  the  proper  complete  subdivision  of  the  total  degrees  of  free 
dom.  Give  the  expected  mean  square  for  each  source  of  variation. 
Indicate  the  proper  F-test  to  test  the  hypothesis  that  the  population 
means  for  the  5  varieties  are  equal. 
Given  the  following  abbreviated  ANOVA: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Groups  

3 

600 

Experimental  units  within  groups  

36 

120 

[Determinations  per  experimental  unit.  .  . 

80 

12 

(a)  Give  the  expected  mean  squares,  assuming  that  we  are  interested 
in  just  these  groups  but  that  experimental  units  and  determina 
tions  are  random  variables. 

(6)  Test  the  hypothesis  that  the  group  population  means  are  equal. 
Interpret  your  result. 

(c)    Compute  the  variance  of  a  group  mean  (per  determination) . 
11.18      Given  the  following  abbreviated  ANOVA: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Among  treatments     

4 

20 

Experimental  units  within  treatments  .  .  . 
Determinations  per  experimental  unit.  .  .  . 

15 
20 

15 
4 

Obtain  estimates  of  all  the  components  of  variance,  and  interpret 
each  in  terms  of  the  model  y;yfc==M+'7~i+£;j+Siyfc  stating  explicitly 
all  assumptions  that  you  make. 
11.19      Given  the  following  abbreviated  ANOVA: 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 

Square 

Expected  Mean 
Square 

Treatments  

3 

1800 

600 

0-!+3o-2+30ov 

Experimental  units 
within  treatments  .... 
Determinations  per 
experimental  unit  .... 

36 
80 

3600 
960 

100 
12 

<r2s+3<r2 
<r! 

(a)   Compute  the  variance  of  a  treatment  mean. 

(6)   Test  the  null  hypotheses  jff:er?  =  0,  and  interpret  your  answer. 

(c)  The  sample  mean  of  treatment  No.  1  is  given  to  be  80.  Compute 
a  95  per  cent  confidence  interval  for  estimating  the  true  popula 
tion  mean  of -treatment  No,  1. 


350  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

11.20      Given,  the  following  abbreviated  ANOVA: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Treatments  

4 

960 

Experimental  units  within  treatments  
determinations  per  experimental  unit  

35 
40 

320 
20 

11.21 


(a)   Test  the  hypothesis  that  the  population  treatment  means  are  all 

equal.  Interpret  your  answer. 
(6)    Give  the  expected  mean  squares  in  the  above  analysis  of  variance. 

(c)  Compute  the  variance  of  a  treatment  mean, 

(d)  Estimate  the  gain  or  loss  in  information  if  the  above  experiment 
were  to  be  repeated  with  10  experimental  units  per  treatment  and 
a   single   determination   per   experimental   unit.    State    all   your 
assumptions. 

Given  the  following  abbreviated  ANOVA: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Among  treatments    

9 

570 

Experimental  units  within  treatments  
Among    determinations    on    same    experi 
mental  unit  

90 
200 

190 
10 

(a) 


(6) 

GO 


11.22 


11.23 


11.24 


Compute  F  to  test  the  hypothesis  that  the  10  treatments  have 
the  same  true  effect. 

Compute  the  variance  of  a  treatment  mean  per  determination  in 
the  above  experiment. 

Assuming  that  the  estimates  of  the  components  of  variance 
would  not  change,  estimate  the  gain  or  loss  in  information  in 
estimating  the  treatment  means  if  20  experimental  units  per 
treatment  were  selected  with  a  single  determination  on  each 
experimental  unit  in  repeating  the  experiment. 

The  cities  and  towns  of  Arizona  have  been  allocated  to  5  strata  (or 
groups)  according  to  population.  In  each  stratum  we  select  at  random 
10  cities  (or  towns) ;  in  each  of  these  cities  we  select  at  random  4 
blocks;  and  in  each  of  these  blocks  we  select  at  random  2  households. 
Indicate  the  proper  subdivision  of  the  degrees  of  freedom  for  a  com 
plete  analysis  of  variance  of  some  item  such  as  the  average  income 
of  the  head  of  each  household. 

Set  up  the  analysis  of  variance  table  and  show  the  degrees  of  freedom 
for  the  following  experiment:  Six  spray  treatments  are  applied  com 
pletely  at  random  in  an  orchard  of  100  trees  (all  being  used).  Each 
treatment  is  applied  to  sets  of  2  trees;  then  the  yield  of  each  tree  is 
estimated  by  obtaining  4  samples  around  the  perimeter.  Note  that 
all  but  2  treatments  contain  8  sets;  the  remaining  2  treatments  con 
tain  9  sets.  Show  the  expected  mean  squares. 

The  following  abbreviated  analysis  of  variance  was  prepared  from 
chemical  determinations  made  on  samples  of  a  legume  hay.  The  hay 


PROBLEMS 


351 


samples  were  obtained  from  a  completely  randomized  experiment 
involving  16  treatments. 

ANALYSIS  OP  VARIANCE  OF  CHEMICAL  DETERMINATIONS 

SAMPLES 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Treatments  

15 

550 

Plots  with,  same  treatment 

220 

Samples  from  plots  treated  alike.  .  . 

256 

20 

11.25 


11.26 


(a) 


State  a  suitable  hypothesis  about  these  treatments,  and  make  the 
proper  test.  The  experimenter  wished  to  reject  the  hypothesis 
only  under  the  condition  of  a  1  per  cent  chance  of  a  Type  I  error. 
What  is  your  conclusion? 

Indicate  the  function  of  replication  in  this  experiment  for  study 
ing  the  effects  of  various  treatments  on  a  legume  hay. 
How  might  the  replication  and  sampling  procedure  for  this  ex 
periment  be  changed  in  order  to  increase  replication  without 
changing  the  total  number  of  samples  to  be  analyzed? 
Estimate  the  possible  gain  in  relative  efficiency  for  your  proposed 
change  in  the  experiment. 

(e)  What  assumption  is  required  for  making  this  calculation? 
In  an  effort  to  develop  objective  methods  of  estimating  the  yield  of 
corn,  an  experimental  survey  was  conducted  in  a  district  of  central 
Iowa.  A  random  selection  of  fields  was  made,  and  within  those  fields 
2  sampling  units  (consisting  of  10  hills  each)  were  selected  at  random 
and  the  grain  yield  determined  by  harvesting  and  weighing.  The 
analysis  of  variance  (on  a  10-hill  s.u.  basis)  is  as  follows: 


(6) 
(c) 

(d) 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 

Square 

Fields  

47 

2098.7 

44.65 

S.u.'s  within  fields.  .  .  . 

48 

554.5 

11.55 

How  much  information  would  have  been  lost  if  only  one  s.u.  per  field 
had  been  taken? 

IData  from  a  sample  survey  of  farms  in  the  Midwest  were  to  be  sum 
marized  by  means  of  the  analysis  of  variance.  Eight  types  of  farming 
areas  were  included  in  the  study,  and  within  each  area  5  counties 
were  selected  at  random.  Within  each  of  the  chosen  counties,  20 
farms  were  selected  at  random  and  farm  management  records  taken 
for  each.  A  partial  list  of  the  summary  calculations  was  as  follows  for 
the  item  "farm  income'7: 

Total  corrected  sum  of  squares  =  8,183,000 
Sum  of  squares  for  among  counties  within  areas  =  352,000 
Mean  square  for  type  of  farming  areas  ==33, 000. 

(a)   Prepare  and  complete  an  analysis  of  variance  for  "farm  income" 
from  the  above  information. 


352 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


11.27 


(6)    What  is  the  variance  of  a  type-of-farming  area  mean  as  deter 
mined  from  your  analysis  of  variance? 

(c)    Suppose  it  had  been  decided  to  select  only  2  counties  in  each 
area  and  sample  50  farms  in  each  county.  What  is  the  relative 
efficiency  of  the  plan  used  to  the  procedure  suggested  here? 
(<2)   The  type-of-farming  areas  included  in  the  study  were  arbitrarily 
selected  upon  the  basis  of  some  known  differences.  Are  the  areas 
different  with  respect  to  "farm  income?" 
(e)    Write   out   the   expected   values    of   the    mean   squares   used   in 

ans wering  (d)  above. 

Given  the  following  abbreviated  analysis  of  variance  of  the  data  from 
a  completely  randomized  experiment  with  4  treatments,  8  experi 
mental  units  per  treatment,  3  samples  per  experimental  unit,  and  2 
determinations  (of  some  chemical  or  physical  characteristic)  per 
sample : 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Treatments  

3 

19,200 

Among  experimental  units  treated  alike  .... 

28 

4  800 

Among  samples  per  experimental  unit  

64 

2,400 

Hetwe^ri  det^rrni  -nations  pfvr  sample            ,  , 

96 

1,200 

Estimate  the  gain  or  loss  in  efficiency  in  estimating  the  treatment 
effects  if  we  had  used  12  experimental  units  per  treatment,  2  samples 
per  experimental  unit  and  1  determination  per  sample. 

11.28  Describe  the  assumptions  underlying  the  application  of  the  analysis 
of  variance  technique. 

(a)  Which  of  these  assumptions  can  the  research  worker  check  for 
any  particular  analysis? 

(&)  Which  assumption  can  be  fulfilled  by  the  research  worker  in  an 
experimental  situation  by  appropriate  procedures? 

(c)  For  what  purposes  do  we  employ  the  analysis  of  variance  tech 
nique? 

(c£)  What  criterion  should  be  applied  for  judging  the  validity  of  an 
^-ratio  obtained  from  an  analysis  of  variance? 

11.29  (a)   Explain  in  your  own  words  the  meaning  in  the  analysis  of  vari 

ance  of  (1)  a  variance  component  and  (2)  a  fixed  effect? 
(6)    Consider  the  following  abbreviated  analysis  of  variance: 

ANALYSIS  OF  VARIANCE  or  CALORIES  CONSUMED  IN  ONE  DAY  FOR  A 
SAMPLE  OF  IOWA  WOMEN  OVER  THE  AGE  OF  30 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Among  zones  

2 

16,960,000 

8  480  000 

Among  counties  in  zones  .  . 

97 

41    128  000 

424  000 

Between  segments  in  counties  in 
zones  

100 

40  ,  000  ,  000 

400  OOO 

Among  individuals  in  segments 
in  counties  in.  zones  

600 

180   000  000 

300   000 

PROBLEMS  353 

For  this  problem  we  shall  assume  that  4  individuals  (women  over 
30)  were  interviewed  in  each  segment  and  that  2  segments  were  se 
lected  at^  random  in  each  county.  The  zones  are  open  country,  rural 
community,  and  urban.  Counties  appearing  in  the  sample  for  the 
zones  were  50,  25,  and  25,  respectively,  yielding  the  97  degrees  of 
freedom  for  counties  in  zones. 

(1)  Estimate  the  variance  components  for  individuals,  segments,  and 
counties  from  this  analysis  of  variance. 

(2)  Test  the  hypothesis:  Calories  consumed  on  this  day  are  the  same 
for  all  zones.  Show  the  mean  squares  used  in  forming  the  ^-ratio, 
and  indicate  in  general  why  these  are  the  proper  mean  squares  to 
use  for  the  test. 

11.30      With  reference  to  Example  11.9  and  Table  11.23,  show  that  the  ex 
pected  mean  squares  for  the  three  comparisons  are: 

Efca  -h 


n4) 


n* 


11.31      With  reference  to  Example  11.11  and  Table  11.25,  show  that  the  ex 
pected  mean  squares  for  the  four  comparisons  are: 


11.32  Given  the  additional  information  that,  in  Problem  11.10,  machine 
A  is  a  standard  machine  and  machines  JS,  (7,  D,  and  E  are  experi 
mental  models,  modify  the  original  analysis  to  assess  the  relative  per 
formance  of  the  five  machines. 

11.33  Apply  the  technique  of  Section  11.10,  to  the  following  problems: 

(«)    11-9  (d)    11.12 

(&)    11.10  (e)    11.13 

(c)    11.11  (/)    11.14 

11.34  If  you  did  not  use  the  technique  of  Section  11.11  in  the  analysis  of 
Problem  11,11,  please  do  so  now. 

11.35  It  is  suspected  that  the  age  of  a  furnace  used  in  curing  silicon  wafers 
influences  the  percentage  of  defective  items  produced.  An  experi 
ment  was  conducted  using  four  different  furnaces  and  the  data  given 
below  were  obtained.  Analyze  and  interpret  the  data. 


354  CHAPTER    IT,    COMPLETELY    RANDOMIZED    DESIGN 

PERCENTAGE  OP  GOOD  WAFERS  IN  8  EXPERIMENTAL  TRIALS 

PER  FURNACE  (THE  SAME  NUMBER  OF  WAFERS  WERE 

USED  IN  EACH  FURNACE  IN  EACH  TRIAL) 


Furnace 

A  (age  1  year) 

B  (age  2  years) 

C  (age  3  years) 

D  (age  4  years) 

95 

95 

80 

70 

92 

85 

80 

65 

92 

92 

82 

70 

90 

83 

78 

72 

92 

83 

77 

72 

94 

88 

75 

66 

92 

89 

78 

50 

91 

90 

78 

66 

11.36  It  is  suspected  that  tlie  environmental  temperature  in  which  batter 
ies  are  activated  affects  their  activated  life.  Thirty  homogeneous 
batteries  were  tested,  6  at  each  of  five  temperatures,  and  the  data 
shown  below  were  obtained.  Analyze  and  interpret  the  data. 

ACTIVATED  LIEE  IN  SECONDS 


Temperature  (°C.) 

0 

25 

50 

75 

100 

55 

60 

70 

72 

65 

55 

61 

72 

72 

66 

57 

60 

73 

72 

60 

54 

60 

68 

70 

64 

54 

60 

77 

68 

65 

56 

60 

77 

69 

65 

11.37  It  is  suspected  that  both  the  machine  on  which  bearings  are  produced 
and  the  operator  of  the  machine  influence  the  critical  dimension, 
namely,  the  inside  diameter.  To  check  on  this,  the  data  given  below 
were  obtained  under  normal  production  conditions.  Analyze  and 
interpret  the  data. 


PROBLEMS 

INSIDE  DIAMETERS  OF  BEARINGS  (IN  INCHES) 


355 


Machine 

1 

2 

3 

Operator 

A 

B 

C 

D                      J3 

1.02 

1.03 

1.05 

1.03 

1.02 

1.03 

1.03 

1.06 

1.03 

1.03 

1,02 

1.03 

1.04 

1.02 

1.04 

1.03 

1.06 

1.02 

1.07 

1.02 

1.06 

1.05 

11.38  An  experiment  was  conducted  to  assess  the  effects  of  temperature 
and  humidity  on  the  effective  resistance  of  a  standard  type  of  re 
sistor.  The  following  data  were  obtained.  Analyze  and  interpret  the 
data. 

CODED  RESISTANCE  VALUES 


Temperature 

—  20° 

F. 

70°F. 

160°F. 

Humidity 

10% 

50% 

10% 

50% 

10% 

50% 

23 

24 

26 

24 

25 

27 

24 

24 

25 

25 

26 

26 

25 

25 

26 

26 

26 

28 

24 

26 

26 

26 

28 

28 

11.39      Given  the  following  abbreviated  analysis  of  variance: 


ANALYSIS   OP  VARIANCE  or  NET  INCOME  PER  CROP  ACRE 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Between  soil  areas      

4 

625 

Soil  conservation  programs  .... 
SAXSCP  

3 
12 

400 
225 

Between  farms  in  subclasses  .  .  . 

80 

100 

(a)  Assuming  both  of  the  main  classifications  are  fixed  effects,  indi 
cate  the  appropriate  /^-ratios  for  tests  of  the  hypotheses:  (1)  soil 
areas  do  not  differ  in  income;  (2)  the  soil  conservation  programs 
have  no  effect  on  income  per  crop  acre. 

(&)  Assuming  that  soil  areas  were  selected  at  random  and,  also,  soil 
conservation  programs  were  selected  at  random  from  a  larger 
number  (not  entirely  realistic,  but  possible),  indicate  the  F-ratios 
for  the  tests  listed  in  (a)  above. 


356 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


11.40  In  the  accompanying  table  pasture-acres  per  farm  for  a  sample  of  36 
farms  in  Audubon  County,  Iowa,  for  the  year  1934  are  presented. 
The  sample  consists  of  3  farms  from  each  soil-tenure  grouping;  there 
are  12  such  groups.  It  is  expected  that  there  may  be  some  interaction 
effects  among  the  soil  and  tenure  classes.  We  shall  consider  the  tenure 
grouping  as  a  fixed  effect  and  the  soil  groupings  as  random  sampling 
from  a  larger  population.  Perform  the  necessary  calculations,  set  up 
the  analysis  of  variance  table,  and  discuss  the  results.  Also  examine 
the  homogeneity  of  variance  in  the  tenure  groups  by  the  Bartlett  test. 

DATA  ON  PASTURE  ACREAGE,  AUDUBON  COUNTY,  IOWA,  1934 


Tenure  Group 

Soil  Group 

I 

II 

III 

IV 

Farm 
Owners      1  

37.0 
40.1 
57.0 

36.0 
52.0 
38.0 

72.6 
65,2 
71.0 

(Pasture  acr 
50.0 
28.6 
37.2 

42.0 

54.5 
58.0 

54.0 
58.0 
29.0 

es  per  farm) 
49.0 
43.7 
27.0 

50.9 
34.0 
43.8 

67.4 
32.5 
43.8 

56.0 
69.0 

54.7 

55.0 
41.0 
54.6 

63.0 
45.0 
60.0 

2  

3  

Tenants     1 

2 

3.     . 

Mixed        1  

2  

3  

11.41  Five  varieties  and  4  fertilizers  were  tested.  From  each  experimental 
plot  3  quadrats  were  selected  at  random  and  their  yields  recorded  as 
follows : 


Varieties 


Fertilizers 

1 

2 

3 

4 

5 

57 

26 

39 

23 

48 

1 

46 

38 

39 

36 

35 

28 

20 

43 

18 

48 

67 

44 

57 

74 

61 

2 

72 

68 

61 

47 

60 

66 

64 

61 

69 

75 

95 

92 

91 

98 

78 

3 

90 

89 

82 

85 

89 

89 

99 

98 

85 

95 

92 

96 

98 

99 

99 

4 

88 

95 

93 

90 

98 

99 

99 

98 

98 

99 

PROBLEMS 


35Z 


11.42 


(a)   Construct  an  analysis  of  variance  table. 

(fe)    On  the  basis  of  the  appropriate  model,  write  the  expected  mean 
squares  conforming  to  the  following  assumptions: 

(1)  varieties  and  fertilizers  random  selections; 

(2)  varieties  and  fertilizers  both  given  sets; 

(3)  varieties  a  random  selection — fertilizers  a  given  set. 

(c)  Test  the  hypothesis  of  equal  variety  means.  Test  the  hypothesis 
of  equal  fertilizer  means. 

(d)  Construct  a  table  showing  the  means  and  their  standard  errors. 

(e)  What  conclusions  do  you  reach  as  a  result  of  this  experiment? 

A  building  superintendent  wishes  to  compare  the  relative  perform 
ance  ratings  of  various  combinations  of  floor  wax  and  length  of 
polishing  time.  Three  waxes  are  to  be  investigated  along  with  3 
polishing  times.  Eighteen  homogeneous  floor  areas  are  selected  and 
2  are  assigned  at  random  to  each  of  the  9  treatment  combinations. 
Analyze  and  evaluate  the  following  data. 


PERFORMANCE  RATINGS 
(HIGH  is  BETTER  THAN  Low) 


Wax 

A 

B 

C 

Polishing  time 
(in  minutes) 

15 

30 

45 

15 

30 

45 

15 

30 

45 

7 

7.5 

8.2 

7 

7.2 

7.1 

8 

9.2 

9.6 

8 

7.4 

8.6 

7 

7.6 

7 

8 

9.4 

9.5 

11.43  An  experiment  was  performed  to  assess  the  effects  of  type  of  material 
and  heat  treatment  on  the  abrasive  wear  of  bearings.  Two  bearings 
were  tested  at  each  of  10  treatment  combinations.  Analyze  and 
interpret  the  following  data. 


AMOUNT  OF  WEAR  (CODED  DATA) 


Material 

A 

B 

C 

D 

R 

Heat 

treatment* 

O        M 

O        M 

O        M 

O        M 

O        M 

23        30 
25        31 

42        45 
44        50 

37        39 

38        39 

41        44 

42        49 

20        24 
25        30 

*  0  =  oven  dried;  M— moisture  saturated. 

11.44  From  each  of  5  lots  of  insulating  material,  10  lengthwise  specimens 
and  10  crosswise  specimens  are  cut.  The  following  table  gives  the 
impact  strength  in  foot-pounds  from  tests  on  the  specimens.  Analyze 
and  interpret  the  data. 


358 


CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 


Lot  Number 

Type  of  Cut 

I 

II 

III 

IV 

V 

1.15 

1.16 

0.79 

0.96 

0.49 

0.84 

0.85 

0.68 

0.82 

0.61 

0.88 

1.00 

0.64 

0.98 

0.59 

0.91 

1.08 

0.72 

0.93 

0.51 

Lengthwise 

0.86 

0.80 

0.63 

0.81 

0.53 

specimens 

0.88 

1.01 

0.59 

0.79 

0.72 

0.92 

1.14 

0.81 

0.79 

0.67 

0.87 

0.87 

0.65 

0.86 

0.47 

0.93 

0.97 

0.64 

0.84 

0.44 

0.95 

1.09 

0.75 

0.92 

0.48 

0.89 

0.86 

0.52 

0.86 

0.52 

0.69 

1.17 

0.52 

1.06 

0.53 

0.46 

1.18 

0.80 

0.81 

0.47 

0.85 

1.32 

0.64 

0.97 

0.47 

Crosswise 

0.73 

1,03 

0.63 

0.90 

0.57 

specimens 

0.67 

0.84 

0.58 

0.93 

0.54 

0.78 

0.89 

0.65 

0.87 

0.56 

0.77 

0.84 

0.60 

0.88 

0.55 

0.80 

1.03 

0.71 

0.89 

0.45 

0.79 

1.06 

0.59 

0.82 

0.60 

1 1 .45  Five  batches  of  ground  meat  are  charged  consecutively  into  a  rotary 
filling  machine  for  packing  into  cans.  The  machine  has  6  filling 
cylinders.  Three  filled  cans  are  taken  from  each  cylinder  at  random 
while  each  batch  is  being  run.  The  coded  weights  of  the  filled  cans 
are  given  below.  Analyze  and  interpret  the  data. 


PROBLEMS 


359 


Cylinder 

Batch 

1 

2 

3 

4 

5 

1 

1 

4 

6 

3 

1 

1 

3 

3 

1 

3 

2 

5 

7 

3 

3 

2 

—  1 

—  2 

3 

2 

1 

3 

1 

1 

0 

0 

—  1 

0 

5 

1 

1 

3 

1 

2 

2 

1 

3 

1 

0 

4 

3 

3 

1 

1 

3 

3 

3 

4 

—  2 

—  2 

3 

0 

0 

3 

0 

3 

0 

1 

0 

1 

4 

2 

1 

5 

1 

2 

0 

1 

—  2 

1 

1 

1 

0 

3 

•* 

5 

2 

_____  i 

1 

6 

0 

0 

3 

3 

3 

1 

0 

3 

0 

1 

1 

3 

4 

2 

2 

11.46  The  following  data  on  the  density  of  small  bricks  resulted  from  an 
experiment  involving  3  different  sizes  of  powder  particles,  3  pres 
sures,  and  3  temperatures  of  firing.  The  27  combinations  of  this 
3X3X3  factorial  were  run  in  duplicate.  Analyze  and  interpret  the 
the  following  coded  data. 


Size 

Pressure 

Temperature 

1900 

2000 

2300 

5-10 

5.0 
12.5 
20.0 

340  375 
388  370 
378  378 

316  386 
338  214 
348  378 

374  350 
334  366 
380  398 

10-15 

5.0 
12.5 
20.0 

260  244 
322   342 
330  298 

388  304 
300  420 
260  366 

266  234 
234  258 
350  284 

15-20 

5.0 
12,5 
20.0 

134  140 
186   30 
40  210 

146   194 
412  428 
436  490 

152  212 
194  208 
230  254 

360 


CHAPTER    11r    COMPLETELY    RANDOMIZED    DESIGN 


11.47  During  the  manufacture  of  sheets  of  building  material  the  perme 
ability  was  determined  for  3  sheets  from  each  of  3  machines  on  each 
day.  The  table  below  gives  the  logarithms  of  the  permeability  in 
seconds  for  sheets  selected  from  the  3  machines  during  a  production 
period  of  9  days.  The  3  machines  received  their  raw  materials  from  a 
common  store.  Analyze  and  interpret  the  data. 


Day 

Machine 

Log  of  Permeability 

1 

1 
2 
3 

1,404 
1.306 
1.932 

1.346 
1.628 
1.674 

1.618 
1.410 
1.399 

2 

1 
2 
3 

1.447 
1.241 
1.426 

1.569 
1.185 
1.768 

1.820 
1.516 
1.859 

3 

1 
2 
3 

1.914 
1.506 
1.382 

1.477 
1.575 
1.690 

1.894 
1.649 
1.361 

4 

1 
2 
3 

1.887 
1.673 
1.721 

1.485 
1.372 
1.528 

1,392 
1.114 
1.371 

5 

1 
2 
3 

1.772 
1.227 
1.320 

1.728 
1.397 
1.489 

1.545 
1.531 
1.336 

6 

1 
2 

3 

1.665 
1.404 
1.633 

1.539 
1.452 
1.612 

1.690 
1.627 
1.359 

7 

1 
2 

3 

1.918 
1.229 
1.328 

1.931 
1.508 
1.802 

2.129 
1.436 
1.385 

8 

1 
2 

3 

1.845 
1.583 
1.689 

1.790 
1.627 
2.248 

2.042 
1.282 
1.795 

9 

1 

2 

3 

1.540 
1.636 
1.703 

1.428 
1.067 
1.370 

1.704 
1.384 
1.839 

References  and  Further  Reading 

1.  Anderson,   H.  L.,   and  Houseman,   E.   E.   Tables  of  orthogonal  polynomial 
values  extended  to  2NT=  104.  Res.  Bui.  297,  Agr.  Exp.  Sta.,  Iowa  State  Univ., 
Ames,  April,  1942. 

2.  Anscombe,  F.  J.,  and  Tukey,  J.  W.  The  analysis  of  residuals.  Unpublished 


REFERENCES   AND    FURTHER    READING  361 

handout,    Gordon    Conference    on    Statistics    in    Chemistry    and    Chemical 
Engineering,  New  Hampton,  N.  H.,  July,  1957, 

3.  Bartlett,  M.  S.  The  use  of  transformations.  Biometrics,  3:39,  1947. 

4.  Bechhofer,  R.  E.  A  sequential  multiple-decision  procedure  for  selecting  the 
best  one  of  several  normal  populations  with  a  common  unknown  variance, 
and  its  use  with  various  experimental  designs.  Biometrics,  14:408—29,  1958. 

5.  Box,  G.  E.  P.,  and  Andersen,  S.  L.  Permutation  theory  in  the  derivation  of 
robust  criteria  and  the  study  of  departures  from  assumption.   Jour.   Roy. 
Stat.  Soc.,  Series  B,  17:1-34,  1955. 

6.  Brownlee,  K.  A.  Statistical  Theory  and  Methodology  in  Science  and  Engineer 
ing.  John  Wiley  and  Sons,  Inc.,  New  York,  1960. 

7.  Chew,  V.  (editor)  Experimental  Designs  in  Industry.  John  Wiley  and  Sons, 
Inc.,  New  York,  1958. 

g^   f  Basic     experimental    designs.     Experimental    Designs    in    Industry. 

John  Wiley  and  Sons,  Inc.,  New  York,   1958. 
9.    Cochran,  W.  G.  Some  consequences  when  the  assumptions  for  the  analysis 

of  variance  are  not  satisfied.  Biometrics,  3:22,  1947. 

10.   .  Testing  a  linear  relation  among  variances.  Biometrics,  7:17,  1951. 

11.   ,  and  Cox,  G.  M.  Experimental  Designs.  Second  Ed.  John  Wiley  and 

Sons,  Inc.,  New  York,  1957. 

12.  David,  H.  A.  Upper  5  per  cent  and  1  percent  points  of  the  maximum  F-ratio. 
Biometrika,  39:422-24,  1952. 

13.  Davies,  O.  L.   (editor)    The  Design  and  Analysis  of  Industrial  Experiments. 
Second  Ed.  Oliver  and  Boyd,  Edinburgh,  1956. 

14.  Dixon,  W.  J.,  and  Massey,  F.  J-  Introduction  to  Statistical  Analysis.  Second 
Ed.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1957. 

15.  Duncan,   D.  B.   Multiple  range  and  multiple  .F-tests.    Biometrics,    11:1—42, 
1955. 

16.   .  Multiple  range    tests    for  correlated    and    heteroschedastic  means. 

Biometrics,  13:164-76,  1957. 

17.  Dunnett,    C.   W.   A   multiple  comparison  procedure  for   comparing   several 
treatments  with  a  control.  Jour.  Amer.  Stat.  Assn.,  50:1096—1121,  1955. 

18.  Eisenhart,    C.   The   assumptions   underlying  the   analysis   of  variance.   Bio 
metrics,  3:1-21,  1947. 

19.  Federer,  W.  T.  Experimental  Design.  Macmillan  Co.,  New  York,  1955. 

20.  Graybill,  F.  A.  An  Introduction  to  Linear  Statistical  Models.  McGraw-Hill 
Book  Company,  Inc.,  New  York,  1961. 

21.  Hamaker,   H.    C.   Experimental  design  in  industry.   Biometrics,    11:257—86, 
1955. 

22.  Hartley,  H.  O.  Some  recent  developments  in  analysis  of  variance.  Communi 
cations  on  Pure  and  Applied  Mathematics,  8:47—72,  1955. 

23.  EZempthorne,  O.   The  Design  and  Analysis  of  Experiments.  John  Wiley  and 
Sons,  Inc.,  New  York,  1952. 

24.  Keuls,  M.  The  use  of  the  "Studentized  range"  in  connection  with  an  analysis 
of  variance.  Euphytica,  1:112-22,  1952. 

25.  Kramer,  C.  Y.  Extension  of  multiple  range  tests  to  group  means  with  un 
equal  numbers  of  replications.  Biometrics,  12:307—10,  1956. 

26.   .  Extension    of    multiple    range    tests    to    group    correlated    adjusted 

means.  Biometrics,  13:13-18,  1957. 

27.  Newman,  D.  The  distribution  of  the  range  in  samples  from  a  normal  popula 
tion  expressed  in  terms  of  an  independent  estimate  of  standard  deviation. 
Biometrika,  31:20-30,  1939. 

28.  Quenouille,  M.  H.  The  Design  and  Analysis  of  Experiment.  Charles  Griffin 
and  Co.,  Ltd.,  London,  1953. 

29.  Satterthwaite,  F.  E.  An  approximate  distribution  of  estimates  of  variance 
components.  Biometrics,  2:110,  1946. 

30.  Scheff<§,  H.  A  method  for  judging  all  contrasts  in  the  analysis  of  variance. 
Biometrika,  40:87-104,  1953. 


362  CHAPTER    11,    COMPLETELY    RANDOMIZED    DESIGN 

31.  Snedecor,  G.  W.  Statistical  Methods.  Fifth  Ed.  The  Iowa  State  University 
Press,  Ames,  1956. 

32.  Tukey,  J.  W.  One  degree  of  freedom  for  nonadditivity.  Biometrics,  5:232—42, 
1949. 

33    ^  Comparing  individual  means  in  the  analysis  of  variance.  Biometrics, 

5:99,  1949. 
34.   .  Quick  and  dirty  methods  in  statistics.  Part  II:  Simple  analyses  for 

standard  designs.  Proc.  Fifth  Ann.  Conv.  A.S.Q.C.,  p.  189,  1951. 
35. .  Allowances  for  various  types  of  error  rates.   Unpublished  material 

presented  before  the  Institute  of  Mathematical  Statistics  and  the  Eastern 

North  American  Region  of  the  Biometric  Society  at  Blacksburg,  Va.,  Mar. 

19,  1952. 

36.  .  The  problem  of  multiple  comparisons.  Unpublished  dittoed  notes, 

Princeton  Univ.,  Princeton,  N.  J.,  1953. 

37.  .  Reply  to  query  number  113.  Biometrics}  11:111—13,  1955. 

38.  Ward,  G.  C.,  and  Dick,  I.  D.  Nonadditivity  in  randomized  block  designs 
and  balanced  incomplete  block  designs.  N.  Z.  Jour.  Sci.  and  Tech.,  33:430— 
35,  1952. 

39.  Yates,   F.   The  design  and  analysis  of  factorial  experiments.   Tech.   Comm. 
No.  35 ,  Imperial  Bureau  of  Soil  Science,  95  pp.,  1937. 


CH  APTE  R    12 

RANDOMIZED  COMPLETE  BLOCK 
DESIGN 

IN  THIS  CHAPTER,  the  most  widely  used  of  all  experimental  designs,  the 
randomized  complete  block  design,  will  be  discussed.  The  discussion 
will  follow  closely  the  pattern  adopted  in  Chapter  11  with,,  once  again, 
the  greatest  attention  being  given  to  methods  of  analysis. 

12.1      DEFINITION  OF  A  RANDOMIZED  COMPLETE  BLOCK 
DESIGN 

A  randomized  complete  block  (RCB}  design  is  a  design  in  which:  (1)  the 
experimental  units  are  allocated  to  groups,  or  blocks,  in  such  a  way 
that  the  experimental  units  within  a  block  are  relatively  homogeneous 
and  that  the  number  of  experimental  units  within  a  block  is  equal  to 
the  number  of  treatments  being  investigated,  and  (2)  the  treatments 
are  assigned  at  random  to  the  experimental  units  within  each  block.  In 
the  foregoing,  the  formation  of  the  blocks  reflects  the  researcher's 
judgment  as  to  potential  differential  responses  from  the  various  experi 
mental  units  while  the  randomization  procedure  acts  as  a  justification 
for  the  assumption  of  independence.  (See  Chapters  10  and  11.) 

Example  12.1 

Six  varieties  of  oats  are  to  be  compared  with  reference  to  their  yields, 
and  30  experimental  plots  are  available  for  experimentation.  However, 
evidence  is  on  file  which  indicates  a  fertility  trend  running  from  north 
to  south,  the  northernmost  plots  of  ground  being  the  most  fertile.  Thus, 
it  seems  reasonable  to  group  the  plots  into  five  blocks  of  six  plots  each 
so  that  one  block  contains  the  most  fertile  plots,  the  next  block  contains 
the  next  most  fertile  group  of  plots,  and  so  on  down  to  the  fifth  (south 
ernmost)  block  which  contains  the  least  fertile  plots.  The  six  varieties 
would  then  be  assigned  at  random  to  the  plots  within  each  block,  a  new 
randomization  being  made  in  each  block. 

Example  12.2 

An  experiment  is  to  be  designed  to  study  the  effect  of  environmental 
temperature  on  the  transfer  time  of  a  certain  type  of  electrical  gap. 
Twelve  different  temperatures  are  to  be  investigated.  A  check  of  the 
stockroom  indicates  that  gaps  are  available  from  six  different  production 
lots.  Since  it  has  previously  been  established  that  gaps  from  different 
lots  exhibit  different  characteristics,  even  when  subjected  to  the  same 
conditions,  some  blocking  is  desirable.  Accordingly,  12  gaps  are  selected 
at  random  from  each  of  the  six  production  lots,  and  each  such  set  of  12 
gaps  is  hereafter  referred  to  as  a  block.  Then  the  12  temperatures  are 

E3631 


364 


CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 


assigned  at  random  to  the  gaps  within  each  block,  a  new  randomization 
being  made  in  each  block. 

Example  12.3 

Ten  rations  are  to  be  tested  for  differences  in  producing  a  gain  in 
weight  for  steers.  Forty  steers  are  available  for  experimentation  and 
they  are  allocated  to  four  blocks  (10  steers  per  block)  on  the  basis  of 
their  weights  at  the  beginning  of  the  feeding  trial,  with  the  heaviest 
steers  being  in  one  block,  the  next  heaviest  steers  being  in  the  second 
block,  and  so  on.  The  10  treatments  (rations)  were  assigned  at  random 
to  the  steers  within  each  block,  as  shown  in  Figure  12.1. 


Block  1 

H 

B 

F 

A 

C 

I 

E 

J 

D 

G 

Block  2 

A 

I 

G 

H 

J 

D 

F 

E 

C 

B 

Block  3 

E 

A 

C 

I 

B 

H 

D 

G 

J 

F 

Block  4 

J 

F 

D 

B 

H 

I 

A 

C 

G 

E 

FIG.  12.1  —  Random  arrangement  of  treatments  as 
described    in    Example    12.3. 


12.2      RANDOMIZED    COMPLETE    BLOCK    DESIGN    WITH 
ONE  OBSERVATION  PER  EXPERIMENTAL  UNIT 

The  basic  assumption  for  a  randomized  complete  block  design  with 
one  observation  per  experimental  unit  is  that  the  observations  may  be 
represented  by  the  linear  statistical  model 


i  =  1,  -  -  - ,  b 
j  -  1,  -  -  -  ,t 


(12.1) 


where  p,  is  the  true  mean  effect,  f3g  is  the  true  effect  of  the  ith.  block, 
TJ  is  the  true  effect  of  the  jth  treatment,  and  e^  is  the  true  effect  of 
the  experimental  unit  in  the  ith  block  which  is  subjected  to  the  y 
treatment.  In  addition, 


=  0   and  e#  is  NID(0,  a). 


As  in  Chapter  11,  either  Model  I  or  Model  II  may  be  assumed  with  re 
spect  to  the  TJ. 

Using  the  symbolism  of  Table  12.1  and  the  following  equations: 


F2  =  total  sum  of  squares 
&       t 

-z 


(12.2) 


ONE   OBSERVATION    PER    EXPERIMENTAL    UNIT 


365 


TABLE   12.1-Synabolic  Representation  of  the  Data  in  a  Randomized 
Complete  Block  Design  With  One  Observation  per  Experimental  Unit 


Treatment 

Block 

1  ...              j  .  .  .                t 

Total 

Mean 

1.,  

Fii                     Fi  -                     Fi 

B 

•y 

i  .  ,  .     ... 

F-                 y  ••                v  - 

*j  i 

fj'h 

3 

Total 
Mean 

TI                     TJ                     Tt 
Y.i                  Y.s                  Y.t 

T 

F.. 

TABLE   12.2-Generalized  ANOVA  for  a  Randomized  Complete  Block 
Design  With  One  Observation  per  Experimental  Unit:  Model  I 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean 

Square 

IVtean 

1 

J\ft/*t 

M 

Blocks     ...         

b  —  1 

K-.-r 

B 

o-2  4-  1  2D  /3i/0 

>-D 

Treatments  

t  —  1 

T 

i—  1 

• 

Experimental  error  

(6—  !)(/!  —  1) 

&VV 

E 

/-i 
o-2 

Total 

to 

T:F* 

=sum  of  squares  due  to  the  mean 


=  among  blocks  sum  of  squares 


=  among  treatments  sum  of  squares 


S 

y— i 


(12.3) 
(12-4) 

(12.5) 


366  CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK   DESIGN 

and 

=  experimental  error  sum  of  squares 

H 


(12.6) 
-  Byy  -  Tyy, 

the  ANOVA  shown  in  Table  12.2  is  obtained. 

Following  the  line  of  reasoning  developed  in  Section  11.2,  it  is  easily 
verified  that  the  hypothesis  H:rj  =  Q  C?  =  l,  -  -  •  ?  0  may  be  tested  by 
computing 

mean  square  for  treatments 

(12,7) 


experimental  error  mean  square 

which,  if  jffistrue,  is  distributed  as  Fwithvi  =  t  —  1  and  *>2=  (6  —  !)(£—  1) 
degrees  of  freedom.  If  the  value  of  F  specified  by  Equation  (12.7)  ex 
ceeds  j^ci-cooi,  *2)>  where  lOOa  per  cent  is  the  chosen  significance  level, 
H  will  be  rejected  and  the  conclusion  reached  that  there  are  significant 
differences  among  the  t  treatments. 

As  before,  it  is  also  possible  to  estimate  cr2  by  s2  —  E.  Then,  too,  if 
Model  II  had  been  assumed,  o>  would  be  estimated  by 

sr  =  (T  -  E)/b.  (12.8) 

In  either  case,  that  is.  Model  I  or  Model  II,  the  estimated  variance  of 
a  treatment  mean  is  given  by 

F(F.y)  =  E/b  (12.9) 

and  the  standard  error  of  a  treatment  mean  is  given  by 


Vs2/b.  (12.10) 

A  IQOy  per  cent  confidence  interval  for  estimating  M/  —  M  +  T/  is  then 
found  by  calculating 

=   Y.J  T  *[ci+-y)/*iooVS7&  (12.11) 

where  ^=(&  —  !)(*—  1)- 

Example  12.4 

The  experiment  described  in  Example  12.3  was  performed  and  the 
data  of  Table  12.3  were  obtained.  Use  of  Equations  (12.2)  through 
(12.6)  yields  the  ANOVA  shown  in  Table  12.4.  Because  the  F-ratio  is 
significant,  we  reject  HIT,  =0(./  =  1,  •  •  •  ,  10)  and  decide  that  in  all 
likelihood  the  10  treatments  (rations)  are  not  equally  effective  in  pro 
ducing  a  gain  in  weight  on  steers.  The  treatment  means  and  their 
standard  error,  given  in  Table  12.3  for  convenience,  may  then  be  used  to 
determine  the  best  treatment  (or  treatments)  and  to  indicate  the 
direction  which  future  research  should  take. 


12.2       ONE    OBSERVATION    PER    EXPERIMENTAL    UNIT 


367 


TABLE   12.3-Gains  in  Weight  (in  Lbs,)  of  Forty  Steers  Fed  Different  Rations 

(Data  coded  for  easy  calculation) 


TV**a  t 

Block 

ment 

1 

2 

3 

4 

Total 

Mean 

A    

2 

3 

3 

5 

13 

3,25 

B           .    . 

5 

4 

5 

5 

19 

4.25 

C  

8 

7 

10 

9 

34 

8.50 

D  

6 

5 

5 

2 

18 

4.50 

E,  

1 

2 

1 

2 

6 

1.50 

F    

3 

5 

7 

8 

23 

5.75 

G           .    . 

8 

8 

7 

8 

31 

7.75 

H 

6 

12 

2 

5 

25 

6.25 

I 

4 

5 

6 

3 

18 

4.50 

J  

4 

4 

2 

3 

13 

3.25 

Total 

47 

55 

48 

50 

200 

Mean 

4  7 

5.5 

4.8 

5.0 

5.0 

Standard  error  of  a  treatment  mean  =-\/(3 -43) /4  =  0.93 

TABLE   12.4-ANOVA  for  Experiment  Described  in  Example  12.3  and 
Discussed  in  Example  12.4  (Data  in  Table  12.3) 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean 
Square 

F-Ratio 

IVTean 

1 

10OO  0 

1000  00 

Blocks 

3 

3  8 

1  26 

4 

0-24.  (10/3)  Y\  Bi 

Treatments  

9 

163.5 

18.17 

i—  a 
O-S-l-^/Q)    ]JTT* 

5.29** 

KxTD^riimpTi'tfil  error 

27 

92  7 

3  43 

J-—1 
<r2 

Total 

40 

1260  0 

**  Significant  at  a =0.01. 

Before  moving  along  to  the  next  topic  connected  with  the  analysis 
of  randomized  complete  block  designs,  there  is  one  point  that  needs 
discussion.  That  is,  why  do  we  not  test  £P:£i  =  0  (i=l,  •  -  •  ,  6)? 
Examination  of  Table  12.2  will  show  that  the  expected  mean  square 
for  blocks  is  of  the  same  form  as  the  expected  mean  square  for  treat 
ments,  and  this  suggests  that  a  logical  procedure  would  be  to  test  Hf 
by  calculating  F  =  B/E.  Why  is  it,  then,  that  the  statistician  says  this 
should  not  be  done?  The  answer  may  be  found  by  noting  the  manner 


368  CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 

in  which  the  randomization  was  performed.  You  will  recall  that  the 
treatments  'were  assigned  at  random  to  the  experimental  units  within 
each  block  but  that  the  blocks  were  formed  in  a  decidedly  nonrandom 
fashion.  Because  of  this  feature  of  the  randomized  complete  block  de 
sign,  a  statistical  test  of  the  block  effect  should  not  be  performed.  [NOTE: 
In  some  cases  where  "blocks"  are  replaced  by  "replications"  and  where 
the  replications  may  be  considered  as  random  samples  of  all  possible 
replications  (i.e..  Model  II  with  respect  to  replications),  an  ^P-test  for 
replications  may  be  appropriate.  However,  even  in  such  a  case,  the 
jF-test  would  be  of  less  importance  than  the  estimation  of  the  com 
ponent  of  variance  for  replications,  o>  ] 

12.3  THE    RELATION    BETWEEN    A    RANDOMIZED    COM 
PLETE   BLOCK    DESIGN  AND    "STUDENT'S"    *-TEST 
OF    jHr:Mz>==0     WHEN     PAIRED    OBSERVATIONS  ARE 
AVAILABLE 

In  Section  7.9,  procedures  were  given  for  testing  the  hypothesis 
H:fj,i  =  fj.2  for  three  different  cases.  In  Section  11.3  it  was  stated  that 
the  analysis  of  a  completely  randomized  design  was  equivalent  to  one 
of  these,  namely,  Case  I.  In  this  section  we  wish  to  show  the  equiv 
alence  of  the  analysis  of  a  randomized  complete  block  design  with  two 
blocks  and  "Student V  t-test  of  If'.jAi  —  juiz  when  paired  observations 
are  available,  that  is,  to  Case  III.  The  crucial  step  is  to  note  the  equiv 
alence  of  "pairs"  and  "blocks."  Once  this  association  of  terms  is  made, 
the  equivalence  of  the  techniques  may  easily  be  demonstrated.  Rather 
than  burden  the  reader  with  the  details  of  the  algebraic  proof  of  the 
equivalence,  we  will  rely  on  the  "power  of  an  example"  to  convince  him 
of  the  truth  of  our  claim.  (NOTE :  The  reader  should  also  reflect  on  the 
obvious  connection  between  the  material  of  this  section  and  the  con 
tents  of  Section  9.13.) 

Example  12.5 

Consider  again  the  experiment  described  in  Example  7.21  and  the 
data  presented  in  Table  7.6.  Denoting  pairs  (samples)  by  blocks  and 
utilizing  Equations  (12.2)  through  (12.6),  the  ANOVA  of  Table  12.5  is 
obtained.  It  is  seen  that  the  F- value  is  significant  at  a.  =  0.05,  and  this 
permits  us  to  reject  the  hypothesis  that  the  two  treatments  (i.e.,  two 
different  steel  balls)  are  doing  an  equivalent  job.  [NOTE:  ^  =  7.89 
=  £2  =  (2.81)2;  see  Example  7.21.]  It  may  be  verified  that  the  two  treat 
ment  means  are  54.2  and  46.2,  respectively.  Also,  the  standard  error  of 
a  treatment  mean  is  determined  to  be  V60. 8/15  =  2.01. 

12.4  SUBSAMPLING      IN     A      RANDOMIZED     COMPLETE 
BLOCK   DESIGN 

When  subsampling  is  employed  in  a  randomized  complete  block  de 
sign,  the  appropriate  statistical  model  is 


12.4       SUBSAMPLING  369 

TABLE   12.5-ANOVA  for  Experiment  Described  in  Examples  7.21  and 

12.5  (Data  in  Table  7.6) 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square 

jF-Ratio 

Mean  

1 

75  601  2 

75  601  2 

Blocks 

14 

2,649  8 

189  3 

15 

<rM-(2/14}  Y^  8* 

Treatments  

1 

480  0 

480.0 

i-i 

2 

cr2H-(15/l)  T"  Tf 

7  89* 

Exp^T~i™~»^nt2»l  e.rrnr     , 

14 

851,0 

60.8 

J-l 

<r2 

Total 

30 

79,582.0 

*  Significant  at  «=O.05. 

J  =   1,   '   *   '  >  t 

k  =  1,  •  -  -  ,  n 

and  the  terms  are  defined  in  the  usual  manner.  The  various  sums  of 
squares  are  found  as  follows : 

]YZ  =  total  sum  of  squares 

,  (12.13) 


=  sum  of  squares  due  to  the  mean 

H  (12.14) 

=  T^/btn^ 

=  among  cells  sum  of  squares  for  the  block  X  treatment  table 
&       t 

XT^  x~^  T   /M       K/T  M9  1 1^^ 

=     /  ;    /  ^   J.  ij/n  —   M-yyi  {LZ.  L3} 

=  sampling  error  sum  of  squares 


B      = 


yy 


block  sum  of  squares 
—  Mvy, 


(12.16) 
(12.17) 


=  treatment  sum  of  squares 


Tj/bn  — 


(12.18) 


37O 

and 


CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK   DESIGN 


Eyy  =  experimental  error  sum  of  squares 

===    £>bt  -L$ity  J-  yy 

where 

T  =  grand  total 

b          t         n 

-  z:  z  2:  r«* 

t^l  y^i  &=! 

7\7  =  total  of  all  observations  in  the  iih  block  that  were 
subjected  to  thejth  treatment 


(12.19) 


(12.20) 


(12.21) 


?i  =  total  of  all  observations  in  the  iih  block 

t         n  t 


(12.22) 


and 


TV  ==  total  of  all  observations  subjected  to  the^th  treatment 


==       S   -*       2  _^r       *     ijjc     ==       S     *       J-    tj - 


(12.23) 


Using  the   preceding   results,   the   ANOVA   shown   in   Table    12.6   is 
obtained. 

TABLE  12.6-Generalized  ANOVA  for  a  Randomized  Complete  Block 
Design  With  n  Samples  per  Experimental  Unit:  Model  I 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square 

F-Ratio 

M!ean  

1 

Ay.  tii  t 

M 

Blocks  

6—1 

B 

2                                  ^A       2 
0.   _I_**0.2_L./**     >  "   Q. 

Treatments  

2—1 

T 

T 

i—  1 
t 

0^^.no.^^rf)n  y^  X2 

T/E 

Experim  en  tal  error  .  . 

(£,—  !)(£_!) 

J—tyy 

E 

/—  i 

a2_|_         2 

Sampling  error  

bt(n—  1) 

C- 

s 

V   ] 

o-2 

-n 

Total 

Un 

y^r2 

12.5      PRELIMINARY  TESTS 


371 


Example  12.6 

An  experiment  was  performed  to  assess  the  relative  effects  of  five 
fertilizers  on  the  yield  of  a  certain  variety  of  oats.  The  location  of  the 
30  experimental  plots  available  for  use  in  the  experimentation  was  such 
that  it  seemed  advisable  to  group  the  plots  into  six  blocks  of  five  plots 
each.  The  treatments  were  then  randomly  assigned  to  the  plots  within 
each  block.  At  the  end  of  the  growing  season  the  researcher  decided  to 
harvest  (for  purposes  of  analysis)  only  three  sample  quadrats  from 
each  plot.  The  data  of  Table  12.7  were  obtained  and  these  led  to  the 
ANOVA  shown  in  Table  12.8.  It  is  noted  that  the  five  fertilizer  means 
are  significantly  different  and,  thus,  it  is  most  important  that  the 
proper  fertilizer  be  recommended  to  the  farmer.  A  tabulation  of  the 
fertilizer  means,  together  with  the  appropriate  standard  error,  would 
be  of  great  help  in  reaching  the  correct  decision. 

TABLE  12.7-Coded  Values  of  Yields  From  Ninety  Sample  Quadrats 


Blocks 

Fertilizer  Treatments 

1 

2 

3 

4 

5 

1  

57 

67 

95 

102 

123 

46 

72 

90 

88 

101 

28 

66 

89 

109 

113 

2  

26 

44 

92 

96 

93 

38 

68 

89 

89 

110 

20 

64 

106 

106 

115 

3  

39 

57 

91 

102 

112 

39 

61 

82 

93 

104 

43 

61 

98 

98 

112 

4  

23 

74 

105 

103 

120 

36 

47 

85 

90 

101 

18 

69 

85 

105 

111 

5  

48 

61 

78 

99 

113 

35 

60 

89 

87 

109 

48 

75 

95 

113 

111 

6   

50 

68 

85 

117 

124 

37 

65 

74 

93 

102 

19 

61 

80 

107 

118 

12.5  PRELIMINARY  TESTS  OF  SIGNIFICANCE 

At  this  time  I  wish  to  digress  for  a  few  moments  from  the  pattern 
established  in  Chapter  11,  and  followed  thus  far  in  the  present  chapter, 
to  discuss  a  matter  of  considerable  importance.  This  topic,  namely, 


372  CHAPTER    12,    RANDOMIZED    COMPLETE   BLOCK    DESIGN 

TABLE   12.8-ANOVA  for  Data  of  Table  12.7;  Discussion  in  Example  12.6 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square 

F-Ratio 

Mean  

1 

573,921.88 

573,921.88 

Blocks  

5 

354.19 

70.84 

JU        2 

<r2-f-3<7-2  -K15/5)  J^Qi 

KeT'tilf  ^ers  .  .  , 

4 

65  246  84 

16311  71 

s 

O^+ScrM-ClS/i)   y~\  Tj 

220  61** 

Experimental 
error  

20 

1,478.76 

73.94 

*                                               3-1 

o-2-i-3<r2 

Sampling  error. 

60 

5,283  33 

88  06 

"17   i^ 

a* 

Grj 

Total 

90 

646,285.00 

Significant  at  ex =0.01. 


the  use  of  preliminary  tests  of  significance,  could  just  as  easily  have 
been  discussed  in  Chapter  11  or  it  could,  without  difficulty,  be  deferred 
until  later.  However.,  Example  12.6  at  the  end  of  Section  12.4  brought 
it  to  my  attention,  and  thus  we  shall  consider  it  at  this  point. 

Some  practitioners  suggest  that  when,  as  in  Example  12.6,  the  experi 
mental  error  mean  square  is  less  than  the  sampling  error  mean  square, 
the  two  sums  of  square  and  their  degrees  of  freedom  be  pooled.  (The 
same  suggestion;  i.e.,  to  pool,  is  also  frequently  made  when  the  experi 
mental  error  mean  square  exceeds,  but  not  significantly,  the  sampling 
error  mean  square.)  That  is,  a  pooled  sum  of  squares  (Ew+Sw)  is 
divided  by  the  pooled  degrees  of  freedom  [(&  —  1)  (2  —  l)+&^(n  —  1)],  and 
this  new  mean  square  is  then  used  as  the  denominator  in  the  .F-ratio 
for  testing  H:r3-  =  Q  (y=  1,  •  -  -  ,  t).  If  such  a  procedure  is  followed,  the 
statistical  test  of  jETrry  —  O  (j—1,  *•*,£)  will  be  based  on  a  preceding, 
or  preliminary,  test  of  significance,  the  hypothesis  H':<r2  =  Q  being 
tested  by  the  preliminary  test.  Because  such  procedures  are  sometimes 
followed,  we  must  make  certain  that  we  understand  their  advantages 
and  disadvantages. 

Problems  of  the  above  type  have  been  investigated  by  Paull  (10), 
and  it  will  pay  us  to  spend  a  few  moments  reviewing  his  conclusions 
and  recommendations.  To  make  the  exposition  easier  to  follow,  we 
shall  tie  it  in  with  Table  12.6  .  Suppose  the  experimenter  decides  he  will 
always  pool  the  two  mean  squares  as  indicated  in  the  preceding  para 
graph,  that  is,  he  will  never  perform  a  preliminary  test  of  significance 
concerning  H':<r2  =  Q.  If,  in  fact,  <r2  does  equal  0,  this  procedure  is  fine. 
But  suppose  a-2  >  0 ;  then  the  denominator  in  the  final  F-test  (of  HiTj  =  0 
for  allj)  tends  to  be  too  small.  Thus,  in  such  a  situation,  the  final  F-test 
tends  to  produce  too  many  significant  results  when  the  mill  hypothesis 


12.6       ESTIMATION    OF    COMPONENTS    OF   VARIANCE  373 

H  is  really  true.  This  is  bad,  for  it  implies  that,  quoting  Paull,  "a  test 
which,  the  research  worker  thinks  is  being  made  at  the  5  per  cent  level 
might  actually  be  at,  say,  the  47  per  cent  level."1 

The  use  of  a  preliminary  test  of  significance  is  clearly  an  attempt  to 
guard  against  such  a  possibility.  It  will  not,  of  course,  eliminate  such 
occurrences  entirely.  To  be  useful,  however,  it  should  keep  the  actual 
(effective)  significance  level  achieved  by  the  final  (or  dependent)  F-test 
close  to  the  value  at  which,  the  research  worker  desires  to  operate. 
Another  property  which  should  be  required  of  a  preliminary  test  is  that 
it  increase  the  power  of  the  final  -F-test  relative  to  the  power  of  a 
"never  pool"  test.  The  recommendations  for  the  performance  of  pre 
liminary  tests  of  significance  for  pooling  mean  squares  in  the  analysis 
of  variance  as  formulated  by  Paull  may  be  found  in  the  reference 
quoted.  However,  if  the  research  worker  follows  the  rule  of  "never 
pooling,"  he  will  not  go  far  wrong,  and  that  is  the  rule  we  shall  adopt 
in  this  text. 

12.6      ESTIMATION  OF  COMPONENTS  OF  VARIANCE  AND 
RELATIVE   EFFICIENCY 

The  problems  of  estimating  components  of  variance  and  predicting 
relative  efficiency  are  no  different  in  a  randomized  complete  block 
design  than  they  were  in  a  completely  randomized  design.  Thus,  they 
will  not  detain  us  long. 

As  before,  <r%  is  estimated  by  s*  =  S  (see  Table  12.6)  and  <r2  is 
estimated  by 

s*  =   (E  -  S)/n.  (12.24) 

If,  as  in  Table  12.8,  the  algebraic  solution  leads  to  a  negative  value  of 
s2,  it  is  customary  to  disregard  the  algebraic  solution  and  use  0  as  the 
estimate.  Of  course,  this  is  a  biased  procedure,  but  it  is  aesthetically 
more  satisfying  since  no  population  variance  can  be  negative. 

The  relative  efficiencies  of  various  allocations  -within  a  randomized 
complete  block  design  will,  as  was  the  case  when  discussing  a  completely 
randomized  design,  be  determined  by  studying  the  variance  of  a  treat 
ment  mean.  It  may  be  shown  that 

^  _  experimental  error  mean  square 


number  of  observations  per  treatment  (12.25) 

=  E/bn  =  (4  +  ns*)/bn. 

Thus,  if  the  estimates  of  the  components  of  variance  remain  unchanged, 
the  efficiency  of  this  design  relative  to  one  in  which  we  might  use  &' 
blocks  and  n'  samples  per  experimental  unit  would  be  predicted  by 

R.R.  of  old  to  new  -  100  [F'(F.yJ/F(F.,.)]  per  cent       (12.26) 

1  A.  E.  Paull,  "On  a  preliminary  test  for  pooling  mean  squares  in  the  analysis  of 
variance,"  Ann.  Math.  Stat.,  Vol.  21,  1950,  p.  541. 


374 

where 


CHAPTER    12,    RANDOMIZED    COMPLETE   BLOCK    DESIGN 


(12.27) 


Example  12.7 

^  Consider  Table  12.9.  It  is  easily  seen  that  s%  =  10  and  s2  =  12,  yielding 
V(Y .j,}  =58/24.  To  determine  the  efficiency  of  the  design  used  relative 
to  one  involving  four  blocks  and  six  samples  per  experimental  unit, 
we  first  calculate  F'(F.y.)  =  [10  +  6(12)  ]/4(6)  =82/24.  Thus,  the 
estimated  relative  efficiency  is  100(82/24)/(58/24)  =  141  per  cent. 

TABLE   12. 9- Abbreviated  ANOVA  on  Yields  of  Ten  Varieties  of  Soybeans 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean 

Square 

Blocks  

5 

3,000 

600 

cr2-Mt<r2+(40/5) 

±rf 

Varieties  

9 

4,500 

500 

o-2+4o-2-K24/4) 

"F  TZ- 

Experimental  error.  .  .  , 
Sampling  error  

45 
180 

2,610 
1,800 

58 
10 

£ 

" 

v 

The  ideas  of  this  section  may  easily  be  extended  to  cases  involving 
many  stages  of  subsampling.  No  examples  will  be  given,  but  several 
of  the  problems  at  the  end  of  the  chapter  will  provide  the  necessary 
practice  in  the  manipulations. 

12.7  EFFICIENCY  OF  A  RANDOMIZED  COMPLETE 
BLOCK  DESIGN  RELATIVE  TO  A  COMPLETELY 
RANDOMIZED  DESIGN 

In  some  instances,  the  investigator  wishes  to  estimate  the  efficiency 
of  his  use  of  a  randomized  complete  block  design  relative  to  what 
might  have  happened  if  the  treatments  had  been  completely  random 
ized  over  all  the  experimental  units.  That  is,  he  wishes  to  know  if  he 
gained  or  lost  in  efficiency  by  grouping  the  experimental  units  into 
homogeneous  groups  (blocks)  .  One  method  of  comparing  the  efficiency 
of  different  designs  is  by  use  of  uniformity  data.  Cochran  (4)  has  dis 
cussed  this  particular  approach,  and  the  reader  is  referred  to  his  article 
for  further  details.  A  second  method  of  comparing  efficiencies  is  to 
consider  algebraically  what  might  have  happened  to  the  experimental 
error  mean  square  under  complete  randomization.  To  accomplish  this, 
it  is  convenient  to  proceed  as  though  dummy  treatments  had  been 
applied  to  the  experimental  units.  That  is,  we  suppose  that  all  experi 
mental  units  were  subjected  to  the  same  (viz.,  no)  treatment  and  then 
proceed  to  estimate  what  the  experimental  error  mean  square  would 


12.7       EFFICIENCY    OF    RANDOMIZED    COMPLETE    BLOCK    DESIGN 


375 


have  been  under  complete  randomization.  Following  this  line  of  reason 
ing,  and  defining  the  efficiency  of  a  randomized  complete  block  design 
relative  to  a  completely  randomized  design  by 


estimated  experimental  error  mean  square  for  a  CR  design 
experimental  error  mean  square  from  the  RCB  design 
it  can  be  shown  that 

(b  —  1)J5  +  b(t  — 


R.E.  = 


(bt  — 


(12.28) 


(12.29) 


where  JB  and  E  refer  to  the  mean  squares  (in  the  randomized  complete 
block  design)  for  blocks  and  experimental  error,  respectively. 

Example  12.8 

Consider  the  ANOVA  presented  in  Table  12.4.  In  this  case,  the  effi 
ciency  of  the  randomized  complete  block  design  relative  to  a  completely 
randomized  design  is  estimated  to  be 


R.E. 


3(1.26)  +4(9)  (3.43) 
39(3.43) 


0.95. 


It  is  seen  that,  because  of  the  small  magnitude  of  B  relative  to  E,  no 
appreciable  gain  in  efficiency  resulted  from  the  formation  of  the  blocks 
and  the  use  of  a  randomized  complete  block  analysis.  That  is,  apart 
from  the  "insurance"  feature  of  the  RCB  design,  the  added  effort  was 
not  worthwhile. 

Example  12.9 

The  data  in  Table  12.10  resulted  from  a  particular  manufacturing 
operation,  the  operation  being  performed  by  one  of  four  different  ma 
chines.  The  data  were  collected  on  five  different  days,  hereafter  referred 
to  as  blocks.  Calculations  yielded  the  abbreviated  ANOVA  shown  in 
Table  12.11.  Proceeding  according  to  Equation  (12.29),  the  randomized 
complete  block  design  is  estimated  to  be  131  per  cent  as  efficient  as  a 
completely  randomized  design  would  have  been. 

TABLE   12,10-Output  From  Four  Machines  Producing  Part  No.  Z-15 
(Output  =  number  of  units  produced  In  one  day) 


Machine 


Day 

A 

B 

C 

D 

1  

293 

308 

323 

333 

2   

298 

353 

343 

363 

3  

280 

323 

350 

368 

4  

288 

358 

365 

345 

5  

260 

343 

340 

330 

376  CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 

TABLE   12, 11- Abbreviated  ANOVA  for  Data  of  Table  12.10 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean  Square 

Blocks               

4 

2,146.2 

536.55 

Treatments               

3 

13,444.8 

4,481.60** 

Experimental  error    

12 

2,626.2 

218.85 

**  Significant  at  <*=0.01. 

12.8      SELECTED  TREATMENT  COMPARISONS 

It  is  often  desirable  to  make  certain  specific  comparisons  involving  a 
selected  number  of  the  treatments.  For  a  completely  randomized  de 
sign,  such  comparisons  were  discussed  in  Sections  11.8  and  11.9.  In  this 
section  the  same  general  topic  will  be  examined  in  conjunction  with  a 
randomized  complete  block  design. 

A  moment's  thought  should  be  sufficient  to  convince  the  reader  that 
the  calculations  will  be  performed  in  the  same  manner  as  indicated 
earlier.  For  example,  if  a  randomized  complete  block  design  with  one 
observation  per  experimental  unit  is  involved,  the  sum  of  squares  for 
a  particular  contrast,  C*,  would  be  given  by 


=  ( i^  «**r/Y  /  b  i: 

\  y— i  /    /  y— i 


2 

Cjk, 


(12.30) 


where  the  Cjk  are  the  coefficients  specifying  the  contrast.  As  before,  if 
t—I  orthogonal  contrasts  are  studied,  the  sum  of  the  t—I  individual 
sums  of  squares  will  equal  the  treatment  sum  of  squares. 

Example  12.10 

Upon  reading  the  complete  description  of  the  project  referred  to  in 
Example  12.9,  certain  additional  information  about  the  four  machines 
is  brought  to  light.  For  example,  machine  A  is  the  standard  type  of 
machine  now  in  use  in  the  industry,  while  machines  B,  C,  and  D  are 
new  designs  which  may  be  considered  as  possible  substitutes.  Further, 
it  is  known  that  B  and  C  contain  moving  parts  made  of  some  aluminum 
alloy,  while  D  does  not  have  this  feature.  Also  known  from  the  manu 
facturers*  specifications  is  the  fact  that  B  is  self-lubricating,  while  C  is 
not.  Therefore,  the  comparisons  represented  symbolically  in  Table 
12.12  seem  to  be  indicated.  Partitioning  of  the  treatment  sum  of 
squares  is  then  carried  out  using  Equation  (12,30),  and  the  abbreviated 
ANOVA  of  Table  12.13  is  obtained. 

12.9  SUBDIVISION  OF  THE  EXPERIMENTAL  ERROR 
SUM  OF  SQUARES  WHEN  CONSIDERING  SELECTED 
TREATMENT  COMPARISONS 

Before  proceeding  to  the  next  general  topic  connected  with  random 
ized  complete  block  designs,  another  digression  seems  desirable.  This 


12.9       SUBDIVISION    OF    EXPERIMENTAL    ERROR    SUM    OF    SQUARES 


377 


TABLE   12.12-Symbolic  Representation  of  the  Selected  Treatment 
Comparisons  Described  in  Example  12.10  (Data  in  Table  12.10) 


Comparison 

Machine 

A 

B 

C 

D 

1 

+  3 
0 
0 

—  1 

+  1 
^ 

—  1 
+  1 
+  1 

—  1 
—  2 
0 

2  

3    

TABLE    12.13-Abbreviated  ANOVA  for  Data  of  Table  12.10  Showing  the 
Subdivision  of  the  Treatment  Sum  of  Squares 
(Discussion  in  Example  12.10) 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean  Square 

Blocks                

4 

2,146.2 

536.55 

Treatments 
A  vs.  rest  

1 

13,142.4 

13,142.4** 

B  and  C  vs.  D  

1 

172.8 

172.8 

B  vs.  C    

1 

129.6 

129.6 

Experimental  error  

12 

2,626.2 

218  85 

**  Significant  at  a  =  0.01. 

time  we  will  be  concerned  with  the  possibility  of  partitioning  the  experi 
mental  error  sum  of  squares.  (NOTE:  You  will  recall  that  this  topic 
was  mentioned  in  the  last  paragraph  of  Section  11.13.) 

When  is  such  a  procedure  in  order?  That  is,  when  should  the  experi 
mental  error  sum  of  squares  be  subdivided?  The  reason  for  sub 
dividing  Eyy  (if  such  a  procedure  is  adopted)  is  that  we  are  not  satisfied 
with  our  assumption  of  homogeneous  variances  of  the  e's.  If  such  an 
assumption  is  questioned — its  validity  may,  of  course,  be  investigated 
using  Bartlett's  test — and  if  selected  treatment  comparisons  are  being 
examined,  it  is  desirable  to  subdivide  Evy  in  a  manner  similar  to  the 
subdivision  of  Tyy.  Such  a  procedure  insures  that  any  particular  treat 
ment  comparison  will  be  tested  against  the  appropriate  error.  That  is, 
the  expected  value  of  the  "error  mean  square  for  testing  Ck"  will  con 
tain  the  same  components  of  variance  (other  than  the  treatment  effects) 
as  the  expected  value  of  the  mean  square  associated  with  C^  In  other 
words,  if  we  are  faced  with  different  variances  o-^  (f  =  l,  •  •  •  ,  Z>; 
y=l,  *  -  •  ,  £),  the  procedure  of  subdividing  Eyy  will  insure  that  the 
expected  mean  squares  for  a  particular  comparison  and  its  associated 
error  will  each  contain  the  same  linear  combination  of  the  c%j.  This,  of 
course,  provides  us  with  unbiased  tests  for  the  comparisons  under  in 
vestigation. 

Since  the  decision  to  implement  the  procedure  (yet  to  be  described) 


378  CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 

for  subdividing  Ew  is  one  which  everyone  will  have  to  make  at  some 
time  or  other,  some  guiding  rule  is  needed.  The  following  appears  to 
be  a  reasonable  rule  :  If  there  is  any  serious  doubt  as  to  the  homogeneity 
of  the  variances,  subdivide  the  experimental  error  sum  of  squares  as  a 
precautionary  measure.  It  should  be  noted,  though,  that  if  the  degrees 
of  freedom  associated  with  the  various  parts  of  Eyy  are  small,  the  re 
sulting  tests  may  be  relatively  insensitive  (i.e.,  of  low  discriminatory 
power).  In  practice  both  the  following  conditions  are  usually  true: 

(1)  The  degree  of  heterogeneity  among  the  error  variances  is,  as 
a  rule,  not  too  great.  Therefore,  for  most  practical  purposes, 
the  variances  may  be  considered  homogeneous. 

(2)  The  numbers  of  degrees  of  freedom  associated  with  the  parts 
into   which  the   experimental  error  sum   of   squares   is   sub 
divided  are  generally  quite  small. 

Consequently,  the  rule  stated  above  should  be  modified  to  read;  Be 
cause  of  the  truth  of  statements  (1}  and  (j£)  above,  it  is  generally  not  wise 
to  subdivide  the  experimental  error  sum  of  squares.  However,  if  the  hetero 
geneity  of  variances  is  such  that  a  subdivision  is  necessary  (regardless  of 
the  fact  that  small  numbers  of  degrees  of  freedom  will  result),  the  sub 
division  should  be  carried  out  in  accordance  with  the  procedure  to  be  ex 
plained  in  the  next  paragraph. 

The  method  of  subdividing  the  experimental  error  sum  of  squares  in 
agreement  with  a  particular  subdivision  of  the  treatment  sum  of  square 
is  as  follows  : 

(1)  Set  up  a  table  showing  the  values  of  the  contrasts  within  each 
block. 

(2)  Calculate  the  portion  of  the  experimental  error  sum  of  squares 
for  a  particular  contrast  using 

(Ek)yy  ===  experimental  error  sum  of  squares  for  Ck 

=  f  ib  el*  -  ci/b\  /  i:  4 

L    i=l  J  /          /=»! 

where 


Cki=     C  c&Yv  (12.32) 

-=i 

and 


Ck  ==          cjkTj  -     T  Cki.  (12.33) 

j—  i  i=»i 

Example  12.11 

Consider  the  experiment  discussed  in  Examples  12.9  and  12.10.  Using 
Equations  (12.32)  and  (12,33)  in  conjunction  with  Tables  12.10  and 
12.12,  we  obtain  Table  12,14.  Then,  using  Equation  (12.31),  we  get,  for 


12.9      SUBDIVISION   OF   EXPERIMENTAL   ERROR   SUM   OF   SQUARES 


379 


example, 


experimental  error  sum  of  squares  associated  with  Ca 

experimental  error  sum  of  squares  associated  with  the  comparison 

"B  versus  C" 


4. 


(_3)2  __  (36)V5]/2 


426.4. 


Similarly,  (EJvy  =  1087.27  and  (#2)^  =  1112.53.  Thus,  we  finally  obtain 
the  abbreviated  ANOVA  shown  in  Table  12.15. 

TABLE   12.14-Sums  for  the  Selected  Treatment  Comparisons  in  Each 

Block  (Data  of  Table  12.10) 


Comparison 

Block 

Ci 

C2 

C5 

1  

—    85 

—  35 

+  15 

2  

—  165 

—  30 

—  10 

3         

—  201 

—  63 

+  27 

4 

—  204 

+33 

+    7 

5  

—  233 

+23 

—    3 

Total 

—  888 

—  72 

+36 

TABLE   12. 15- Abbreviated  ANOVA  for  Data  of  Table  12.10  Showing 
The  Subdivision  of  the  Experimental  Error  Sum  of  Squares 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 

Square 

Expected  Mean  Square 

* 

Blocks  

4 

2,146.20 

536.55 

<r'+(4/4)  ±  /*' 

Treatments  : 
A  vs   rest 

1 

13,142.40 

13,142.40 

<r2  +  (25/60)  (3-n  —  r2  —  T3  — 

r^3 

B  and  C  vs.  D  .  .  . 
B  vs   C 

1 

1 

172.80 
129.60 

172.80 
129.60 

<r2+(25/30)(r2+r3-2r4)2 
0.2^  (25/10)  (r3  —  r2)2 

Experimental  error: 
A  vs   rest 

4 

1,087.27 

271  .  82^) 

B  and  C  vs.  D  .  .  . 
^  vs.  C  

4 
4 

1,112.53 
426.40 

278.  13  > 
106.  60J 

cr* 

*  The  symbol  o-2  was  used  in  each  expected  mean  square  as  a  matter  of  convenience.  If 
the  variances  are  homogeneous,  it  is  correct;  if  the  variances  are  not  homogeneous,  the 
symbol  <r2  would  be  replaced  by  various  linear  combinations  of  the  o\y. 

The  procedure  explained  and  illustrated  in  this  section  can  sometimes 
be  used  to  advantage  when  analyzing  a  particular  set  of  data.  How 
ever,  the  reader  should  realize  that  it  is  a  special  technique  and  will, 
therefore,  be  used  only  rarely. 


38O  CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 

12.10  ALL     POSSIBLE     COMPARISONS    AMONG     TREAT 
MENT   MEANS 

Since  the  problem  of  making  all  possible  comparisons  among  treat 
ment  means  in  a  randomized  complete  block  design  is  handled  in 
exactly  the  same  manner  as  similar  comparisons  were  handled  in  a 
completely  randomized  design,  no  additional  discussion  is  necessary. 
The  reader  is,  therefore,  referred  to  Section  11.10  for  the  appropriate 
details. 

12.11  RESPONSE  CURVES   IN   A   RANDOMIZED 
COMPLETE   BLOCK  DESIGN 

Once  again  it  is  sufficient  to  state  that  the  techniques  explained  in 
the  preceding  chapter  are  directly  applicable  to  the  present  situation. 
Thus,  the  reader  is  referred  to  Section  11.11  for  the  computational 
details.  However,  to  emphasize  the  "sameness/*  an  illustrative  example 
will  be  presented.  (NOTE:  The  reader  will  find  it  rewarding  to  com 
pare  the  following  example  with  Example  11.15.) 

Example  12.12 

Considering  the  data  of  Table  12.16  and  using  the  methods  described 
in  Section  11.11,  the  abbreviated  ANOVA  shown  in  Table  12.17  is 
obtained. 

TABLE   12. 16- Yields  (Converted  to  Bushels/Acre)  of  a  Certain  Grain 

Crop  in  a  Fertilizer  Trial 


Block 

Level  of  Fertilizer 

No 
Treat 
ment 

10  Ibs, 
per  Plot 

20  Ibs. 
per  Plot 

30  Ibs. 
per  Plot 

40  Ibs. 
per  Plot 

1 

20 

25 
23 
27 
19 

25 

29 
31 
30 
27 

36 

37 
29 
40 
33 

35 
39 
31 
42 
44 

43 
40 
36 
48 
47 

2  

3  

4  

5  

Treatment 
totals 

114 

142 

175 

191 

214 

12.12 


FACTORIAL  TREATMENT  COMBINATIONS   IN   A 
RANDOMIZED  COMPLETE   BLOCK   DESIGN 


Because  of  tlie  detail  with  which  the  analysis  of  factorial  treatment 
combinations  was  discussed  in  connection  with  a  completely  random 
ized  design  (see  Section  11,12),  only  a  summary  discussion  seems 
appropriate  here.  Accordingly,  all  that  will  be  given  are  two  linear 


12.12       FACTORIAL    TREATMENT    COMBINATIONS 


381 


TABLE   12. 17- Abbreviated  ANOVA  for  Data  of  Table  12.16  Showing 
the  Isolation  of  the  Linear  and  Quadratic  Portions  of  the  Treatment 

Sum  of  Squares 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of  Squares 

Mean 

Square 

Blocks  

4 

154   16 

38  54 

Treatmen  ts 

4 

1256  56 

314   14 

Linear  

1 

1240  02 

1240  02** 

Quadratic  

1 

10  41 

10  41 

[Deviations    from    re 
gression  .  . 

2 

6   13 

3   07 

TC-jrp^Hrnenfal  errnr 

16 

193.44 

12   09 

**  Significant  at  «  =  0.01. 

statistical   models,   their  associated   ANOVA's,    and   some   numerical 
examples. 


Two -Factor  Case 


ijk 


Pi 


(12.34) 


Three- Factor  Case 

Yjjki  =  JJL  +  pi  +  0.3  +  fa  + 


j  =  1,  • 
k  =  1, . 


—  1> 

=  1, 
=  1, 
=  1, 


(12.35) 


,  a 
,  b 
,  c. 


In  the  preceding  equations,  M  is  the  true  mean  effect,  p^  is  the  true 
effect  of  the  ith  replicate  (or  block),  the  various  terms  involving  a,  (3 
and  y  are  the  true  effects  of  the  several  factors  and  their  interactions, 
and  the  e's  are  the  true  effects  of  the  experimental  units.  The  general 
ized  ANOVA's  associated  with  Equations  (12.34)  and  (12.35)  are 
shown  in  Tables  12.19  and  12.18,  respectively. 

Example  12.13 

Consider  the  data  in  Table  12.20.  Performing  the  usual  calculations, 
we  obtain  the  abbreviated  ANOVA  shown  in  Table  12,21.  Testing 
#1:0^  =  0  (.7  =  1,  2),  we  calculate  ^  =  32.00/3.16  =  10.2,  and  this  leads  to 
the  rejection  of  Hi.  Testing  £T2:/3;b  =  0  (&  =  1,  2,  3,  4),  we  calculate 
p  =  5.47/3.16=  1.73,  and  this  does  not  permit  H%  to  be  rejected.  To  test 
#3:(o!0)/fc  =  0  (.7  =  1,  2;  fc  =  l,  2,  3,  4),  we  calculate  F  =  12.80/3.16  =  4.05 
and  this  leads  to  rejection  of  H$  at  the  5  per  cent  significance  level  but 


TABLE  12.18-Generalized  ANOVA  for  a  Three-Factor  Factorial  in  a 


Expected  Mean  Square 

Source  of 

Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 

Square 

Model  I 

M 

Mean  

Myy 
T> 

•p 

Replicates,  .  .  . 

Treatments 

f—  1 

•K-w 

1 

A 

A 

n-i+rbt:  T!<£/fo-l) 

6-1 


AB. 


(a- 1)0- 1 


By 


Cyy 


B 


a*+rab 


1-1 


AB 


AC     

(a_l)(c_l) 

(AQyy 

AC 

(rz+rb  >  .  >  .  (a7)yi/(0""1-)w"'1^ 

/-i  1-1 

6        c^              j 

EC  

(d-l)fr-l) 

(BC)yy 

BC 

<r*+ra^T,  X)  (fryOfci/C^  l)(c—  1) 

fc-i  1-1 

a        6        c                    » 

ABC  

(0-1)  (6-  !)(<?-!) 

(ABQyy 

ABC 

;—l  jfe-1    i-1 

Experimental 

error  

(r-lXafa-l) 

Ryy 

E 

** 

—  _T 

EV2 

Total 

TdOC 

* 

**+rc  i:  : 


,--1 


[3821 


Randomized  Complete  Block  Design 


Expected  Mean  Square 


Model  II 


Model  in 
(a  and  b  fixed,  c  random) 


Model  III 
(a  fixed,  5  and  c  random) 


j-i 


[3831 


ft 

-s 


a 

o 
O 


O 


5 

o 


5-, 

5 

O 


oS 
O 


I— i      -C> 

'S       I 


d 


. 

1  * 

S  ,8 


d 

d 


"2 


I 


1 


4- 

Nb 


b 
"b 


4- 


b 

o 

b 


" 

4- 


5 

*b 


S 

4- 

wb 


^ 

-W2 


a    Z 

c»      oi 


(D 

S 
o 

ck 


J? 


w 


"-M      rj 
O      O 


I      * 

g  a 


e* 


53 


B 


5 
S 


12.12       FACTORIAL    TREATMENT    COMBINATIONS 


385 


TABLE   12. 20- Yields  of  Soybeans  at  the  Agronomy  Farm,  Ames,  Iowa,  1949 

{In  bushels  per  acre) 


Date  of 
Planting 

Fertilizer 

Replicate 

1 

2 

3 

4 

Early 

Check 
Aero 

Na 
K 

Check 
Aero 

Na 
K 

28.6 
29.1 
28.4 
29.2 

30.3 
32.7 
30.3 
32.7 

36.8 
29.2 
27.4 
28.2 

32.3 
30.8 
32.7 
31.7 

32.7 
30.6 
26.0 
27.7 

31.6 
31.0 
33.0 
31.8 

32.6 
29.1 
29.3 
32.0 

30.9 
33.8 
33.9 
29.4 

Late  

TABLE   12.21-Abbreviated  ANOVA  for  Data  of  Table  12,20 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean  Square 

Replicates.      .         

3 

7.31 

2.44 

4 

o-*+(8/3)  52  P* 

Dates  of  planting  

1 

32.00 

32.00 

i-i 

2 

<r2+(16/l)  2D«j- 

Fertilizers  

3 

16.40 

5.47 

3—1 

o-2+(8/3)  X)  A 

Fertilizers  X  dates  of  planting. 
Experimental  error  

3 
21 

38.40 
66.43 

12.80 
3.16 

Jb-l 

«rH-cv3)i:]fc<«0)i* 

j-i  jb—i 

Or* 

not  at  the  1  per  cent  significance  level.  [NOTE :  Depending  on  whether  we 
use  <x  =  0.05  or  ex  =  0.01,  the  recommendations  will  differ.  If  a.  =  0.05,  dif 
ferent  fertilizers  would  probably  be  suggested  for  each  date  of  planting; 
if  a:  =  0.01,  it  is  possible  that  the  same  recommendation  (concerning 
fertilizers)  would  be  made  for  each  date  of  planting.] 

Example  12.14 

An  experiment  such  as  described  in  Example  10.18  was  performed. 
The  resulting  data  are  given  in  Table  12.22.  The  associated  ANOVA  is 
presented  in  Table  12.23.  It  will  be  noted  that  none  of  the  factors  led  to 
significant  results. 

So  far  in  this  section  we  have  summarized  the  cases  involving  two 
and  three  factors  with  one  observation  per  experimental  unit.  How 
ever,  there  are  two  other  topics  associated  with  factorials  in  a  random 
ized  complete  block  design  which  also  deserve  our  attention  at  this 


386 


CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 


TABLE   12.22~Surge  Voltages  Resulting  From  the  Experiment  Described 
in  Example  10.18  and  Discussed  in  Example  12.14 


Heat 

Tempera 

Replicate 

Electrolyte 

Paper 

ture 

(a) 

(*) 

to 

I 

II 

0 

0 

0 

6.08 

6.79 

0 

0 

1 

6.31 

6.77 

0 

1 

0 

6.53 

6.73 

1 

0 

0 

6.04 

6.68 

0 

1 

1 

6.12 

6.49 

1 

0 

1 

6.09 

6.38 

1 

1 

0 

6.43 

6.08 

1 

1 

1 

6.36 

6.23 

TABLE    12.23-ANOVA  for  Data  of  Table  12.22 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

IVIean  .  .         

1 

651  .6533 

651.6533 

Repli  cates 

1 

0.2997 

0.2997 

Treatments 
A  

1 

0  .  1462 

0.1462 

2.21 

B    

1 

0.0018 

0.0018 

0.03 

C 

1 

0.0232 

0  0232 

0.35 

AB   

1 

0.0001 

0.0001 

0.00 

AC 

1 

0  .  0047 

0.0047 

0.07 

BC  

1 

0.0176 

0.0176 

0.27 

ABC  

1 

0.0883 

0.0883 

1.33 

Experimental  error  

7 

0.4632 

0.0662 

Total 

16 

652.6981 

time.  These  are:  (1)  subsampling  and  (2)  the  analysis  of  response 
curves.  Each  of  these  will  now  be  discussed. 

Subsampling  in  a  randomized  complete  block  design  which  incor 
porates  factorial  treatment  combinations  leads  to  analyses  such  as 
shown  in  Tables  12.24  and  12.25.  Since  no  new  techniques  are  involved, 
numerical  examples  will  not  be  given.  The  reader  is  referred  to  Sections 
11.12  and  12.4  for  further  details. 

As  was  mentioned  in  Section  11.12,  when  factorial  treatment  combi 
nations  are  involved,  it  is  possible  to  subdivide  the  treatment  sum  of 
squares  into  several  parts  such  as  (A.i^yy)  (Ao)vy,  •  •  •  ;  CB.L)IW* 
(Bo)w*  *  •  •  ;  (A-jJBi^yy,  {AQBL}yVy  (A  iJB  Q}  yV  y  and  so  on.  The  pro 
cedure  to  be  followed  will  parallel  that  presented  in  Section  11.11,  the 
only  difference  being  the  refinements  introduced  to  subdivide  the 
interaction  sum  of  squares.  Because  of  this,  the  technique  will  be  pre 
sented  in  terms  of  two  numerical  examples. 


12.12       FACTORIAL   TREATMENT    COMBINATIONS 


367 


TABLE   12. 2 4- Abbreviated  ANOVA  for  a  Two-Factor  Factorial  in  a 

Randomized  Complete  Block  Design  With  n  Samples  per 

Experimental  Unit 


Source  of  Variation 


Degrees  of  Freedom 


Expected  Mean  Square 


Replicates . . 

Treatments: 

A 


r—1 

a—\ 
6—1 


AB. 


ncrz-{-rn 


I-IX&-D 


Experimental  error. 
Sampling  error. 


TABLE    12.25-Abbreviated  AISTOVA  for  a  Two-Factor  Factorial  in  a 

Randomized  Complete  Block  Design  With  n  Samples  per  Experimental 

Unit  and  d  Determinations  per  Sampling  Unit 


Source  of  Variation 


Degrees  of 

Freedom 


Expected  Mean  Square 


Replicates. .  . 
Treatments: 
A 

B 


r—1 

a— I 
b-l 


AB. 


Experimental  error . 

Sampling  error 

Determinations.  .  .  . 


rab(n—l*) 
rabn(d —  1) 


Example  12.15 

Consider  the  data  of  Table  12.26.  The  first  step  in  the  analysis  is  the 
formation  of  the  aX&  table  shown  in  Table  12.27.  Remembering  that 
each  entry  in  Table  12.27  is  the  sum  of  r  =  2  observations  and  using  the 
polynomial  coefficients  given  in  Table  11.26,  we  find  that 

,  ,    ,  [(-3)(35)  +  (-1X42)  -4-  (1)(59)  +  (3)(60)]» 


(2)(3)[(-3)2  +  (-1)2  +  (I)2  +  (3)*] 

(-1X59)  +  (1X60)]* 


(I)2] 


70,53 


1.50 


388 


CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK   DESIGN 

-1)(35)  +  (3)  (42)  +  (-3)  (59)  H 

(2) (3)  [(-I)"  +  (3)2  +  (-3)2  +  (I)2] 

(0)(65)  +  (1)(67)]* 

=  .56 


5.63 


(I)2] 


(-2)(65) 


(2)(4)[(1)*+  (-2)"  + 
where  the  divisors  are: 


.02 


(1)  for    (.Ai^yy,    (Ao)yy,   and    (^L  c)w  =  (r6)  X  (sum   of  the   squares   of   the 
coefficients) ; 

(2)  for  (Bz^yy  and  (Bo)yyi(ra)  X(sum  of  the  squares  of  the  coefficients). 

TABLE  12.26-Coded  Data  for  Use  in  Illustrating  the  Calculation  of  the 

Linear,  Quadratic,  .  .  .  Effects  in  a  Two-Factor  Factorial  Experiment 

Conducted  in  a  Randomized  Complete  Block  Design 


Replicate 

<3l 

<Z2 

a3 

a4 

1  

f6i 
l&a 

7 
5 

8 
6 

9 

11 

7 
10 

2  

i 

u, 

[6l 

•^2 

4 

7 
6 

6 

9 
6 

10 

9 
10 

12 

8 
11 

1 

U3 

6 

7 

10 

12 

TABLE   12.27-aX&  Table  Formed  From  the  Data  of  Table  12.26 


ai 

a2 

a3 

a4 

Totals 

b-L  

14 

17 

18 

15 

64 

62  

11 

12 

21 

21 

65 

&3 

10 

13 

20 

24 

67 

Totals 

35 

42 

59 

60 

196 

To  illustrate  the  computation  of  the  various  sums  of  squares  which 
comprise  (AB^yy,  let  us  take  the  case  of  (AQBiJ)yy  as  an  example.  It 
should  be  understood  that  any  other  of  the  sums  of  squares  may  be 
found  in  a  similar  manner  if  the  appropriate  word-substitution,  for 
quadratic  and  linear,  is  made  in  the  next  few  sentences.  Obtain  the  sum 
of  the  products  of  the  a  quadratic  polynomial  coefficients  by  the  totals 
in  the  cells  of  the  aX&  table  for  each  level  of  &.  Then  apply  the  b  linear 
polynomial  coefficients  to  these  "sums"  and  obtain  the  usual  sum  of 
products.  Square  this  last  sum,  and  divide  the  squared  quantity  by  the 
product  of  the  sums  of  squares  of  the  two  sets  of  polynomial  coefficients 
(quadratic  for  a  and  linear  for  6)  used  in  the  computation.  Also  divide 
by  r,  the  number  of  replicates,  since  each  total  in  the  a  X  b  table  was  the 
sum  of  r  observations.  The  resulting  value  is  the  sum  of  squares  due  to 

.  For  our  numerical  example  we  have 

for  6i:     (1X14)  +  (-1X17)  +  (-1X18)  +  (1)(15)  =  -  6 


12.12       FACTORIAL    TREATMENT    COMBINATIONS 


389 


for 

for 

and  hence 

(AQBL")yy     = 


4-  (-1)(12)  +  (-1X21)  -f-  (1X21)  -  -  1 
,:     (1)(10)  4-  (-1X13)  -I-  (~D(20)  4-  (1)(24)  =  1 

2[(1)2  4-  (-1)2  4-  (-1)2  4-  (D2][(-l)2  +  (O)2  4-  (I)2] 


3.06. 


The  reader  should  verify  that  the  remaining  sums  of  squares  in  Table 
12.28  may  be  found  in  a  like  manner. 

TABLE    12.28-Abbreviated  ANOVA  for  Data  of  Table  12.26 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean  Square 

Replicates  

1 

1    50 

1   50 

Treatments  : 
AL  

1 

70.53 

70.53 

AQ  

1 

1.50 

1   5O 

Ac  

1 

5.63 

5   63 

BL  

1 

56 

56 

Bo  

1 

.02 

,02 

ALBL  

1 

25   31 

25.31 

A£,BQ  

1 

2.60 

2   6O 

AQ&L 

1 

3   06 

3  06 

A.oBo 

1 

.20 

20 

AC&L  

1 

.32 

.32 

AC&Q  

1 

2.60 

2.6O 

Experimental  error 

11 

3.5O 

.32 

Example  12.16 

Consider  next  the  data  of  Table  12.29.  To  save  time,  the  calculation 
of  the  sums  of  squares  will  be  illustrated  for  only  three  effects  :  A&,  AQB^ 
and  AQBCCL.  To  check  these  results,  the  reader  will  find  it  advantageous 
to  form  the  aX&Xc  and  the  aXb  tables.  Then, 


(0)(1404)  -f-  (1)(925)]* 


(2)(4)(6)[(-l)2  +  (O)2  -f-  (I)2] 
[(1X325)  +  (-2)(432) 


330.04 


996.67 


(-2)* 


1066.06 


(-1)2  4-  (I)2  4-  (3)«J 


)2][(-D2  +  (3)*  + 
(I)2  +  (3)2  4-  (5)2] 


-3)2  4- 


where 
D  = 

4-  (~3)24-  (-1 
and  the  divisors  are; 

(1)  for  (^1  £,)  yy '  (rbc)  X(sum  of  the  squares  of  the  coefficients); 

(2)  for    (.AQBj^yjfi  (re)  X  (product    of   the   sums    of   the    squares    of   the 
coefficients)  ; 

(3)  (AQBcCi^vy*  (r)  X  (product    of    the    sums    of    the    squares    of    the 
coefficients) . 


39O 


CHAPTER    12,    RANDOMIZED    COMPLETE   BLOCK    DESIGN 


TABLE   12.29-Hypothetical  Data  for  Use  in  Illustrating  the  Computation 

of  Certain  Sums  of  Squares  in  a  Three-Factor  Factorial,  the  Basic 

Design  Being  a  Randomized  Complete  Block 


"O  f>-rVli 

<Z] 

L 

a 

2 

a 

3 

cate 

&1 

£2 

bz 

&4 

61 

b* 

63 

?>4 

61 

b* 

bs 

&4 

1  

c\ 

C% 
£3 

7 
23 
9 

7 
18 
18 

9 

25 
24 

7 
15 
23 

15 
13 
12 

36 

35 
43 

60 
61 
62 

15 
18 
14 

24 
30 
31 

29 
26 
24 

17 
11 
15 

19 
8 
23 

2  

c± 
c*> 

LC6 
'^1 

c* 
cz 

7 
6 
10 

7 
20 
9 

13 
8 
12 

6 
19 
22 

25 
20 
30 

11 

25 
26 

36 
7 
11 

7 
16 

24 

11 
15 
10 

15 
13 
13 

12 
46 
42 

35 
30 
40 

63 
18 
27 

60 
64 
66 

26 
28 
12 

20 
20 
15 

32 
15 
17 

25 
30 

32 

15 
32 
29 

30 

25 
25 

12 
13 
8 

20 
15 
15 

5 
6 
7 

20 
10 
22 

c± 
cs 

^ 

8 
8 
9 

15 
10 
12 

26 
20 
28 

30 
8 
11 

13 
17 
8 

10 
40 

45 

66 
20 
30 

25 
30 
15 

34 
18 
19 

15 
35 
30 

15 
15 
10 

4 
5 
8 

The  numerators  for  (<Az,)i/2/  and  (AQBi^yy  were  found  by  the  same 
procedures  as  similar  quantities  in  the  preceding  example.  The  numera 
tor  for  {AoBcCi^yy  was  obtained  by  a  simple  extension  of  the  same 
principles.  In  this  case  the  extension  may  be  explained  as  follows: 
Operate  with  the  CL  coefficients  on  the  c  entries  in  the  aX&Xc  table  for 
each  level  of  b  within  each  level  of  a  and  obtain  a  set  of  sums  of  products, 
one  for  each  level  of  6  within  each  level  of  a.  Next,  use  the  Be  coeffi 
cients  and  operate  on  the  sums  just  obtained.  This  will  provide  us  with 
some  "BcCz  totals"  for  each  level  of  a.  Then  use  the  AQ  coefficients  to 
give  us  the  numerator  for  (AQBcCiJ)yy.  The  procedure  should  now  be 
clear,  and  the  extension  to  any  number  of  factors  will  be  a  simple,  even 
if  a  time-consuming,  job. 

12.13      MISSING    DATA    IN   A    RANDOMIZED   COMPLETE 
BLOCK   DESIGN 

Many  times,  even  after  considerable  effort  has  been  expended  and 
due  diligence  exercised  in  planning  an  experiment,  there  are  things 
which  occur  to  the  disadvantage  of  the  research  worker.  One  of  the 
most  common  of  these  "disturbances"  is  the  problem  of  missing  obser 
vations.  Missing  observations  arise  for  many  reasons:  An  animal  may 
die,  an  experimental  plot  may  be  flooded  out,  a  worker  may  be  ill  and 
not  turn  up  on  the  job,  a  jar  of  jelly  may  be  dropped  on  the  floor,  or 
the  recorded  data  may  be  lost.  What  effect  does  this  have  on  our 
methods  of  analysis?  Since  most  experiments  are  designed  with  at  least 
some  degree  of  balance,  or  symmetry,  any  missing  observations  will 
usually  destroy  this  balance.  Thus,  we  now  expect  our  original  planned 
analysis  to  be  complicated  and  some  modifications  in  procedure  to  be 
required.  We  could,  of  course,  in  many  instances  treat  the  data  as  a 
case  of  disproportionate  subclass  numbers  and  use  methods  of  analysis 
appropriate  to  such  situations  (see  Chapter  13).  Ho wever,  other 


12.13       MISSING    DATA  391 


approaches  are  sometimes  open  to  the  statistician,  and  we  shall  exam 
ine  these  in  this  section,  pointing  out  the  difficulties  which  arise  and 
indicating  the  computational  procedures  to  be  followed  in  each  case. 

First,  let  us  mention  two  cases  of  missing  data  in  a  randomized  com 
plete  block  design  which  present  no  difficulties  as  to  computational 
procedures:  (1)  a  complete  block  is  missing,  or  (2)  a  treatment  is 
completely  missing.  When  one  or  more  complete  blocks  are  missing 
we  simply  proceed  with  the  standard  type  of  analysis,  provided  we 
still  have  at  least  two  blocks  remaining;  that  is,  we  analyze  the  data 
as  though  we  had  planned  only  on  the  number  of  blocks  which  are 
actually  available.  For  the  case  in  which  no  data  are  available  on  one 
or  more  treatments  (assuming  we  still  have  at  least  two  treatments 
remaining),  we  may  again  proceed  in  the  regular  manner.  However, 
in  this  instance,  the  research  worker  should  certainly  inquire  into  the 
reasons  for  the  lack  of  data  on  certain  treatments.  It  is  apparent  that 
many  things  might  have  caused  such  a  happening,  each  of  which  could 
possibly  lead  to  different  decisions  or  recommendations  on  the  part  of 
the  experimenter.  Without  a  specific  example,  further  discussion  on 
such  points  can  only  be  of  a  vague  nature;  rather  than  continue  in 
general  terms,  we  shall  postpone  further  remarks  on  this  type  of 
situation  until  the  need  arises. 

A  more  commonly  occurring  situation  is  the  one  in  which  one  obser 
vation  is  missing.  Here  we  run  into  difficulty  in  our  analysis.  Either 
we  must  treat  the  data  by  methods  appropriate  to  disproportionate 
frequencies,  or  we  must  find  some  other  scheme  which  we  hope  will  be 
simpler  to  apply.  One  such  device  is  to  estimate  a  value  to  replace  the 
missing  observation  and  then  to  proceed  with  the  usual  analysis  for 
randomized  complete  block  designs.  How  does  one  obtain  an  estimate 
of  the  missing  observation?  The  estimation  procedure  currently  favored 
by  statisticians  is  to  assign  that  value  for  the  missing  observation 
which  will  minimize  the  experimental  error  sum  of  squares  when  the 
regular  analysis  is  performed.  To  mathematics  students  this  is  another 
familiar  problem  in  differential  calculus;  calling  the  missing  observa 
tion  M,  the  experimental  error  sum  of  squares  is  computed,  or  rather 
the  algebraic  expression  for  the  experimental  error  sum  of  squares  is 
formulated,  and  by  differentiating  this  expression  with  respect  to  M 
and  equating  to  0,  a  solution  may  be  obtained.  For  the  student  not 
proficient  in  mathematics,  this  procedure  may  be  summarized  by  the 
following  formula  which  will  provide  an  estimate  of  the  missing  obser 
vation  in  accordance  with  the  above  principle: 


tT  —  bB  —  S 

'  C12.36) 


where 


t==  number  of  treatments 

6  —  number  of  blocks 

T  =  sum  of  observations  with  the   same  treatment   as  the  missing 

observation 

B  =  sum  of  observations  in  the  same  block  as  the  missing  observation 
of  all  the  actual  observations. 


392 


CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 


This  value,  that  is,  M,  is  then  entered  in  the  appropriate  place  in  the 
table  of  data  and  the  augmented  data  are  analyzed  in  the  customary 
manner. 

We  are  now  ready  to  construct  our  analysis  of  variance  table  and  to 
test  the  hypothesis  £T:ry  =  0(j  =  1,  •  -  •  ,  Z).  However,  certain  changes 
must  be  made  in  the  form  of  our  analysis  of  variance  table  if  we  are  to 
avoid  biased  results.  The  first  change  is  easy  to  apply  and  proceeds  as 
follows:  Reduce  the  degrees  of  freedom  associated  with  both  experi 
mental  error  and  total  by  1.  That  is,  the  degrees  of  freedom  for  experi 
mental  error  become  (6  —  !)(£ — 1)  —  1  and  the  degrees  of  freedom  for 
total  become  bt — 1.  The  second  change  is  a  little  more  cumbersome  to 
apply.  Before  detailing  this  change,  let  us  discuss  what  it  is  and  why  it 
is  necessary.  It  may  be  proved  that,  under  the  null  hypothesis,  the 
expected  value  of  Tw/(t—l'),  the  treatment  mean  square  calculated 
from  the  augmented  data,  is  greater  than  <r2,  the  expected  value  of  the 
experimental  error  mean  square.  Thus  any  test  of  hypothesis  which 
does  not  correct  for  this  fact  will  be  a  biased  test  and  can  only  be 
considered  approximate.  The  correction  for  this  bias,  the  second  change 
mentioned  above,  is  to  decrease  the  treatment  sum  of  squares,  Tyy, 
by  the  amount 

[B  —  (t  —  l)^f]2 
Correction  for  bias  =  Z  =  ,  (12  .37) 

*(*  -  1) 
which  gives  us  a  new  treatment  sum  of  squares 

T'yy  =   Tyy  —  Z,  (12.38) 

and  the  analysis  of  variance  indicated  in  Table  12.30  is  finally  obtained. 

TABLE   12. SO-Generalized  ANOVA  for  a  Randomized  Complete  Block 
Design  With  One  Missing  Observation 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

F-Ratio 

]VIean      

1 

Mw 

M 

Blocks        

6—1 

JStnt 

B 

Treatments    

t—  1 

Tf 

Tf 

T'/E 

Experimental  error 

(£_!)(£  __!)__  1 

EWJ 

E 

Total 

bt—l 

73  F2  —  Z 

Example  12.17 

An  experiment  was  conducted  by  Tinker  (13)  to  investigate  the 
consistency  of  blink-rates  during  reading.  Data  were  recorded  for  six 
successive  5-minute  periods  of  reading.  As  we  have  extracted  only  part 
of  the  available  data  for  our  example,  care  should  be  exercised  in 
drawing  conclusions  from  the  analysis  -which  follows.  The  original  paper 
should  be  consulted  by  those  desiring  further  information  on  the  sub 
ject  matter.  We  will  assume  that  the  experiment  was  performed  on 
four  individuals — they  will  be  our  blocks — and  the  six  periods  will 
represent  the  treatments.  Moreover,  to  illustrate  the  techniques  of  this 


12,13       MISSING    DATA 


393 


section,    we   will   assume   that   the   observation   on   Subject   A   for   the 
fourth  period  is  missing.  Our  observed  data  are  given  in  Table  12.31. 

TABLE    12.31— Number  of  Blinks  for  Successive  Five-Minute  Periods  of  Reading 


Periods 

Sub 

jects 

1 

2 

3 

4 

5 

6 

A  

24 

23 

28 

30 

41 

B  

18 

17 

17 

19 

19 

18 

C  

41 

41 

49 

39 

19 

27 

D  

46 

69 

74 

58 

54 

50 

Adapted  from  M.  A.  Tinker,  "Reliability  of  Blinking  Frequency  Employed  as  a  Measure 
of  Readability,"  Jour.  E,xp.  Psych.,  XXXV,  421. 

Substituting  ia   our  formula,    we   find   our   estimate    of  the   missing 
value  to  be 

tT  +  bB  —  S        6(116)  -f-  4(146)  —  821 


M 


30.6. 


(*-  1)(Z>-D  5(3) 

The  correction  for  bias  in  the  treatment  sum  of  squares  is  found  to  be 

[JB  -  (f  -  l)Af]»        [146  -  5(30.6)]2 

Z  =  *=!)  " 6(5)  -  ^ 

and  so  we  arrive  at  the  analysis  of  variance  presented  in  Table  12.32. 

TABLE   12. 3 2- Abbreviated  ANOVA  of  Number  of  Blinks  During  Reading 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of  Squares 

Mean  Square 

Blocks                     .... 

3 

5233.82 

1744.61 

Treatments 

5 

339.90 

67.98 

Experimental  error  

14 

1068.40 

76.29 

The  test  of  the  null  hypothesis  ^r:ry  =  0(y=  1,  -  -  *  ,  6)  gives  rise  to  an 
F-value  less  than  unity.  Noting  that  F'*=l/F  is  not  significant,  we 
conclude  that  the  blink-rate  is  consistent  during  reading  when  meas 
ured  over  six  successive  5-minute  periods.  It  is  evident  that  there  are 
wide  differences  among  individuals,  a  fact  which  is  not  surprising  and 
which  confirms  our  judgment  in  performing  the  experiment  as  we  did, 
that  is,  by  removing  the  inter-individual  differences  which  otherwise 
would  have  appeared  as  part  of  the  experimental  error  sum  of  squares. 
(NOTE:  Actually,  the  value  we  have  assumed  to  be  missing  was  re 
corded  as  27  in  the  original  source  of  data.  It  will  pay  the  reader  to  do 
the  analysis  with  the  true  value  entered  in  the  table  for  comparison  with 
the  approximate  solution  presented  above.  This  should  give  him  an 
indication,  but  only  an  indication,  of  how  reliable  the  estimation  pro 
cedure  is.) 

If  two  or  more  values  are  missing,  the  same  general  procedure  (using 
the  calculus)  may  be  followed  to  provide  estimates.  For  th.e  person 
not  familiar  with,  the  requisite  mathematical  tecliniques,  equivalent 
results  may  be  obtained  by  use  of  the  following  iterative  method. 
Suppose  that  two  values  are  missing:  For  one  of  these  substitute  the 


394  CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 

mean  of  all  recorded  observations,  and  then  estimate  the  second  mis 
sing  value  using  Equation  (12.36);  next,  place  this  estimate  in  its 
proper  place  in  the  table,  remove  the  general  mean  from  its  position 
in  place  of  the  first  missing  observation,  and  then  estimate  a  value 
for  the  first  missing  observation  using  Equation  (12.36).  After  about 
two  cycles  you  will  find  very  little  or  no  change  in  successive  estimates 
of  the  same  missing  value.  When  this  point  is  reached,  you  have  the 
estimated  values.  This  procedure  may  easily  be  extended  to  cases  where 
three  or  more  observations  are  missing. 

What  changes  are  necessary  before  one  proceeds  with  the  usual  F- 
tests  if  we  have  been  forced  to  estimate  several  missing  values?  First, 
we  must  reduce  the  degrees  of  freedom  associated  with  both  experi 
mental  error  and  total  by  the  number  of  observations  estimated.  Sec 
ond,  the  treatment  sum  of  squares  must  be  reduced  by  a  specified 
quantity  to  avoid  a  biased  test  procedure.  If  we  have  only  two  missing 
observations  (not  in  the  same  block),  the  necessary  correction  for  bias 
is  given  by 

-          \_B>  -  (jt  -  1)M']*  +  [B"  -(jt-  VM"\* 

Z,   =  }  ( 1 2  .  o  9 ) 

t(t  -  i) 

where 

t  —  number  of  treatments 
B'  =  total  of  all  the  observations  in   the    same   block   as   the    first 

missing  observation 
J5"  =  total  of  all  the  observations  in  the  same  block  as  the  second 

missing  observation 

M'  =  estimate  of  the  first  missing  observation. 
M"  =  estimate  of  the  second  missing  observation. 

If  more  than  two  observations  are  missing,  or  if  two  observations  are 
missing  in  the  same  block,  a  formula  giving  the  correction  for  the  bias 
in  the  treatment  sum  of  squares  may  be  found  in  Yates  (14,  15). 

Problems 

12.1  The  folio  wing  data  are  from  an  experiment  involving  a  randomized 
complete  block  design.  Complete  the  appropriate  analysis  of  vari 
ance,  and  test  the  hypothesis  that  the  true  effects  of  the  four  treat 
ments  are  equal.  State  all  your  assumptions. 


PROBLEMS 


395 


Treat 

ment 

Block 

1 

2 

3 

4 

1 

20 

18 

16 

17 

2 

18 

18 

16 

2O 

3  .  . 

20 

18 

17 

18 

4  

20 

16 

20 

17 

5  

19 

16 

16 

2O 

12.2  In  a  randomized  complete  block  experiment  with  5  treatments  in  10 
replications,  the  variance  among  the  5  treatment  means  was  100. 
Complete  the  following  abbreviated  ANOVA,  and  test  the  hypothesis 
that  the  5  treatment  effects  are  the  same. 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 

Square 

Replicates  

90 

Treatments  

Experimental  error  .  .  . 

5 

12.3  Upon  calculating  the  analysis  of  variance  of  the  yields  of  6  varieties 
planted  in  8  randomized  complete  blocks,  the  3  sums  of  squares,  for 
varieties,    for    blocks,    and    for    experimental    error    (or    remainder), 
were    each    245.    Complete,    as    far    as    is    possible,    the    appropriate 
analysis    of   variance,    and    compute    a    value    of    F   for    testing    the 
significance  of  the  differences  among  varieties.  Interpret  your  result 
in  terms  of  the  appropriate  model,  and  give  your  conclusions. 

12.4  Examine  the  results  given  below  to  learn  about  the  effectiveness  of 
chalk  and  lime  applications  in  neutralizing  soil  acidity  and  thus  in 
creasing  the  stand  of  beets. 


Number  of  Beets  per  Plot 


Block 

Control 

Chalk 

Lime 

1  

49 

135 

147 

2 

37 

151 

131 

3 

114 

143 

103 

4         ... 

140 

146 

147 

396  CHAPTER    12,   RANDOMIZED    COMPLETE    BLOCK   DESIGN 

12.5       Analyze  the  data  in  the  following  table  and  interpret  the  results. 

RATIO  OF  DRY  TO  WET  WHEAT 


Nitrogen 

Applied 

Block 

None 

Early 

Middle 

Late 

1 

0.718 

0.732 

0.734 

0.792 

2  

0.725 

0.781 

0.725 

0.716 

3  

0.704 

1.035 

0.763 

0.758 

4  

0.726 

0.765 

0.738 

0.781 

12.6  To  study  the  relative  efficiencies  of  5  different  types  of  filter,  an  ex 
periment  is  to  be  performed  using  a  certain  brand  of  oil.  Fifteen 
quarts  of  oil  (in  1-qt.  tins)  are  purchased  and  the  same  amount  of 
foreign  material  is  added  to  each  quart.  Since  only  5  tests  can  be 
performed  in  any  one  day,  we  proceed  as  follows:  (1)  allocate,  at 
random,  the  15  quarts  into  three  groups  of  5  each;  (2)  allocating  the 
groups  to  the  days,  assign  the  treatments  at  random  to  the  quarts 
within  groups;  (3)  perform  the  experiment;  and  (4)  collect,  analyze 
and  interpret  the  data. 

AMOUNT  or  FOREIGN  MATERIAL  CAUGHT  BY  FILTER 


Type  of  Filter 

"Rlnr^L-c 

(Days) 

A 

B 

C 

D 

E 

1 

16.9 

18.2 

17.0 

15.1 

18.3 

2 

16.5 

19.2 

18.1 

16.0 

18.3 

3 

17.5 

17.1 

17.3 

17.8 

19.8 

12.7  In  a  paired  experiment  there  were  10  pairs  with  the  sum  of  the 
squares  of  the  deviations  of  the  differences  from  their  mean  being 
]£d2  =  360.  The  totals  for  the  2  treatments  were  7^  =  160  and 
2^=  120.  Complete  the  following  abbreviated  ANOVA. 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Pairs  or  replications 

100 

Treatments  ,  ,  .  . 

Experimental  error  

PROBLEMS 


397 


12.8  Discuss  the  following  statement:  "If  only  one  sample  is  obtained  from 
each,  experimental  unit,   e.g.,   if  one  small  sample  is  taken   from  a 
field  plot  to  estimate  the  effect  on  the  whole  plot,  n  is  set  equal  to  1 
in  Table  12.6,  and  the  line  for  sampling  error  is  omitted.  However, 
if  the  whole  plot  in  our  field  plot  example  is  harvested,   then  the 
sampling  error  is  reduced  to  0,  and  we  have  an  analysis  as  in  Table 
12.2." 

12.9  In  a  randomized  complete  block  experiment  on  the  accuracy  of  de 
termination  of  ascorbic  acid  concentration  in  turnip  greens  (Heinze- 
Kanapaugh  method),  4  weights  of  sample  were  tried  in  5  replications. 
Two  determinations,  A.  and  B,  were  made  on  each  sample.  The  results 
(in  micrograms  per  milliliter  of  filtrate)  were  as  follows : 


Sample 
Weight 
(Grams) 

Replication 

1 

2 

3 

4 

5 

A 

B 

A 

B 

A 

B 

A 

B 

A 

B 

5  

34.2 
12.8 
5.8 
3.5 

37.2 
12.8 
8.2 
3.5 

47.0 
21.5 
10.2 
5.0 

52.5 
22.0 
13.0 
6.0 

48.5 
24.5 
16.5 
9.8 

46.5 
23.0 
11.0 
6.8 

44.2 

17.8 
9.5 

5.2 

44.2 
17.8 
15.2 
3.5 

42.5 
17.0 
11.0 
3.8 

43.5 
17.5 
10.5 
4.7 

2  

1  

0.5           

Complete  the  analysis  of  variance  for  these  data. 
12.10      Given  the  following  abbreviated  ANOVA: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Expected 
Mean 
Square 

Replicates  

3 

176 

Treatments  

7 

352 

"KxTvcTimeTi'fcaJ.  error              .  . 

21 

88 

Sampling  error         

96 

40 

Determinations  

256 

10 

(a)  Give  the  experimental  error  mean  square  in  the  above  analysis 
for  the  following: 

(1)  if  10  had  been  added  to  each  determination, 

(2)  if  each  determination  had  been  multiplied  by  10. 

(6)  Fill  in  the  expected  mean  squares  in  the  above  table,  assuming  we 
are  interested  in  just  these  8  treatments  but  that  replicates, 
samples,  and  determinations  may  be  considered  as  random 
variables. 


398  CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 

12,11      Given  the  following  abbreviated  ANOVA: 


12.12 


12.13 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Blocks  

3 

Treatments  

8 

Experimental  error  
Samples  within  plots.  .  .  . 

24 
144 

1084 
381 

if 


(a)  What  is  the  construction  of  the  experiment? 

(b)  What  is  the  variance  of  a  treatment  mean? 

(c)  Give  the  answer  to  (6)  if  we  have  only  1  sample  per  plot. 

(d)  What  is  the  maximum  precision  obtainable  by  sampling;  i.e., 
we  take  k  samples,  k — >  co  ? 

We  conducted  a  field  experiment  to  estimate  the  effect  of  9  fertilizers 
on  the  yield  of  oats.  Instead  of  harvesting  each  plot  completely,  we 
took  12  samples,  3  by  3  feet,  from  each  plot.  The  abbreviated 
ANOVA  is  as  follows: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Replicates  

3 

384 

Xreatrnents  

8 

960 

Experimental  error,     

24 

192 

Among  samples  within  plots  

396 

24 

(a)  Assuming  that  the  components  of  variance  do  not  change,  esti 
mate  the  gain  or  loss  in  information  in  the  above  experiment,  had 
6  replicates  been  used  with  8  samples  per  plot. 

(6)  What  would  the  above  mean  squares  be  if  the  analysis  of  variance 
had  been  computed  using  the  totals  of  the  12  samples  in  each 
plot? 

Given  the  following  abbreviated  ANOVA: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Expected 
Mean 
Square 

Replicates  

3 

288 

Treatments  

7 

432 

Experimental  error  

21 

144 

Among  samples  within  experimental  units  .  . 
Among  detenriinatioris  per  sample.  ........ 

96 
256 

72 
6 

(a)   Compute  the  variance  of  a  treatment  mean. 
(6)    Give  the  expected  mean  squares. 


PROBLEMS 


399 


12.14 


(c)  Compute  the  gain  or  loss  In  efficiency,  or  information,  had  6  repli 
cates  been  used  with  8  samples  from  each  experimental  unit  and 
1  determination  per  sample. 

(d)  Give  the  experimental  error  mean  square  for  the  following: 

(1)  if  10  had  been  added  to  each  determination, 

(2)  if  each  determination  had  been  multiplied  by  10. 

(e)  Test  the  hypothesis  that  there  are  no  differences  among  the  true 
effects  of  the  eight  treatments. 

A  chemist  is  confronted  with  the  problem  of  just  where  he  should 
expend  his  efforts  in  the  following  situation:  A  series  of  8  soil  treat 
ments  are  applied  in  a  randomized  complete  block  design  with  2  repli 
cations,  3  soil  samples  from  each  plot  are  taken  in  the  field,  each 
sample  is  divided  into  2  portions  in  the  laboratory,  and  duplicate  de 
terminations  for  each  portion  are  analyzed  for  nitrate  nitrogen.  The 
following  mean  squares  are  given: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Treatments.  .  .                  .                ..... 

7 

11700 

Experimental  error    .     

7 

1300 

Samples  within  pints  ,,,,,. 

32 

100 

Portions  within  samples    .          .  .       .... 

48 

20 

Determinations  within  portions 

96 

16 

12.15 


Find  the  expected  mean  squares  and  estimate  the  variance  compo 
nents.  What  might  be  his  gain  or  loss  in  efficiency  in  future  experi 
ments  if  he  used  6  replicates,  but  still  continued  to  run  only  24 
analyses  per  treatment,  e.g.,  2  samples  per  plot,  2  portions  per  sample 
and  1  determination  per  portion? 

In  an  experiment  to  test  the  effect  of  6  treatments  on  some  soil 
characteristic,  we  obtained  the  following  abbreviated  ANOVA.  A 
total  of  6  soil  samples  was  selected  at  random  from  each  plot,  and 
2  chemical  determinations  were  made  of  each  sample. 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Replicates  

4 

240 

Treatments    .  ,  

5 

360 

Experimental  error            

20 

120 

Samples  within  plots    

150 

60 

Determinations  per  sample  

180 

4 

(a)   Compute  the  variance  of  a  treatment  mean  (per  determination). 

(6)  Estimate  the  gain  or  loss  in  efficiency  in  the  above  experiment  if 
we  had  taken  8  samples  per  plot  and  had  made  only  1  determi 
nation  per  sample. 


4OO 


CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 


12.16      Given  the  following  abbreviated  A1STOVA  for  a  randomized  complete 
block  design: 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Expected 
Mean  Square 

Blocks  

9 

.4074 

Treatments.  .  

3 

1.1986 

Experimental  error.  .  . 

27 

,6249 

(a)   Complete  the  analysis;  fill  in  expected  mean  squares. 
(6)    Estimate  the  efficiency  of  this  design  relative  to  a  completely 
randomized  design. 

(c)  Compute  the  standard  error  for  a  treatment  mean  and  for  the 
difference  between  2  treatment  means. 

(d)  The  treatment  means  are  1.464,  1.195,  1.325,  and  1.662.  What 
mean  or  means  do  you  suspect  might  represent  different  popula 
tions? 


12.17  The  following  data  give  the  gains  in  weight  of  pigs  in  a  comparative 
feeding  trial.  Analyze  and  interpret  the  data,  paying  attention  to 
the  comparison  of  Rations  I,  II,  and  III  with  Rations  IV  and  V. 

GAINS  03?  PIGS  iisr  A  COMPARATIVE  FEEDING  TRIAL 


Replicate 

Ration  I 

Ration  II 

Ration  III 

Ration  IV 

Ration  V 

1 

165 

168 

164 

185 

201 

2  

156 

180 

156 

195 

189 

3 

159 

180 

189 

186 

173 

4  

167 

166 

138 

201 

193 

5  

170 

170 

153 

165 

164 

6  

146 

161 

190 

175 

160 

7     

130 

171 

160 

187 

200 

8  

151 

169 

172 

177 

142 

9 

164 

179 

142 

166 

184 

10  . 

158 

191 

155 

165 

149 

PROBLEMS 


401 


12.18 


12.19 


12.20 


The  following  data  are  extracted  from  a  larger  experiment  concerned 
with  oat-seed  treatment.  The  following  yields  in  grains  were  obtained 
with  2  rates  of  the  same  compound  over  7  replicates: 


R£ 

Lte 

Replicate 

1 

2 

Check 

1  

360 

391 

408 

2  

436 

382 

409 

3  

413 

414 

340 

4. 

353 

416 

324 

5  

328 

375 

304 

6  

269 

422 

268 

7  

220 

227 

290 

What  conclusion  do  you  draw?  In  a  separate  column  are  the  yields 
for  the  untreated  seed.  Is  seed  treatment  worth  the  added  expense  in 
this  instance? 

Results  similar  to  those  in  Problem  12.18  are  also  available  for  flax. 
What  advice  would  you  give  about  the  use  of  Ceresan  M  as  against 
224,  and  about  the  advisability  of  seed  treatment? 


Replicate 

Ceresan  M 

224 

Check 

1  

19.2 

14.4 

13.2 

2  

14.8 

24.6 

19.2 

3  

26.7 

22.9 

17.4 

4    

17.6 

22.7 

16.4 

5  

22.1 

22.0 

15.8 

6  

21.7 

22.0 

14.6 

7  

23.9 

20.4 

12.5 

8  

19.1 

16.0 

13.0 

A  project  studying  farm  structures  was  concerned  with  the  insulation 
of  poultry  houses.  The  data  obtained  from  a  study  of  a  set  of  model 
structures  (total  number  of  eggs  over  4  replicates  of  each  treatment) 
were  as  follows: 

Standard  house+laying  mash 250 

3"  wall  insulation+laying  mash 280 

3"  wall  insulation+laying  mash+cod  liver  oil 350 

6"  wall  insulation+laying  mash 310 

6"  wall  insulation+laying  mash+cod  liver  oil 400 

Construct  a  reasonable  set  of  4  orthogonal  comparisons  based  on  the 
above  treatments.  Calculate  the  sum  of  squares  for  one  of  your  com 
parisons  and  test  for  significance.  The  following  is  part  of  the  original 
analysis: 


4O2 


CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Treatments  

4 

3470 

Experimental  error       

12 

1728 

12.21  Assume  a  randomized  complete  block  experiment  with  4  treatments 
and  8  replicates.  One  of  the  treatments  is  a  control  or  check  and  the 
other  3  are  different  methods  of  treatment.  Assume  that  the  mean 
effect  of  all  32  experimental  units  is  40,  that  the  mean  effect  for  the 
control  is  34,  and  that  the  mean  effect  for  method  B  is  42.  Also  the 
following  abbreviated  ANOVA  is  given: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Replicates  

7 

32 

Treatments  

3 

64 

Experimental  error  

21 

16 

(a)   What  is  the  experimental  error  variance  per  experimental  unit? 
(6)    Compute  the  coefficient  of  variation  per  experimental  unit, 

(c)  Compute  the  variance  of  a  treatment  mean. 

(d)  Is  the  difference  between  the  mean  effects  of  control  and  method 
B  significant  at  the  1  per  cent  level? 

(e)  Compute  and  interpret  the  95  per  cent  confidence  interval  esti 
mate  of  the  mean  difference  between  the  control  and  method  B. 

12.22     An  experiment  was  conducted  to  assess  the  relative  merits  of  5  dif 
ferent  gasolines.  Since  vehicle  to  vehicle  variations  in  performance 
are  inevitable,  the  test  was  run  using  5  cars,  hereafter  called  blocks. 
The  following  descriptions  of  the  5  gasolines  are  available: 
A:   control 

B:   control+ additive  X  manufactured  by  company  I 
C:   control  +  additive  Y  manufactured  by  company  I 
D:  control -4-additive  IT  manufactured  by  company  II 
E:   control  +  additive  V  manufactured  by  company  II. 
The  data,  in  miles  per  gallon,  are  given  below.  Please  analyze  and 
interpret  the  data. 


Blocks  (Cars) 

Treatments 

(Gasolines) 

1 

2 

3 

4 

5 

A 

22 

20 

18 

17 

19 

B 

28 

24 

23 

19 

25 

C 

21 

23 

25 

25 

27 

D 

26 

21 

21 

22 

20 

E 

27 

25 

22 

20 

24 

PROBLEMS 


4O3 


12.23 


12.24 


12.25 


12.26 


12.27 


Subdivide  the  experimental  error  sum  of  squares  in  each,  of  the  follow 
ing  problems  in  accordance  with  the  principles  given  in  Section  12.9: 

(a)  12.4  (d)   12.19 

(b)  12.17  (e)    12.20 

(c)  12.18  (f)    12.22 

Using  the  technique  presented  in  Section  11.10,  analyze  further  the 
data  given  in  the  following  problems: 

(a)  12.1  (d)   12.6  (g)    12.19 

(b)  12.4  (e)    12.17  (h)  12.20 

(c)  12.5  (f)    12.18  (i)     12.22 

In  Table  12.31  we  presented  some  data  on  blinking  rates  in  successive 
5-minute  periods  of  reading.  After  substituting  for  a  missing  observa 
tion,  these  data  were  analyzed  using  a  randomized  complete  block 
design.  Ignoring  the  fact  that  we  had  to  estimate  a  missing  observa 
tion,  discuss  critically  the  use  of  a  randomized  complete  block  design 
in  analyzing  data  of  this  type.  If  you  feel  that  the  use  of  a  random 
ized  complete  block  design  was  unjustified,  state  reasons  to  support 
your  contention  and  give  what  you  believe  to  be  an  appropriate 
method  of  handling  such  data.  Examine  all  your  assumptions 
carefully. 

Five  levels  of  fertilizer,  0,  10,  20,  30,  and  40,  were  applied  to  corn  in  a 
randomized  complete  block  design.  A  preliminary  analysis  of  vari 
ance  gave  the  following  results: 


d.f. 

Replicates 4 

Fertilizers 4 

Experimental  error 16 


M.S. 
2500 
2800 
1500 


The  sums  of  the  yields  in  the  5  plots  of  each  level  were: 
Level  0  10  20  30  40 


Total  yield 


20 


140 


260 


300 


280 


What  additional  computations  would  you  make  to  interpret  the 
effect  of  treatments?  Make  these  computations,  and  interpret  the 
results. 

The  strength  index  of  cotton  fibers  was  thought  to  be  affected  by  the 
application  of  potash  to  the  soil.  A  randomized  complete  block  experi 
ment  was  conducted  to  get  evidence.  Here  is  a  summary  of  the  plot 
strength  indexes: 


Treatment 

/"|~>                                -J               j-v-C        "C?""        /*"% 

Replications 

^Jrounas  ot  &.2.(J 
per  Acre) 

1 

2 

3 

36         

7.62 

8.00 

7.93 

54    

8.14 

8.15 

7.87 

72    

7.76 

7.73 

7.74 

108    

7.17 

7.57 

7.80 

144  

7.46 

7.68 

7.21 

404 


CHAPTER    12,  RANDOMIZED    COMPLETE   BLOCK  DESIGN 


12.28 


Analyze  the  data.  Plot  the  mean  strength  index  for  each  treatment, 
Y,  against  the  pounds  of  fertilizer  per  acre,  X.  The  sum  of  squares 
attributable  to  regression  is  0.5662  with  1  degree  of  freedom  (verify 
this).  Subtract  this  from  your  sum  of  squares  for  treatments  (4 
degrees  of  freedom).  The  remainder  (3  degrees  of  freedom)  is  the 
sum  of  squares  of  deviations  from  regression.  Complete  the  analysis 
of  variance.  Test  the  hypothesis  of  0  regression.  What  conclusions 
do  you  draw? 

The  following  are  the  yields  (tons  per  acre)  of  sugar  beets  on  plots 
which,  2  years  earlier,  had  been  treated  with  lime: 


TV*a.a  frn  pin  f 

Replications 

(Tons  per  Acre) 

1 

2 

3 

4 

5 

1            

13.7 

13.3 

12.6 

14.7 

10.8 

2  

16.9 

17.1 

14.7 

15.7 

15.4 

3..  .  

17.3 

17.1 

16.9 

16.2 

14.6 

4  

17.8 

16.5 

17.9 

15.7 

16.3 

12.29 


Analyze  the  data.  Test  the  hypothesis  that  there  is  no  effect  of 
treatment.  Plot  the  treatment  means,  Y,  against  the  rate  of  applica 
tion  of  Erne,  X.  Do  you  think  the  regression  is  linear?  As  evidence, 
divide  the  sum  of  squares  for  treatments  into  2  parts :  attributable  to 
regression,  35.52;  and  remainder.  Test  the  null  hypothesis  that  there 
is  no  deviation  from  linear  regression.  Instead  of  thinking  about  re 
gression,  you  might  have  divided  the  treatment  sum  of  squares  into 
these  2  parts:  (1)  due  to  difference  between  mean  of  first  treatment 
and  mean  of  the  other  3  combined,  43.02;  and  (2)  differences  among 
means  of  the  last  3  treatments.  What  conclusions  do  you  reach? 
Consider  an  experiment  to  assess  the  relative  effects  of  4  different 
treatments  (i.e.,  packing  pressures)  on  the  function  time  of  a  certain 
explosive  actuator.  Casings  are  available  from  4  different  production 
lots.  Four  casings  were  randomly  selected  from  each  of  the  lots  and 
the  treatments  were  assigned  at  random  within  each  lot.  Given  the 
data  shown  below  (operation  time  in  milliseconds),  analyze  and 
interpret  the  results. 


Packing  Pressures  (psi) 

Blocks 

(Lots) 

10,000 

20,000                  30,000 

40,000 

1 

12 

17                           10 

12 

2 

11 

16                             9 

11 

3 

10 

15                             8 

11 

4 

9 

15                             8 

10 

PROBLEMS 


405 


12.30     Analyze  and  interpret  the  following  data  on  yields  of  sweet  potatoes 
obtained  with  various  combinations  of  fertilizer    (n  =  N,   p  —  PzOs, 


Replicate  1 

Replicate  2 

npk 

Yield 

npk 

Yield 

npk 

Yield 

npk 

Yield 

npk 

Yield 

npk 

Yield 

133 

45 

211 

39 

333 

70 

212 

83 

211 

56 

133 

65 

111 

34 

313 

62 

311 

40 

221 

52 

321 

49 

112 

48 

221 

42 

222 

65 

212 

45 

322 

65 

333 

92 

311 

56 

323 

69 

233 

92 

132 

53 

313 

101 

122 

75 

332 

79 

213 

58 

123 

56 

121 

54 

111 

50 

312 

86 

213 

95 

331 

51 

332 

91 

223 

69 

331 

61 

232 

74 

222 

81 

232 

72 

131 

73 

322 

85 

132 

89 

223 

109 

231 

84 

122 

56 

112 

55 

113 

60 

123 

90 

113 

68 

323 

103 

312 

82 

321 

75 

231 

78 

233 

122 

131 

98 

121 

64 

Totals 

509 

608 

554 

713 

707 

675 

Grand 

Totals 

1671 

2095 

12.31      Given  the  following  abbreviated  ANOVA: 


Source  of  Variation 

Degrees  of 
Freedom 

Mean 
Square 

Replicates  

4 

70 

Treatments  : 
A                   

3 

50 

B                   

3 

160 

AB              

9 

40 

Experimental  error  

60 

10 

Interpret  the  effects  of  a  and  &  assuming  that: 

(1)  the  various  levels  of  both  a  and  6  are  fixed  or  selected; 

(2)  the  various  levels  of  both  a  and  6  are  random  variables; 

(3)  the  levels  of  a  are  fixed,  but  the  levels  of  &  are  random; 

(4)  the  levels  of  a  are  random,  but  the  levels  of  &  are  selected. 

12.32  Mr.  X  sprayed  apple  leaves  with  different  concentrations  of  a  nitro 
gen  compound,  then  determined  the  amounts  of  nitrogen  (ing.  per 
sq.  dcm.)  remaining  on  the  leaves  immediately  and  at  two  subsequent 
times.  The  object  was  to  learn  the  rate  at  which  the  nitrogen  was  ab 
sorbed  by  the  leaves.  There  were  two  replications  of  each  treatment. 
The  first  entry  in  each  cell  of  the  table  is  for  the  first  replication. 


4O6 


CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK    DESIGN 


Time 

Levels  of  Nitrogen 

n\ 

n* 

ns 

t0  

2.29 

2.24 

0.46 
0.19 

0 
0.26 

6.50 
5.94 

3.03 
1.00 

0.75 
1.16 

8.75 
9.52 

2.49 
2.04 

1.40 
1.81 

ti  

tz  

12.33 


Obtain  the  analysis  of  variance  which,  subdivides  the  8  degrees  of 
freedom  for  treatments  into  individual  comparisons:  N&,  NQ,  TL, 
TQ,  NLTL,  NLTQ,  NQTL,  and  NQTQ. 


23  FACTORIAL  FIELD  PLAN  WITH  YIELDS 


Replicate  1 

Replicate  2 

Replicate  3 

Replicate  4 

(1)      7          b  24 
abc  39        ac  31 
a  30          c  21 
be  27       ab  39 

ab  36        be  31 
(1)    19        ac  36 
abc  41          b  30 
c  30          a  33 

a  28        ac  31 
c  24          b  19 
ab  35       (1)   13 
be  26      abc  36 

abc  66       (1)    11 
a  31         be  29 
c  21        ac  33 
&  25        <zZ>  43 

12.34 


Complete  the  analysis  of  variance,  computing  the  treatment  sum  of 
squares  for  each  of  the  individual  treatment  effects,  and  subdividing 
the  experimental  error  corresponding  to  the  subdivision  of  the  treat 
ment  sum  of  squares. 


ANALYSIS  or  VARIANCE 


Source  of 
Variation 

Degrees 
of  Free 
dom 

Mean 

Square 

Replicates  
A  

3 
1 

192 
100 

B  

1 

2500 

AB  

1 

900 

Experimental 
error  

9 

32 

TABLE 


<Z0 

#1 

Sum 

&o    - 

120 

80 

200 

bl  

160 

240 

400 

Sum 

280 

320 

600 

(1)  Interpret  the  effects  of  a  and  6  assuming  both  are  fixed  variates. 

(2)  Compute  and  interpret  the  95  per  cent  confidence  interval  esti 
mate    of    the    true    mean    difference    between    treatments 
and 


PROBLEMS 


4O7 


12.35      The  following  yields  of  grass  -were  reported  for  one  year  in  dry  matter 
per  1/57-acre  plots.  This  was  a  randomized  complete  block  design. 


El 
(Ha 

ephant  Gr£ 
rvests  per  y 

LSS 

ear) 

Gu 
(Ha 

ate  mala  Gi 
rvests  per  y 

•ass 
ear) 

Blocks 

2 

3 

4 

2 

3 

4 

1  

109 

222 

187 

277 

246 

252 

2  

97 

125 

163 

293 

263 

181 

3 

133 

134 

143 

260 

194 

224 

4 

113 

173 

179 

325 

190 

248 

Discuss  the  complete  2X3  factorial  experiment,  displaying  the  perti 
nent  estimates;  outline  tentative  conclusions  before  making  the 
analysis  of  variance  and  tests  of  hypotheses. 

12.36  Analyze  and  interpret  the  following  set  of  experimental  data:  crop — 
oats;  location — Flathus,  Correctionville;  year — 1944;  comment — 
yield  in  bushels  per  acre. 


Replicate 

'  1  *r^c*  "f~TT"l  f-T~l  "f" 

Treatment 

1 

2 

3 

Total 

/yt\fp\k\               .  . 

32.2 

33.9 

34.6 

100.7 

7^27?l™l             .... 

37.4 

40.9 

38.9 

117.2 

J^1??2&1              .... 

30.6 

39.4 

33.8 

103.8 

fl%fp%k\  ...        ... 

52.4 

48.0 

43.9 

144.3 

>¥l\'P\k%         

29.9 

34.5 

36.5 

100.9 

n%'f)~\mfe%     

42.2 

29.9 

34.1 

106.2 

^^l7^2^2 

31.8 

32.5 

34.2 

98.5 

46.6 

49.5 

46.7 

142.8 

Total 

303.1 

308.6 

302.7 

914.4 

12.37  An  experiment  was  conducted  to  assess  the  effects  of  3  raw  material 
sources  (i.e.,  suppliers)  and  4  mixtures  (i.e.,  compositions)  on  the 
crushing  strength  of  concrete  blocks.  Twenty-four  blocks  were  se 
lected,  2  at  random  from  those  manufactured  by  each  of  the  12 
treatments,  and  the  experiment  was  conducted  as  a  randomized  com 
plete  block  with  2  replicates.  The  resulting  data  are  given  below. 
Analyze  and  interpret. 


4O8 


CHAPTER    12,    RANDOMIZED    COMPLETE    BLOCK   DESIGN 


Suppliers 

Replicate 

Mixtures 

^1 

B 

C 

D 

1 

1 
2 

57 
46 

65 
73 

93 
92 

102 
108 

2 

1 
2 

26 
38 

44 
67 

81 
90 

96 
99 

3 

1 
2 

39 
40 

57 
60 

96 
100 

105 
116 

12.38  The  following  is  a  randomized  complete  block  design  with  two  missing 
plots.  Fill  in  estimates  for  the  missing  values,  and  complete  the 
analysis  of  the  data. 


Trea 

tment 

Block 

1 

2 

3 

4 

Block  Totals 

1  

43 

35 

37 

42 

157 

2  

45 

39 

40 

47 

171 

3.        ... 

42 

30 

M" 

43 

115  +M" 

4  

Mf 

43 

48 

49 

140+-34T 

5  

41 

34 

36 

44 

155 

Treatment 
Totals  

1714-M' 

181 

161+lf" 

225 

73S+M'+M" 

References  and  Further  Reading 

1.  Anderson,  R.  L.,  and  Bancroft,  T.  A.  Statistical  Theory  in  Research.  McGraw- 
Hill  Book  Company,  Inc.,  New  York,  1952. 

2.  Brownlee,  K.  A.  Statistical  Theory  and  Methodology  in  Science  and  Engineer 
ing.  John  Wiley  and  Sons,  Inc.,  New  York,  1960. 

3.  Chew,  V,  (editor)  Experimental  Designs  in  Industry.  John  Wiley  and  Sons, 
Inc.,  New  York,  1958. 

4.  Cochran,  W.  G.,   Catalogue  of  uniformity  trial  data.  Jour.  Roy.  Stat.  Soc. 
(SuppL),  4  (No.  2)  1937. 

5    f  and  Cox,  G.  M.  Experimental  Designs.  Second  Ed.  John  Wiley  and 

Sons,  Inc.,  New  York,  1957. 

6.  Davies,  O.  L.   (editor)    The  Design  and  Analysis  of  Industrial  Experiments. 
Second  Ed.  Oliver  and  Boyd,  Edinburgh,  1956. 

7.  Dixon,  W.  J.,  and  Massey,  F.  J.  Introduction  to  Statistical  Analysis.  Second 
Ed.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1957. 

8.  Federer,  W.  T.  Experimental  Design.  Macmillan  Co.,  New  York,  1955. 

9.  Kempthorne,  O.  The  Design  and  Analysis  of  Experiments.  John  Wiley  and 
Sons,  Inc.,  New  York,  1952. 

10.   Paull,  A.  E.  On  a  preliminary  test  for  pooling  mean  squares  in  the  analysis 
of  variance.  Ann.  Math.  Stat.,  21:539,  1950. 


REFERENCES   AND   FURTHER    READING  4O9 

11.  Quenouille,    M.  H.   The  Design  and  Analysis  of  Experiment.  Charles  Griffin 
and  Co.  Ltd.,  London,  1953. 

12.  Snedecor,  G.  W.  Statistical  Methods.  Fifth  Ed.  The  Iowa  State  University 
Press,  Ames,  1956. 

13.  Tinker,  M.  A.  ^Reliability  of  blinking  frequency  employed  as  a  measure  of 
readability.  Jour.  Eocptl.  Psych.,  35:418,  1945. 

14.  Yates,  F.  The  analysis  of  replicated  experiments  when  the  field  results  are 
incomplete:.  Emp.  Jour.  Exptl.  Agr.,  1:129,  1933. 

15.   .  Incomplete  randomized  blocks.  Ann.  Eugen.,  7:121,  1936. 


CH  APTE  R    13 

OTHER  DESIGNS 

THE  COMPLETELY  RANDOMIZED  and  randomized  complete  block  designs 
discussed  in  Chapters  11  and  12,  respectively,  are  only  two  of  the 
many  useful  statistical  designs  that  have  been  developed  for  special 
situations.  Unfortunately,  not  all  of  the  available  designs  can  be  dis 
cussed  in  this  book.  However,  after  considering  such  factors  as  fre 
quency  of  use  and  potential  contribution  to  more  efficient  experimen 
tation,  a  select  group  of  designs  and  analysis  techniques  has  been 
chosen  for  presentation  in  this  chapter.  Persons  desiring  information 
on  other  designs  should  consult  a  professional  statistician  and/or  refer 
to  the  references  at  the  end  of  this  chapter. 

13.1      LATIN   AND   GRAECO-LATIN   SQUARES 

The  Latin  square  (LS}  design  is  frequently  used  in  agricultural  and 
industrial  experimentation.  It  is  a  special  design  that  permits  the  re 
searcher  to  assess  the  relative  effects  of  various  treatments  when  a 
double  type  of  blocking  restriction  is  imposed  on  the  experimental 
units.  Viewed  in  this  way,  the  Latin  square  design  is  a  logical  extension 
of  the  randomized  complete  block  design  and  two  examples  should  be 
sufficient  to  illustrate  the  ideas  involved. 

Example  13.1 

Suppose  we  have  5  fertilizer  treatments  to  be  investigated  and  25 
plots  available  for  experimentation.  If  the  soil  shows  a  fertility  trend 
in  two  directions  (say  N— >S  and  E— >W),  it  would  seem  reasonable  to 
set  up  blocks  of  (5)  plots  in  bath  directions.  This  is  precisely  what  is 
done  under  the  names  rows  and  columns.  The  treatments  are  then 
applied  at  random,  subject  to  the  restriction  that  each  treatment  appear 
but  once  in  each  row  and  each  column. 

Example  13.2 

Consider  the  problem  of  testing  4  machines  to  see  if  they  differ  sig 
nificantly  in  their  ability  to  produce  a  certain  manufactured  part.  It  is 
well  known  that  different  operators  and  different  time  periods  in  the 
work  day  will  have  an  effect  on  production.  Thus,  we  set  up  4  operators 
as  "columns"  and  4  time  periods  as  "rows"  and  then  assign,  at  random, 
the  machines  to  the  various  cells  in  the  square,  subject  to  the  restriction 
that  each  machine  be  used  only  once  by  each  operator  and  in  each  time 
period. 

These  two  examples  should  acquaint  the  reader  with  the  basic  concepts 
involved  in  a  Latin  square  design.  The  idea  of  a  square  is  evident,  of 

[4101 


13.1        LATIN    AND    GRAECO-LATIN    SQUARES  411 

course,  since,  if  m,  treatments  are  to  be  investigated,  we  need  mz  ex 
perimental  units, 

The  basic  assumption  for  a  Latin  square  design  with  one  observation 
per  experimental  unit  is  that  the  observations  may  be  represented  by 
the  linear  statistical  model 

yV/Gfc)  —  M  +  pi  +  ry  +  Tk  +  e;y(fc) ;  i  =   1,   -  •   •  ,  nt  (13 .  1) 

£   —    1      ...       w 

K    —     13  ,    77Z, 

where 

vn.  tn  m 

-A  =  0 


and  the  €#<*>  are  independently  and  normally  distributed  with  mean  0 
and  common  variance  <r2.  The  subscript  k  is  placed  in  parentheses  to 
indicate  that  it  is  not  independent  of  i  and  j.  The  constants  pt-,  y^,  and 
Tk  are,  of  course,  the  true  effects  associated  with  the  ith  row,  Jth 
column,  and  fcth  treatment,  respectively. 

Because  of  the  possible  economies  due  to  reduced  sample  sizes,  the 
Latin  square  design  has  great  appeal  to  researchers  in  all  fields.  In  par 
ticular,  the  engineer  has  been  a  prolific  user  of  the  Latin  square  design, 
but,  unfortunately,  he  has  not  always  used  the  design  "wisely.  An 
examination  of  the  postulated  statistical  model  will  show  that  the 
interactions  among  rows,  columns,  and  treatments  have  been  assumed 
to  be  0.  In  many  engineering  or  industrial  experiments  involving  a 
Latin  square  design  (where  the  rows  and  columns  usually  refer  to 
real  chemical,  physical  or  other  factors) ,  it  is  precisely  this  assumption 
that  appears  to  have  been  overlooked  by  the  researcher.  (NOTE: 
When  information  about  interactions  is  lacking  or  when  the  assumption 
of  0  interaction  is  of  doubtful  validity,  a  full  factorial  should  be  run.) 

Having  pointed  out  the  advantages  and  limitations  of  a  Latin  square 
design,  let  us  now  summarize  the  appropriate  calculations.  These  are: 

=  total  sum  of  squares 

771  7M.  fft,  7n  tn  T^fL  _ 

-2       _  y^  y^  Tr2       _  y- 

/-  -^  .{   -*   -*-  iy  (&)  s  > 


Myy  =  sum  of  squares  due  to  the  mean 

(13  .  3j 


Ryy  =  row  sum  of  squares 

2  (13.4) 


Cyy  =  column  sum  of  squares 

2  (13.5) 


412 


CHAPTER    13,    OTHER    DESIGNS 

yy  =  treatment  sum  of  squares 


and 


=  experimental  error  sum  of  squares 

*—  ~        X    y   -«^  """""      JML  yy     ~~~~     JK-yy  \-s  yy     ™~~       JL   yy 


(13.6) 


(13.7) 


where  Jfg,-,  Cj,  and  Tk  represent  the  indicated  row,  column,  and  treat 
ment  totals,  and  T  denotes  the  total  of  all  the  observations.  The  result 
ing  ANOVA  is  shown  in  Table  13.1. 

TABLE  13.1-Generalized  ANOVA  for  an  mXm  Latin  Square 
Design  With.  One  Observation  per  Experimental  Unit 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected 
Mean  Square 

F-Ratio 

Mean  

1 

MM 

M 

Rows 

m  —  I 

JR 

02  _j_  tm/(m  —  l}]  y^  P!- 

0!nlvimT»R  .    .    .    ,        -.,-,, 

m—1 

c 

i—  i 
<r*+\m/(in  —  l')]  £T/ 

Treatments  

T 

T 

y-i 
o^+Iw/Cw  —  1)]  2  £ 

T/E 

Experimental  error. 

(m—l)(m  —  2} 

E 

i—  i 

<r2 

Total 

m* 

TF* 

Example  13.3 

The  data  shown  in  Table  13.2  resulted  from  an  experiment  such  as 
described  in  Example  13.2.  Assuming  that  time  periods,  operators,  and 
machines  do  not  interact  (either  pairwise  or  as  a  complete  set),  the 
ANOVA  of  Table  13.3  is  obtained.  This  leads  to  the  conclusion  that 
there  are  significant  differences  among  the  outputs  of  the  4  machines. 
Further  examination  of  the  data  should  permit  selection  of  the  most 
productive  machine  or  machines. 

Should  a  single  observation  be  missing  in  an  experiment  conducted 
according  to  an  mXm  Latin  square  design,  its  value  may  be  estimated 
using 


M  = 


m(R 


T}  — 


(m  —  1)  (m  —  2) 


(13.8) 


13.1        LATIN   AND    GRA  ECO -LATIN    SQUARES 


413 


TABLE  13.2-Number  of  Units  Produced  by  Four  Machines 
in  a  Latin  Square  Design 

(The  random  assignment  of  the  machines  is  shown 
by  the  letters  in  parentheses) 


Time 
Periods 

Operators 

1 

2 

3 

4 

1  

31  (Q 
39  (£>) 
57  (£) 
85  {A) 

43  (Z>) 
96  (A) 
33  (C) 
46  (£) 

67  01) 
40  (J5) 
40  (£>) 
48  (C) 

36  (J5) 
48  (C) 
84  (A} 
50  (Z>) 

2. 

3 

4  

TABLE  13. 3- Abbreviated  ANOVA  for  Data  of  Table  13.2 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean 
Square 

Time  periods   ... 

3 

408.188 

136.06 

J^        2 

<r2-K4/3)  5Z  Pi 

Operators  

3 

88.688 

29.56 

*—  i 
<T2-f-(4/3)  ]T  T* 

Jv/Cacnines   

3 

4946.688 

1648.90** 

*-i 
<72-f-(4/3)  2^  rl 

Experimental  error  ..... 

6 

515.874 

85.98 

Jt-i 
o-2 

**  Significant  at  <*  =  0.01. 

where 

R  =  sum  of  observations  in  the  same  row  as  the  missing  observation 

C  =  sum  of  observations  in  the  same  column  as  the  missing  obser 
vation 

T  =  sum  of  observations  with  the  same  treatment  as  the  missing 
observation 

>§f==sum  of  all  the  actual  observations. 

After  substituting  the  value  of  M  in  the  table,  the  various  sums  of 
squares  are  calculated  as  indicated  above.  However,  it  must  be  remem 
bered  that  the  treatment  sum  of  squares  so  calculated  (Tvy)  will  be 
biased  upwards,  and  a  correction  must  be  applied  before  we  test  the 
hypothesis  H:rk  =  Q  (/b=  1,  -  •  •  ,  m).  This  correction  is  made  by  com 
puting  a  new  treatment  sum  of  squares  (T7^)  defined  as 


Tf 

JL    <UU 


Tyy 


where 


jr     


[S  —  R  —  C  —  (m  —  1)  T\* 
(m  —  1)20»  -  2)2 


(13.9) 


(13.10) 


414  CHAPTER    13f    OTHER    DESIGNS 

Remember  that  the  degrees  of  freedom  associated  with  experimental 
error  and  total  are  each  reduced  by  one  (in  view  of  the  single  missing 
observation);  that  is,  the  degrees  of  freedom  for  experimental  error 
ar&  now  (m  —  l)(m  —  2)  —  1,  and  the  degrees  of  freedom  for  total  are 
now  m2—  1.  No  example  will  be  given  for  the  above  technique,  but 
one  or  two  of  the  problems  at  the  end  of  this  chapter  will  illustrate  the 
principles  involved. 

By  now  the  reader  should  be  sufficiently  adept  at  the  calculations 
involved  in  analyses  of  variance  so  that  lengthy  discussions  of  such 
topics  as  subsampling,  selected  treatment  comparisons,  factorials, 
analysis  of  response  functions,  estimation  of  components  of  variance, 
and  predictions  of  the  relative  efficiencies  of  various  allocations  of  the 
observations  in  terms  of  experimental  and  sampling  units  would  be  a 
waste  of  time.  Accordingly,  we  will  do  no  more  than  state  that  the 
techniques  introduced  in  Chapters  11  and  12  may  easily  be  extended 
and  adapted  for  use  with  Latin  square  designs.  However,  to  make  cer 
tain  that  the  previously  mentioned  extensions  and  adaptations  are  made 
properly,  a  few  problems  requiring  their  use  have  been  included  in  the 
set  at  the  end  of  this  chapter. 

Before  terminating  our  discussion  of  the  Latin  square  design,  mention 
must  be  made  of  its  efficiency  relative  to  completely  randomized  and 
randomized  complete  block  designs.  (NOTE:  This  discussion  will,  of 
course,  be  closely  related  to  that  of  Section  12.7.)  If  we  designate  the 
mean  squares  in  the  Latin  square  for  rows,  columns,  and  experimental 
error  by  -R,  C,  and  E,  respectively,  we  may  readily  evaluate  the  effi 
ciency  of  a  Latin  square  design  relative  to  either  a  completely  ran 
domized  or  randomized  complete  block  design.  For  the  efficiency  of  a 
Latin  square  design  relative  to  a  completely  randomized  design,  we 
calculate 

R  +  C  +  (m  —  1)JE 
R.E.  =  -  -  -  —  -  (13.11) 


If,  however,  we  wish  to  compare  a  Latin  square  design  with  what 
might  have  happened  had  a  randomized  complete  block  design  been 
utilized  (assuming  the  rows  were  used  as  blocks),  the  following  formula 
is  appropriate: 

C  +  (m  —  1)JS 

R.E.  =  -  -  -  —  -  (13.12) 

mE 

If  columns  were  used  as  the  blocks,  we  put  R  in  place  of  C  in  Equation 
(13.12). 

The  concept  of  a  Latin  square  design  can  be  extended  rather  easily 
to  that  of  a  Graeco-Latin  square  (G~LS}  design.  Rather  than  go  into  the 
details  of  a  Graeco-Latin  square,  we  shall  only  indicate,  by  example, 
the  nature  of  the  design.  Those  persons  interested  in  using  such  a  de 
sign  are  advised  to  consult  a  professional  statistician. 


13.2       SPLIT    PLOTS  415 

Example  13.4 

Chew  (21)  describes  an  experiment  which  could  be  used  to  compare 
five  formulations  (<x,  /3,  y,  d,  e)  for  making  concrete  bricks,  using  material 
from  5  batches,  prepared  on  each  of  5  days,  and  tested  on  5  different 
machines  (A,  B}  C,  D,  N).  One  possible  randomization  of  a  Graeco- 
Latin  square  design  for  this  situation  is  shown  in  Table  13.4,  It  will  be 
noted  that:  (1)  each  Latin  letter  appears  exactly  once  in  each  row  and 
each  column,  (2)  each  Greek  letter  appears  exactly  once  in  each  row 
and  each  column,  and  (3)  each  Latin  letter  appears  exactly  once  with 
each  Greek  letter, 

TABLE  13.4-Symbolic  Representation  of  the  Graeco-Latin  Square  Used 

in  Example  13.4 


"O  f\w«5i 

Columns  (Days) 

(Batches) 

1 

2 

3 

4 

5 

1.  .                .  -   . 

Ace 

By 

Ce 

D8 

JES 

2    . 

BQ 

C5 

Da. 

Ey 

Ae 

3  

Cy 

-De] 

E8 

Ad 

BOL 

4  

D5 

^  JL 

Ea. 

Ay 

Be 

C8 

5.  ... 

JSe 

A3 

Bd 

COL 

Dy 

As  tempting  as  Graeco-Latin  squares  are  to  the  industrial  experi 
menter  (because  of  the  potential  savings  in  numbers  of  observations) , 
they  should  be  used  with  caution.  This  recommendation  sterns  from  the 
same  type  of  limitation  that  was  emphasized  for  Latin  square  designs, 
namely,  no  interactions  are  tolerated. 

13.2     SPLIT   PLOTS 

A  fairly  simple  design  which,  is  frequently  used  in  experimental  work 
is  the  split  plot  OSP)  design.  In  this  design  we  are  concerned  with  two 
factors,  but  we  wish  more  precise  information  on  one  of  them  than  on 
the  other.  Let  us  assume  that  we  have  factors  a  and  b  and  desire  more 
accurate  information  on  b  than  on  a.  The  usual  scheme  is  to  assign  the 
various  levels  of  factor  a  at  random  to  the  whole  plots  (main  plots)  in 
each  replicate  as  in  a  randomized  complete  block  design.  Following 
this,  the  levels  of  b  are  assigned  at  random  to  the  split  plots  (sub-plots) 
within  each  whole  plot.  Under  such  a  scheme  of  randomization,  which 
may  arise  not  only  from  the  desire  for  more  precise  information  on  one 
factor  than  on  another  but  also  because  of  the  nature  of  the  factors  and 
the  way  in  which  they  must  be  applied  to  the  experimental  units,  the 
analysis  of  variance  appears  as  in  Table  13.5. 

Example  13.5 

An  experiment  similar  to  that  described  in  Example  10.19  was  per 
formed.  However,  in  this  case,  there  were  six  replicates,  three  tem 
peratures,  and  four  levels  of  electrolyte.  (NOTE:  In  contrast  to  Ex- 


416  CHAPTER    13,    OTHER    DESIGNS 

TABLE  13.5-Generalized  ANOVA  for  a  Split  Plot  Design 


Source  of 
Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Expected  Mean 
Square 

F-  Ratio 

1 

r—  1 
a—  1 

(r-l)O-l) 

£-1 

O_  !)(&-!) 

(r_D  *Ca_D 

Myv 
Rvv 

Ayy 

CEOw 

J?VJ/ 

C^B)W 

CEOiw 

M 

R 
A 

Ea 

B 
AB 

Eb 

Whole  plots 
Replicates. 
A  

o-i+bo-z+rb  y^  ct.i/(a,—  1) 
a—  i 

2    ,          2 
CT1-J-0CT2 

<r!+rai;/sJ/(6-l) 

A—I 

°-?+r  z;  i  (<*/?)**/(«-  1)  (&-1) 

j_i  fc—  i 

2 
CTl 

^4/^a 

Whole  plot 
error.  .  .  . 
Split  plots 

B  

B/Eb 
AB/Eb 

AB 

Split  plot 
error  .... 

Total 

rob 

]£F* 

ample  10.19,  heat  paper  was  not  a  factor  in  this  experiment.)  The  data 
are  given  in  Table  13.6  and  the  resulting  ANOVA  in  Table  13.7.  No 
calculational  details  are  reported,  since  these  are  assumed  to  be  straight 
forward.  Further  interpretation  of  the  data  is  impossible  because  of  lack 
of  information  regarding  the  exact  nature  of  the  treatments. 

TABLE  13.6— Activated  Lives  (in  Hours)  of  72  Thermal  Batteries  Tested 

in  a  Split  Plot  Design  Which  Used  Temperatures  as  Whole  Plots 

and  Electrolytes  as  Split  Plots 


Replic 

ate 

Temperature 

Electrolyte 

I 

2 

3 

4 

5 

6 

Low        

A 

2.17 

1.88 

1.62 

2.34 

1.58 

1.66 

M!  edium  

B 
C 
D 

A 

1.58 
2.29 
2.23 

2.33 

1.26 
1.60 
2.01 

2.01 

1.22 
1.67 
1.82 

1.70 

1.59 
1.91 
2.10 

1.78 

1.25 
1.39 
1.66 

1.42 

0.94 
1.12 
1.10 

1.35 

High  

B 
C 
D 

A 

1.38 
1.86 
2.27 

1.75 

1.30 
1.70 
1.81 

1.95 

1.85 
1.81 
2.01 

2.13 

1.09 
1.54 
1.40 

1.78 

1.13 
1.67 
1.31 

1.31 

1.06 
0.88 
1.06 

1.30 

B 
C 
D 

1.52 
1.55 
1.56 

1.47 
1.61 
1.72 

1.80 
1.82 
1.99 

1.37 
1.56 
1.55 

1.01 
1.23 
1.51 

1.31 
1.13 
1.33 

13.3       COMPLETE    FACTORIALS    WITHOUT    REPLICATION 

TABLE   13.7-Abbreviated  ANOVA  for  Data  of  Table  13.6 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean  Square 

Whole  plots 
Replicates  

5 

4   1499 

0  .  8300 

Temperatures  .  . 

2 

0    1781 

O  0890 

\Vlaole  plot  error  

10 

1  .  3622 

0.1362 

Split  plots 
Electrolytes    .  . 

3 

1.9625 

0  6542** 

Temperature  X 
electrolyte  

6 

0   2105 

0  0351 

Split  plot  error 

45 

1    2586 

0  0280 

Significant  at  a:  =  0.01. 


Before  leaving  the  subject  of  split  plot  designs,  we  must  take  note 
tliat  the  principle  of  "splitting"  may  be  carried  on  for  several  stages; 
that  is,  we  may  employ  split-split  plot  designs,  etc.  For  more  detailed 
discussion  of  such  designs  and  for  some  illustrative  examples,  one 
should  consult  the  references  at  the  end  of  the  chapter. 


13.3      COMPLETE  FACTORIALS  WITHOUT   REPLICATION, 
FRACTIONAL   FACTORIALS,   AND    INCOMPLETE 
BLOCKS 

Because  most  experimenters  are  interested  in  investigating  the  effects 
on  a  response  variable  of  the  simultaneous  variation  of  many  factors, 
a  large  number  of  designs  incorporate  factorial  treatment  combinations. 
However,  as  the  number  of  factors  increases,  the  size  of  the  experiment 
becomes  prohibitive.  In  addition,  it  becomes  difficult  to  control  the 
magnitude  of  the  experimental  error  within  reasonable  bounds. 

In  an  attempt  to  reduce  the  experimental  error  to  a  reasonable  mag 
nitude,  the  principle  of  confounding  (see  Chapter  10)  was  utilized  to 
create  a  group  of  designs  known  as  incomplete  block  designs.  These 
designs  are  so  named  because  not  all  the  treatment  combinations  are 
present  in  each  block,  that  is,  the  blocks  are  incomplete.  With  ade 
quate  replication,  these  designs  proved  very  useful  in  agricultural 
experimentation . 

Since  incomplete  block  designs  are  not  usually  included  in  a  first 
course  in  statistical  methods,  the  decision  has  been  made  to  omit  dis 
cussion  of  them  from  this  book.  However,  several  of  the  references 
listed  at  the  end  of  the  chapter  discuss  at  length  the  methods  of 
analysis  appropriate  to  such  designs. 

When  engineers  and  physical  scientists  became  interested  in  statis 
tically  designed,  multi-factor  experiments,  they  decided  that  both 
replicated  complete  factorials  and  incomplete  block  designs  were  un- 


418 


CHAPTER    13,    OTHER    DESIGNS 


satisfactory  in  that  they  required  too  many  experimental  units. 
Further,  it  was  evident  that,  as  a  general  rule,  the  experimental 
errors  in  industrial  experiments  were  much  smaller  than  those  en 
countered  in  agricultural  experiments.  Because  of  the  small  experi 
mental  errors,  one  common  approach  has  been  to  avoid  replication 
(i.e.,  subject  only  one  experimental  unit  to  each  treatment  combina 
tion)  and  to  estimate  the  experimental  error  by  pooling  the  mean 
squares  associated  with  the  higher  order  interactions.  (NOTE:  This  is 
equivalent  to  assuming  that  the  true  high  order  interaction  effects 
are  0.)  This  technique,  referred  to  in  the  title  of  this  section  as  complete 
factorials  without  replication,  is,  as  we  have  said,  used  quite  often. 
(NOTE:  Actually,  it  is  a  completely  randomized  design  involving 
factorial  treatment  combinations  and  utilizing  only  one  experimental 
unit  per  treatment  combination.) 

Rather  than  devote  a  lot  of  space  to  the  discussion  of  the  models 
and  assumptions  for  the  many  possible  situations,  let  us  consider  an 
example.  It  is  hoped  that  this  will  prove  sufficient  for  a  reasonable 
understanding  of  the  principles  involved.  For  those  persons  who  wish 
to  consider  the  matter  more  thoroughly,  I  again  recommend  the  refer 
ences  listed  at  the  end  of  the  chapter. 

Example  13.6 

Davies   (28)    considered  a  laboratory  experiment  to  investigate  the 
yield  of  an  isatin  derivative  as  a  function  of  acid  strength   (a),  time 
of  reaction  (b),  amount  of  acid  (c),  and  temperature  of  reaction  (d).  Two 
levels  of  each  factor  were  used,  namely: 
a:  87  per  cent,  93  per  cent 
b:    15  minutes,  30  minutes 
c:    35  ml.,  45  ml. 
d:  60°C.,  70°C. 
The  data  shown  in  Table  13.8  led  to  the  ANOVA  of  Table  13.9. 

TABLE   13.8-Yield  of  Isatin  Derivative 
(g.  per  10  g.  of  base  material) 


Acid 
Strength 
(a) 

Reaction 
Time 
(*) 

Temperature  of  Reaction  (d) 

60±1 

70  ±1 

Amount  of  acid  (e) 

Amount  of  acid  (c) 

35  ml.               45  ml. 

35  ml.               45  ml. 

87 
93 

15  rain. 
30  rain. 

15  min. 
30  min. 

6.08  (1)          6.31  (e) 
6.53  (b)          6,12  (be) 

6.04  (a)          6.09  (ae) 
6.43  (ab)        6.36  (abc) 

6.79  (d)          6.77  (ed) 
6.73  (bd)        6.49  (bed) 

6.68  (ad)        6.38  (aed) 
6.08  (abd)     6.23  (abed) 

Source:  O.  31.  Davies,  (editor),  Design  and  Analysis  of  Industrial  Experiments.  Second 
Edition.  Oliver  and  Boyd,  Edinburgh,  1956,  p.  275,  Table  7.7.  By  permission  of  the  author 
and  the  publishers. 


13,3       COMPLETE    FACTORIALS    WITHOUT    REPLICATION 

TABLE   13. 9- Abbreviated  ANOVA  for  Data  of  Table  13.8 


419 


Source  of  Variation 

Degrees  of 
Freedom 

Mean  Square 

Main  effects 
A  

i 

0~f  A  /C  "2 

B  

i 

.  14oo 
Onrn  Q 

C  

i 

.UUlo 
Ono  1  1 

D  

i 

-Uzoo 

OOOOQ 

Two-factor  interactions 
AB  

i 

.  ZWo 
Or\r\r\c\ 

AC  

i 

.  uuuu 

Of\f~\A  /£ 

AD  

1 

,UU4fcO 
01  r\A  r\ 

BC  

1 

-  1U4U 
Or\-i  *7j< 

BD  

1 

.U17o 
00  co  c 

CD  

1 

.  ZDZ.S 

Of\r\OQ 

Experimental  error 
(pooled  high  order  interactions)  . 

5 

.  UU^O 

0.0385 

Source:  O.  L.  Davies,  (editor),  Design  and  Analysis  of  Industrial  Experiments,  Second 
Edition,  Oliver  and  Boyd,  Edinburgh,  1956,  p.  277,  Table  7.72.  By  permission  of  the 
author  and  the  publishers. 


As  helpful  as  it  was,  the  approach  taken  in  the  two  preceding  para 
graphs  and  illustrated  in  Example  13.6  (i.e.,  the  utilization  of  a  com 
plete  factorial  without  replication)  was  not  enough.  Experiments  were 
still  too  large  to  suit  the  researcher.  Some  other  way  had  to  be  found 
to  reduce  the  size  and  cost.  One  such  attempt  was  the  development 
of  fractional  factorials  in  which  only  some  (a  fraction)  of  the  treatment 
combinations  are  actually  investigated. 

Once  an  experimenter  has  decided  that  some  form  of  fractional  fac 
torial  is  appropriate  for  his  needs,  the  question  naturally  arises:  "Which 
treatment  combinations  should  be  included  in  the  experiment?"  The 
answer  to  this  question  depends,  of  course,  on  what  assumptions  the 
experimenter  is  willing  to  make  or,  to  phrase  it  differently,  on  what 
information  he  is  willing  to  forego.  As  we  all  know,  you  can  seldom 
get  something  for  nothing,  and  the  desired  smaller  experiment  with 
its  associated  savings  can  only  be  achieved  at  a  cost,  namely,  the 
cost  of  giving  up  part  of  the  information  usually  derived  from  complete 
factorials. 

To  illustrate  the  nature  of  a  fractional  factorial,  let  us  consider  a 
one-half  replicate  of  a  26  factorial.  If  we  have  32  experimental  units  and 
subject  them  to  the  treatment  combinations  shown  in  Table  13.10,  the 
principles  introduced  in  Section  10.14  may  be  invoked  to  show'  the 
equivalences  (i.e.,  confoundings)  of  effects  listed  in  Table  13.11.  If  the 
experimenter  is  willing  to  assume  that  all  interaction  effects  involving 
three  or  more  factors  are  0,  this  fractional  factorial  is  adequate  to 


TABLE  13.10-Treatment  Combinations  To  Be  Used  in  a  One-Half  Replicate 
of  a  26  Factorial  in  Which  the  Defining  Relation  is  I  =  ABCDEF* 


Experimental 
Unit 

Treatment 
Combination 

Experimental 
Unit 

Treatment 
Combination 

1  

(1) 

17 

/7/7 

2  

de 

18 

CLUr 

3  

ef 

19 

&& 
nJ*>-F 

?4  

yj 
df 

20 

(LQr&J 
ft-f 

15  

ab 

21 

aj 

-L.J 

*6  

abde 

22 

ud 

Z,- 

|7  

dbef 

23 

o& 

"kjj*-f 

8  

abdf 

24 

uaej 
t/ 

19  

ac 

25 

PJ 
ffj 

10  

acde 

26 

CQf 

11  

acef 

27 

ce 
ffj^-f 

12  

acdf 

28 

CCL6J 
f-f 

13  

\J,V*MJ 

be 

29 

CJ 

siT-.^J 

14  

bcde 

30 

duCCL 

15  

beef 

31 

cibcc 

16  

bcdf 

32 

dbcdef 

sif^^f 

^^/ 

Q>OCj 

-  The  use  of  the  symbol  I  rather  than  M  (as  in  Chapter  10)  is  to  agree  with  convention 
The  equality  sign  is  used  as  an  abbreviation  for  "is  completely  confounded  with." 


TABLE   13.11-Confounded  Effects  in  a  One-Half 

Replicate  of  a  26  Factorial  in  Which  the 

Defining  Relation  is  I^ABCDEP 


I=ABCDEF 
A=BCDEF 
B  =  ACDEF 
AB^CDEF 

C=ABDEF 
AC=BDEF 
BC=ADEF 
ABC=DEF 


>  =  ACEF 
ABD  =  CEF 

CD=ABEF 
ACD  =  BEF 


ABCD=EF 


E^ABCDF 


BE=ACDF 
A  BE  =  CDF 

CE=ABDF 
ACE=BDF 
BCE=ADF 
ABCE^DF 

DE=ABCF 


BDE^ACF 
ABDE=CF 

CDE=ABF 
ACDE=BF 
BCDE=AF 
ABCDE=*F 


13.4       UNEQUAL    BUT    PROPORTIONATE    SUBCLASS    NUMBERS 


421 


TABLE  13.12-Abbreviated  A1SFOVA  for  the  Experiment 

of  Table  13.10 


Source  of  Variation 


Mean 

A 

B 

C 

D 

R 

F 

AB 

AC 

AD 

AR 

AF 

BC 

BD 

BE. 

BF 

CD 

CR 

CF 

DR 

DF 

RF 

Experimental  error 

(higher  order  interactions) . 


Total 


Degrees  of 
Freedom 


1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 
1 

10 


32 


estimate  all  main  effects,  all  two-factor  interactions,  and  experimental 
error.  Under  such  an  assumption,  the  appropriate  ANOVA  is  as  given 
in  Table  13.12. 

Fractional  factorials  have  wide  application  in  industrial  experimen 
tation.  Thus,  it  will  pay  research  workers  in  both  engineering  and  the 
physical  sciences  to  become  better  acquainted  with  these  valuable  aids 
to  efficient  experimentation.  As  with  other  topics  mentioned  in  this 
section,  it  is  felt  that  a  detailed  discussion  is  beyond  the  scope  of  this 
book.  For  this  reason,  the  interested  reader  is  referred  to  the  publica 
tions  listed  at  the  end  of  the  chapter. 

13.4      UNEQUAL  BUT  PROPORTIONATE  SUBCLASS 
NUMBERS 

The  reader  will  have  noticed  that  practically  all  the  recommended 
statistical  designs  require  a  balanced  configuration,  that  is,  an  equal 
number  of  observations  in  each  group.  The  one  exception  was  the 
completely  randomized  design.  However,  even  in  that  case,  it  was 


422 


CHAPTER    13,    OTHER    DESIGNS 


noticed  that  unequal  numbers  of  observations  in  the  subgroups  could 
lead  to  difficulties  in  interpretation.  (See  Section  11.4.) 

In  this  section  we  propose  to  examine  one  other  case  of  unequal  fre 
quencies  which  presents  little  difficulty  in  the  way  of  calculation.  This 
case  involves  a  factorial  set  of  treatment  combinations  in  which  the 
cells  of,  say,  the  aX&  table  contain  different  numbers  of  observations 
but  these  numbers  happen  to  be  proportional.  That  is,  the  number  of  ob 
servations  in  the  (ij)th  cell  are  such  that  n^^u^Vj  where  u\\ u*  *  -  •  :ua 
are  the  proportions  in  the  rows  and  #1:^2-  -  •  •  :v&  are  the  proportions 
in  the  columns.  Rather  than  go  into  details,  a  numerical  example  will 
be  given  and  it  is  hoped  that  this  will  be  sufficient  to  illustrate  the 
ideas  involved.  Persons  desiring  further  details  should  consult  the 
references  at  the  end  of  the  chapter. 

Example  13.7 

Suppose  we  have  3  varieties  of  oats  to  be  tested  for  yield  differences  and 
that  we  also  wish  to  investigate  the  effects  of  3  fertilizers.  There  are  28 
experimental  plots  available  to  the  researcher.  Further,  we  will  assume 
that  from  previous  experiments  we  already  know  considerably  more 
about  varieties  B  and  C  than  about  variety  A;  thus,  we  shall  plant 
variety  A  on  twice  as  many  plots  as  varieties  B  and  C.  It  is  also  con 
sidered  desirable  to  assign  the  3  fertilizers  to  the  plots  in  the  ratio 
3:2:2;  that  is,  we  shall  apply  fertilizer  No.  1  to  12  plots  and  each  of 
fertilizers  No.  2  and  No.  3  to  8  plots.  The  assignment  of  the  treat 
ment  combinations  to  the  plots  was  made  completely  at  random, 
and  the  resulting  yields,  in  bushels  per  acre,  are  recorded  in  Table  13.13. 
Calculating  the  various  sums  of  squares  in  the  usual  manner,  we  arrive 
at  the  abbreviated  ANOVA  of  Table  13.14. 

TABLE   13. 13- Yields  of  3  Varieties  of  Oats  Subjected  to  3  Different 

Fertilizer  Treatments 
(In  bushels  per  acre) 


c\<*  + 

Fertilizer 

vjat 
Variety 

1 

2 

3 

A    

50,  51,  52,  56,  60   55 

42,  40   38   38 

55   56   56 

58 

B  

65,  69,  67 

50,  50 

62,  62 

C 

67,  67,  69 

48   50 

65,  67 

TABLE   13. 14- Abbreviated  ANOVA  for  Data  of  Table  13.13 


Source  of  Variation 

Degrees  of 
Freedom 

Sum  of 
Squares 

Mean 
Square 

Treatments 
Varieties  

2 

818.9 

409   45 

Fertilizers    . 

2 

1455   0 

727    50 

Varieties  X  fertilizers  

4 

52.3 

13   075 

Among  plots  treated  alike 

19 

100   5 

5   289 

13.5        UNEQUAL    AND    DISPROPORTIONATE    SUBCLASS    NUMBERS  423 

13.5      UNEQUAL  AND    DISPROPORTIONATE   SUBCLASS 
NUMBERS 

Let  us  now  examine  the  case  where  our  data  may  be  represented  by 
tlie  model 

Y*Sh  =  /*  +  «<  +  fo  +  («£)*  +  €„•*  i  =   1,   •   •   -  ,  a          (13  .  13) 

J  =   1,  •   •   '  ,  6 


where  the  various  terras  are  defined  as  before  but  the  n^  are  not  equal 
for  the  various  cells  of  the  aX&  table.  Further,  the  n^  are  not  propor 
tionate  as  they  were  in  Section  13.4.  What  difficulties  in  analysis  result 
from  this  fact?  Why  is  it  that  we  refer  to  the  case  of  "unequal  and  dis 
proportionate  subclass  numbers"  as  an  undesirable  situation?  The 
answer  is,  of  course,  because  we  encounter  complications  in  analyzing 
such  data.  Let  us  now  take  note  of  some  of  the  problems  that  arise. 

Suppose  that  we  went  ahead,  ignoring  the  fact  that  the  subclass 
numbers  are  disproportionate,  and  calculated  the  various  sums  of 
squares  in  the  usual  fashion.  If  this  procedure  were  followed,  we 
would  find  that  the  sums  of  squares  so  calculated  (assuming  that  each 
sum  of  squares  was  calculated  directly;  that  is,  no  sum  of  squares  was 
obtained  by  subtraction)  would  not  sum  up  to  agree  with  the  total 
sum  of  squares.  In  other  words,  because  of  the  disproportionality  of  the 
subclass  numbers,  the  different  comparisons  with  which  the  sums  of 
squares  are  associated  are  -nonorthogonal.  This,  of  course,  would  lead  to 
biased  test  procedures  unless  some  adjustment  were  made.  The  other 
major  difficulty  which  arises  when  dealing  with  cases  involving  dis 
proportionate  subclass  numbers  is  that  the  simple  (unweighted)  treat 
ment  means  obtained  from  the  data  are  biased  estimates  of  the  true 
treatment  effects.  This  could  lead  to  serious  errors  if  inferences  were 
made  without  attempting  to  correct  for  the  above-mentioned  bias. 

What  then,  should  be  the  method  of  analysis  for  such  situations? 
The  usual  approach  is  to  utilize  regression  techniques  and  obtain  a 
general  least  squares  solution.  However,  because  of  the  many  varia 
tions  which  may  be  employed  (e.g.,  different  models  and/or  different 
orders  of  estimating  the  unknown  parameters),  neither  detailed  ex 
planations  nor  numerical  illustrations  of  such  solutions  will  be  included 
in  this  book.  If  you  should  encounter  a  situation  in  which  a  general 
least  squares  solution  is  required,  I  would  suggest  that  you  do  three 
things:  (1)  review  the  contents  of  Chapter  8;  (2)  study  the  appropriate 
sections  of  some  of  the  references  at  the  end  of  this  chapter;  and  (3) 
consult  a  professional  statistician. 


424  CHAPTER    13,    OTHER    DESIGNS 

13.6      RESPONSE  SURFACE  TECHNIQUES 

One  of  tlie  most  significant  contributions  to  statistical  methodology 
in  recent  years  has  been  the  development  of  systematic  procedures 
for  determining,  experimentally,  those  levels  of  the  factors  under  inves 
tigation  which  produce  an  optimum  response.  These  procedures,  fre 
quently  referred  to  as  response  surface  techniques,  can  be  of  value  to 
researchers  in  almost  every  field  of  specialization.  Unfortunately,  a 
satisfactory  description  of  the  many  ramifications  of  these  techniques 
is  more  than  can  be  accomplished  in  this  text.  Therefore,  we  shall  be 
content  with  a  few  general  observations  on  the  topic  and  then  refer  the 
reader  to  other  sources  where  these  ideas  are  discussed  in  greater  detail. 

Response  surface  techniques  are,  in  essence,  a  blending  of  regression 
analysis  (Chapters)  and  experimental  design  (Chapters  10,  11,  12,  and 
13)  to  provide  an  economical  means  of  locating  a  set  of  experimental 
conditions  (i.e.,  a  combination  of  factor  levels)  which  will  yield  a 
maximum  (or  minimum)  response.  However,  one  very  important  fea 
ture  has  been  added.  That  feature  is  the  sequential  nature  of  the  explora 
tion  of  the  response  surface.  While  it  is  true  that  most  research  is  of  the 
continuing  variety  (and  therefore  sequential),  the  majority  of  the 
techniques  discussed  heretofore  in  this  book  have  been  of  the  nonse 
quential  type.  Thus,  the  insertion  of  the  sequential  element  into  the 
pattern  of  the  investigation  is,  from  one  point  of  view,  a  long  overdue 
step. 

In  capsule  form,  the  steps  involved  in  the  application  of  response 
surface  techniques  are  as  follows : 

(1)  Choose  base  levels  of  the  factors  to  be  investigated.  (Depend 
ing  on  the  judgment  of  the  experimenter,  these  levels  may  be 
close  to  or  far  removed  from  the  optimum  levels.) 

(2)  Since,  at  this  stage,  linear  effects  are  thought  to  be  dominant 
over  nonlinear  effects,  select  one  other  level  of  each  factor. 

(3)  Utilizing  either  a  complete  or  fractional  2n  factorial,  estimate 
(by  examining  the  effects;  i.e.,  the  linear  regression  coefficients) 
the  direction  in  which  the  greatest  gain  may  be  expected. 

(4)  Moving  in  this  direction,  that  is,  along  the  path  of  steepest 
ascent,  to  the  extent  that  the  experimenter  deems  reasonable, 
a  second  experiment  (again  utilizing  a  complete  or  fractional 
2n  factorial)  is  performed. 

(5)  Repeat   steps   (3)   and   (4)   until  a  near-stationary  region  is 
found. 

(6)  Then,  utilizing  a  complete  or  fractional  3n  factorial  or  a  com 
posite  design1  to  estimate  the  second  order  effects,  the  nature 
of  the  response  surface  may  be  explored  in  the  near-stationary 
region  and  the  optimum  conditions  located. 

It  will  be  realized  that  the  preceding  steps  are  only  an  indication  of 

1  A  composite  design  is  essentially  a  complete  2n  factorial  with  sufficient  points 
added  to  permit  estimation  of  the  second  order  effects. 


13.8       OTHER   DESIGNS    AND   TECHNIQUES  425 

the  procedure.  Depending  on  the  problem  and  the  assumptions  that 
the  experimenter  is  willing  to  make,  the  "rules"  may  be  modified.  (For 
example,  3n  factorials  or  composite  designs  might  be  used  in  step  No.  3.) 
However,  regardless  of  the  details,  the  philosophy  of  sequential  experi 
mentation  has  much  to  recommend  it.  In  fact,  the  concept  of  exploring 
a  response  surface  in  a  sequential  manner  with  the  objective  of  locating 
a  maximum  (or  minimum)  point  on  the  surface  is  one  with  which  all 
experimenters  should  be  familiar.  For  those  who  are  interested  in 
pursuing  this  topic  further,  an  excellent  exposition  is  available  in 
Davies  (28).  Both  the  theory  and  the  application  of  response  surface 
techniques  are  also  discussed  in  a  number  of  the  other  references  listed 
at  the  end  of  this  chapter. 

13.7  RANDOM    BALANCE 

Another  recent  contribution  to  experimental  design  is  the  concept  of 
random  balance  investigations  in  multi-multi-factor  experiments. 
As  proposed  by  Satterthwaite  (43),  the  random  balance  technique 
permits  the  researcher  to  screen  a  large  number  of  possible  contributing 
factors  in  an  experiment  involving  a  limited  number  of  test  runs.  That 
is,  random  balance  is  a  device  for  considering  (simultaneously)  the  many 
factors  involved  and,  as  is  always  important,  keeping  the  size  of  the 
experiment  within  reasonable  bounds.  When  the  experiment  has  been 
performed,  examination  of  the  results  should  permit  isolation  of  the 
more  important  factors  for  further  investigation. 

In  a  random  balance  experiment,  all  factors  and  levels  are  consid 
ered  by  choosing  at  random  the  level  of  each  factor  to  be  used  in  forming 
a  particular  treatment  combination.  (NOTE:  From  a  practical  point  of 
view,  the  following  restriction  on  complete  randomization  has  been 
found  desirable:  Each  level  of  a  particular  factor  should  be  used  an 
equal,  or  nearly  equal,  number  of  times.)  Since  random  balance  experi 
mentation  was  first  proposed,  there  has  been  much  discussion,  both  pro 
and  con,  as  to  its  worth.  Personally,  I  believe  that  random  balance  has 
much  to  recommend  it  and  that  we  will  see  a  rapid  increase  in  its  use, 
especially  in  industrial  experimentation.  However,  the  theory  on  which 
it  is  based  has  not  been  fully  explored,  and  thus  the  controversy  over 
its  merits  continues.  For  those  interested  in  the  possibilities  and/or 
wisdom  of  using  random  balance  in  their  own  experimentation,  I  sug 
gest  a  careful  reading  of  the  appropriate  references  at  the  end  of  this 
chapter. 

13.8  OTHER   DESIGNS  AND  TECHNIQUES 

As  was  stated  in  the  opening  paragraph  of  this  chapter,  the  number 
of  designs  and  analysis  techniques  that  have  been  developed  for  special 
purposes  are  many.  Thus,  it  has  been  possible  to  mention  only  a  few 
in  this  book.  The  two  most  common  designs,  the  completely  random 
ized  design  and  the  randomized  complete  block  design,  were  discussed 
in  detail  in  Chapters  11  and  12,  respectively.  In  this  chapter  a  few  of 


426 


CHAPTER    13,    OTHER    DESIGNS 


the  more  specialized  designs  and  techniques  have  been  described  or 
alluded  to.  An  examination  of  some  of  the  references  which  follow  will 
bring  many  other  special  designs  to  your  attention.  It  is  my  hope  that 
the  presentation  thus  far,  brief  through  it  has  sometimes  been,  will 
have  whetted  your  appetite  and  that  you  will  continue  your  readings 
and  studies  in  the  related  areas  of  experimental  design  and  research 
techniques. 

Problems 

13.1        A  5X5  Latin  square  was  laid  out  to  test  the  effects  of  5  fertilizers  on 
the  yield  of  potatoes.  Perform  a  complete  analysis  of  the  data. 


Column 

"O  i-v-vjr-r 

Row 

1 

2 

3 

4 

5 

Totals 

1  

A  449 

B  444 

C  401 

Z>299 

£292 

1885 

2.. 

B  463 

C  375 

Z>323 

£264 

A  415 

1840 

3  

C  393 

Z?353 

£278 

A  404 

^425 

1853 

4  

D371 

E  241 

A  441 

B  410 

C  392 

1855 

5   . 

£  258 

A  430 

B  450 

C  385 

jD347 

1870 

Column 
totals 

1934 

1843 

1893 

1762 

1871 

9303 

Treatment  totals 


A:  2139 

B:  2192 

C:  1946 

D:  1693 

E:  1333 

13.2        Shown  below  are  the  yields  (cwt.  per  1/40-acre  plots)  of  sugar  cane 
in  a  Latin  square  experiment  comparing  fertilizers. 


A  14 
B  19 
D23 
C  21 

£23 


E22 
D21 
A  15 
#46 
C  16 


£20 
A  16 
C20 
£24 


C  18 
£23 
B  18 
£>21 
A  17 


D25 
C  18 
£23 
A  18 
B  19 


A :  No  fertilizer 

B :  Complete  inorganic  fertilizer 
C:   10  tons  manure  per  acre 
£>:  20  tons  manure  per  acre 
£:  30  tons  manure  per  acre 

What  conclusions  do  you  draw  from  this  experiment? 

13.3        Analyze  the  following  data  from  a  cacao  experiment  consisting  of  3 
separately  randomized  Latin  squares.  The  3  treatments  were: 

A :  No  fertilizer  (check) 

B:    1.5  Ibs.  superphosphate  per  tree 

C:   3  Ibs.  superphosphate  per  tree 


PROBLEMS 


427 


The  field  plans  of  the  squares,  together  with  plot  yields  in  average 
pods  per  tree,  are  as  follows: 


B 

C 

A 

41 

25 

15 

A 

B 

C 

20 

32 

24 

C 

^ 

^ 

22 

12 

21 

C 

B 

.4 

27 

28 

3 

^ 

C 

B 

4 

17 

9 

B 

A 

C 

22 

4 

17 

A 

C 

B 

11 

15 

17 

B 

A 

C 

24 

14 

33 

C 

B 

A 

22 

20 

15 

Note:  Do  not  consider  a  transformation  because  these  are  averages 
(rounded)  from  the  trees  on  1/15-acre  plots.  The  total  numbers  of 
pods  were  large  enough  to  approximate  a  continuous  distribution. 
13.4         Five  levels  of  a  fertilizer  were  tried  in  a  5X5  Latin  square.  This  is 
the  analysis: 

H> agrees  of  Mean 

Freedom  Square 

Rows  4  25 

Columns  4  20 

Treatments  4  28 

Error  12  15 

The  sums  of  the  yields  in  the  5  plots  of  each  level  were: 
Level  12345 

Sum  of  yields  2  14  26  3O  28 

Subdivide  the  4  degrees  of  freedom  for  treatments  into 

d.f. 

Linear  regression  1 

Second  degree  term  1 

Remainder  2 

Is  any  comparison  significant? 


428  CHAPTER    13r    OTHER    DESIGNS 

13.5  Crop — wheat;  Location — R.  W.  Gt.  Harpenden  (175);  Year — 1935; 
Type — 6X6  Latin  square;  Comment — yield  in  pounds  of  grain  per 
1/40-acre  plot 


Total 

4 

0 

2 

1 

3 

5 

77.2 

88.0 

89.7 

92.6 

72.1 

76.2 

495.8 

3 

4 

0 

5 

1 

2 

93.2 

95.8 

94.1 

93.9 

91.6 

67.3 

535.9 

5 

2 

3 

4 

0 

1 

90.2 

87.0 

86.1 

85.5 

93.4 

68.5 

510.7 

2 

3 

1 

0 

5 

4 

72.5 

76.7 

96.3 

95.3 

95.9 

78.2 

514.9 

0 

1 

5 

2 

4 

3 

84.2 

96.5 

98.5 

81.6 

90.1 

81.8 

532.7 

1 

5 

4 

3 

2 

0 

77.0 

91.9 

95.1 

86.3 

82.8 

60.5 

493.6 

Total     494.3 

535.9 

559.8 

535.2 

525.9 

432.5 

3083.6 

Treatment 
Treatments  Total 

0 — No  (NH02SO4 515  .5 

1 — (NH4)2SO4  applied  Oct.  26  at  0.4  cwt.  of  N/A 522.5 

2 — (NH4)2SO4  applied  Jan.  19  at  0.4  cwt.  of  N/A 480.9 

3 — (NHLOaSCX  applied  Mar.  18  at  0.4  cwt.  of  N/A 496.2 

4 — (NELOaSCU  applied  Apr.  27  at  0.4  cwt.  of  N/A 521 .9 

5 — (NHOaSO*  applied  May  24  at  0.4  cwt.  of  N/A 546.6 

Analyze  and  interpret  the  above  data. 

13.6  We  wish  to  conduct  a  field  experiment  to  test  the  yielding  ability  of 
6  varieties  of  soybeans  and  have  available  an  area  of  land  sufficient 
for  36  plots.  Indicate  the  proper  subdivision  of  the  total  degrees  of 
freedom  for  the  following  experimental  designs: 

(a)    completely  randomized 

(6)    randomized  complete  block 

(c)    Latin  square. 

Indicate,  by  means  of  arrows,  the  proper  ^P-tests  for  testing  variety 
differences  in  each  design. 

13.7  Given  that  the  data  shown  below  resulted  from  an  experiment  such 
as  described  in  Example  13.4,  perform  the  analysis  and  give  your 
interpretations  of  the  results. 


PROBLEMS 

CRUSHING  STRENGTHS  (CODED  VALUES)* 


429 


Columns  (Days) 

Rows 
(Batches) 

1 

2 

3 

4 

5 

1 

257 

230 

279 

287 

202 

2 

245 

283 

245 

280 

260 

3 

182 

252 

280 

246 

250 

4 

203 

204 

227 

193 

259 

5 

231 

271 

266 

334 

338 

*  The  treatments  are  assumed  to  have  been  Imposed  exactly  as  shown  In  Example  13.4. 

13.8  An  experiment  was  conducted  to  assess  the  relative  resistances  to 
abrasion  of  four  grades  of  leather  (A,  B,  C,  £>).  A  machine  was  used 
in  which  the  samples  could  be  tested  in  any  one  of  four  positions. 
Since  different  runs  (replications)  are  known  to  yield  variable  results, 
it  was  decided  to  make  four  runs.  A  Latin  square  design  was  utilized 
and  the  following  results  obtained.  Analyze  and  interpret  the  data. 


Position 


Run 

1 

2 

3 

4 

1 
2 
3 
4 

118(5) 
127  (D) 
174(4) 
130(C) 

136(Z>) 
141(3) 
173(C) 
170(4) 

168(4) 
129(C) 
126CB) 
125(Z>) 

135(C) 
151(4) 
134(1?) 
95(5) 

13.9  Another  experiment  such  as  described  in  Problem  13.8  was  con 
ducted  at  a  second  laboratory.  In  this  case,  the  data  shown  below 
were  obtained.  Analyze  and  interpret.  (NOTE:  M  represents  a 
missing  observation.) 


Run 

4 

2 

1 

3 

2 
3 
1 
4 


-4(150) 
2?(130) 


£(98) 


Position 


£(145) 
C(172) 
Z?(132) 
-4(171) 


-4(170) 
£(115) 
C(132) 


C(133) 
£(127) 
.4(170) 
Z?(120) 


13  10  The  experiment  described  in  Problem  13.8  was  conducted  once  more, 
this  time  at  a  third  laboratory.  Analyze  and  interpret  the  data  which 
follow.  [HINT:  Use  Equation  (13.8)  and  the  iterative  technique  dis 
cussed  in  Section  12.13.] 


430 


CHAPTER    13,    OTHER 


Run 

Position 

1 

2 

3 

4 

1 
2 
3 
4 

C(131) 
£>(139) 
B(157) 
,4(185) 

.D(Af') 
-4(196) 
C(133) 
£(146) 

.4(167) 
£(140) 
Z>(140) 
C(M") 

5(136) 
C(148) 
,4(184) 
D(150) 

13.11  On  checking  the  original  data  sheets,  it  was  discovered  that  the 
technician  took  two  independent  abrasion  readings  on  the  samples 
tested  in  the  experiment  described  in  Problem  13.8.  The  second  set 
of  readings  is  reproduced  below.  Pooling  these  data  with  those  given 
in  Problem  13.8,  analyze  and  interpret  the  complete  results. 


Run 

Position 

1 

2 

3 

4 

1 

2 
3 
4 

120(5) 
125(Z>) 
175C4) 
132(C) 

130(Z>) 
142(5) 
180(C) 
17004) 

165  (-4) 
120(C) 
120(5) 
130(1?) 

140(C) 
14004) 
140(1)) 
102(5) 

13.12  An  experiment  was  performed  to  compare  the  effects  of  three 
catalysts  on  the  yield  of  a  chemical  process.  Three  runs  were  started, 
one  using  catalyst  A,  another  using  B,  and  the  third  C.  After  3  days, 
a  sample  was  drawn  from  each  run  and  an  analysis  performed.  A 
similar  operation  (i.e.,  taking  samples  and  performing  the  analyses) 
was  performed  after  5  days.  The  whole  experiment  was  repeated  four 
times.  Analyze  and  interpret  the  resulting  data. 

CODED  YIELDS  OF  AN  UNSPECIFIED  CHEMICAL  PROCESS 


Catalyst 


A 

B 

C 

Replicate 

3  days 

5  days 

3  days 

5  days 

3  days 

5  days 

1 

68 

82 

90 

96 

82 

88 

2 

83 

79 

68 

80 

71 

78 

3 

66 

75 

70 

91 

68 

78 

4 

66 

76 

84 

92 

74 

80 

13.13  A  split-split  plot  design  was  used  in  an  experiment  concerned  with  the 
yield  of  cotton.  Four  replications  (or  blocks)  were  involved.  Each 
main  plot  was  subjected  to  one  of  two  levels  of  irrigation,  each  sub 
plot  was  subjected  to  one  of  three  rates  of  planting,  and  each  sub- 


PROBLEMS 


431 


subplot  was  subjected  to  one  of  three  levels  of  fertilizer  application. 
Analyze  and  interpret  the  following  experimental  yields. 

CODED  YIELDS  FROM  COTTON  GROWING  EXPERIMENT 


Rate  of 

Block 

Planting 

(Density  of 

T'VrtiliT'^T 

Irrigation 

Plants) 

JL    t*.L  LiiljoCJ. 

Rate 

1 

2 

3 

4 

Light 

Thin 

ISTone 

9.0 

8.2 

8.5 

8.2 

Average 

9.5 

8.1 

8.8 

7.9 

Heavy 

10.6 

9.4 

8.8 

8.6 

Medium 

None 

9.0 

9.7 

11.1 

7.8 

Average 

8.9 

9.7 

10.3 

8.5 

Heavy 

9.3 

10.4 

9.1 

8.6 

Thick 

None 

8.1 

7.4 

8.2 

8.5 

Average 

9.0 

8.1 

7.6 

8.8 

Heavy 

9.6 

7.5 

9.4 

8.4 

Heavy 

Thin 

None 

8.1 

10.3 

6.0 

7.2 

Average 

8.6 

10.8 

10.4 

11.6 

Heavy 

10.2 

10.4 

11.5 

11.6 

Medium 

None 

12.2 

9.8 

9.1 

11.0 

Average 

11.0 

9.5 

11.7 

13.2 

Heavy 

12.0 

12.4 

11.6 

^3.0 

Thick 

None 

7.9 

13.4 

12.0 

11.7 

Average 

10.0 

14.2 

12.2 

13.8 

Heavy 

12.5 

14.0 

13.8 

13.4 

13.14     The  following  data  resulted  from  an  unreplicated  complete  factorial. 
Analyze  and  interpret.  State  all  your  assumptions. 

CODED  YIELDS  OP  A  CHEMICAL  PROCESS 
Concentration  of  Solvent 


JL  dUJJCrfcLLUiC 

Low 

Medium 

High 

100 

44 

46 

42 

200 

51 

55 

55 

300 

50 

50 

48 

13.15 


In  a  manufacturing  company,  the  micrometers  used  in  checking 
quality  are  themselves  checked  by  use  of  gauge  blocks.  However, 
there  are  5  departments  and  each  has  its  own  micrometers  and 


432 


CHAPTER    13,   OTHER   DESIGNS 


gauge  blocks.  Because  of  a  suspicion  that  there  is  too  much  variation 
among  micrometers  and/or  gauge  blocks,  the  quality  control  engi 
neer  ran  a  test  utilizing  a  random  sample  of  instruments.  Analyze 
and  interpret  the  following  data. 


Gauge 
Block 

Micrometer 

1 

2 

3 

4 

5 

A 

0.0110 

0.0115 

0.0130 

0.0151 

0.0121 

B 

0.0135 

0.0127 

0.0132 

0.0155 

0.0128 

C 

0.0127 

0.0124 

0.0132 

0.0152 

0.0130 

13.16  An  experiment  was  run  to  investigate  the  effect  of  temperature,  type 
of  powder,  amount  of  powder,  and  packing  pressure  on  the  function 
time  of  an  explosive  actuator.  An  unreplicated  complete  factorial 
yielded  the  following  data.  Analyze  and  interpret. 

FUNCTION  TIME  (IN  MILLISECONDS) 


Type  of  Powder 

A 

B 

C 

Amount  of  Powder  (Mg.) 

5         10       15 

5        10       15 

5        10       15 

Temperature 
(°F.) 

Packing 
Pressure 
(psi) 

-50 

10,000 
15,000 
20,000 

7.4     7.0     6.8 
7.5     7.2     6.7 
7.4     7.4     6.0 

5.4     5.0     4.8 
5.5     5.2     4.7 
5.4     5.4     4.0 

7.2     6.9     6.6 
7.2     6.6     6.5 
7.2     6.7     6.2 

75 

10,000 
15,000 
20,000 

6.6     6.6     5.8 
6.8     6.6     6.6 
6.8     6.2     5.9 

4.6     4.6     3.8 
4.8     4.6     4.6 
4.8     4.2     3.9 

6.8     7.2     4.9 
6.9     7.0     5.0 
7.0     7.1     5.0 

200 

10,000 
15,000 
20,000 

5.1     5.1     5.1 
5.1     4.8     4.9 
5.2     4.7     5.0 

3.1     3.1     3.1 
3.1     2.8     2.9 
3.2     2.7     3.0 

6.0     4.9     4.8 
6.4     4.8     4.1 
5.9     4.9     2.0 

13.17  A  complete  but  unreplicated  factorial  was  used  to  investigate  the 
effects  of  type  of  metal  (a  qualitative  factor),  amount  of  primary 
initiator  (a  quantitative  factor),  and  packing  pressure  (a  quanti 
tative  factor)  on  the  firing  time  of  explosive  switches.  Analyze  and 
interpret  the  following  data: 


PROBLEMS 


433 


FIRING  TIMES  (IN  MILLISECONDS) 


Metal 

Primary 
Initiator 
(Mg.) 

Packing  Pressure  (psi) 

12,000 

20,000 

28,000 

2  s  al 

5 

12.3 

10.6 

15.2 

10 

10.4 

9.5 

15.0 

15 

8.8 

9.1 

14.5 

teflon 

5 

12,4 

11.7 

15.0 

10 

11.0 

11.0 

14,6 

15 

11.0 

9.8 

14.6 

13.18  A  certain  type  of  capacitor  was  to  be  tested  to  assess  its  perform 
ance  as  a  function  of  a  number  of  specified  factors.  The  four  factors 
considered  were: 

(a)  ^potted  (+)  or  not  potted  (  — ) 
(6)==  wedged  (+)  or  not  wedged  (—) 

(c)  —  impregnated  (+)  or  not  impregnated  (  — ) 

(d)  =  high  temperature  (+)  or  low  temperature  (— ). 

The  performance  characteristic  measured  was  the  high  voltage 
breakdown  when  the  capacitors  were  subjected  to  a  voltage  rise  of 
250v/sec.  Some  hypothetical  data  which  could  have  resulted  from 
such  an  experiment  are: 


Capacitor 

Level  of  Factor 

High  Voltage 
Breakdown  (kv) 

abed 

1 

—                    __ 

10.7 

2 

+                    —                    —                    + 

11.4 

3 

—                    +                    —                    + 

12.2 

4 

-j-                   4_                    —                    _ 

13.0 

5 

—                    —                    +                   + 

10.6 

6 

-}-                  —                  .4-                  — 

12.1 

7 

—                  4-                 +                 — 

12.0 

8 

+                 +                 +                 + 

13.2 

Analyze  and  interpret  the  one-half  replicate  of  a  2*  factorial  described 
above. 

13.19  An  experiment  was  to  be  performed  to  assess  the  effects  of  the 
following  factors  on  the  surge  voltage  of  a  specific  model  of  thermal 
battery:  temperature,  humidity,  amount  of  electrolyte,  amount  of 
heat  paper,  and  type  of  electrolyte.  These  five  factors  were  denoted 
as  a,  by  c,  d,  and  e,  respectively.  Since  this  was  only  a  preliminary 
experiment  (in  the  development  phase)  and  since  all  three-,  four-, 
and  five-factor  interactions  could  be  assumed  to  be  negligible,  a 


434 


CHAPTER    13,    OTHER    DESIGNS 


one-half  replicate  of  the  25  factorial  was  performed.  Analyze  and 
interpret  the  following  data. 


Treatment  Combination 

Surge  Voltage  (volts) 

(i) 

14.0 

ae 

14.6 

be 

11.7 

ab 

16.3 

ce 

11.2 

ac 

16.6 

be 

15.6 

abce 

10.2 

de 

13.9 

ad 

13.8 

bd 

15.1 

abde 

13.2 

cd 

14.6 

acde 

14.3 

bcde 

12.6 

abed 

15.4 

References  and  Further  Reading 

1.  Anderson,    H.    E.    Random   permutations   of   selected   powers   of   2   and   3. 
Sandia   Corporation  Monograph  SCR-4%8,   Sandia   Corp.,   Albuquerque,   N. 
Mex.,  Aug.,  1961. 

2.  Anderson,  R.  L.  Complete  factorials,  fractional  factorials,  and  confounding. 
Experimental    Designs   in    Industry.    (Editor:    V.    Chew).    John    Wiley    and 
Sous,  Inc.,  New  York,  1958. 

3    -  ^  and   Bancroft,   T.   A.  Statistical   Theory  in  Research.   McGraw-Hill 
Book  Company,  Inc.,  New  York,  1952. 

4.  Anscombe,  F.  J.  Quick  analysis  methods  for  random  balance  experimenta 
tion.  Technometrics,  1  (No.  2):195-209,  May,  1959. 

5.  Bicking,  C.  A,  Experiences  and  needs  for  design  in  ordnance  experimenta 
tion.  Experimental  Designs  in  Industry.  (Editor:  V.  Chew).  John  Wiley  and 
Sons,  Inc.,  New  York,  1958. 

6.  Box,  G.  E,  P.  Multi-factor  designs  of  first  order.  Biometrika,  39:49,  1952. 

7.  -  .  The  exploration  and  exploitation  of  response  surfaces;  some  general 
considerations  and  examples.  Biometrics,  10:16,  1954. 

8.  -  ,  and  Coutie,  G.  A.  Application  of  digital  computers  in  the  explora 
tion  of  functional  relationships.  Proceedings  of  the  Institution  of  Electrical 
Engineers,  Vol.  103,  Part  B,  SuppL  No.  1,  1956. 

,  and  Hunter,  J.  S.  A  confidence  region  for  the  solution  of  a  set  of 


simultaneous   equations   with  an  application  to   experimental  design. 
metrika,  41:190,  1954. 
10.  -  ,  and  -  .  Multifactor  designs.  Report  prepared  under  Office  of 


11. 


12. 


Ordnance  Contract  No.  DA-36-034-ORD-1177  (RD),  1954. 


and 


Ann.  Math.  Stat.,  27,  1956. 


-.  Multifactor    designs    for    exploring    response    surfaces. 


and 


-.  Experimental  designs  for  exploring  response  surfaces. 


Experimental  Designs  in  Industry.  (Editor:  V.  Chew).  John  Wiley  and  Sons, 
Inc.,  New  York,  1958. 


REFERENCES   AND    FURTHER    READING  435 

13.   ,  and   Wilson,    K.   B.   On   the   experimental  attainment  of  optimum 

conditions.  Jour.  Roy.  Stat.  Soc.,  Series  B,  13:1,  1951. 

14.   ,  and  Youle,  P.  V.  The  exploration  and  exploitation  of  response  sur 
faces;  an  example  of  the  link  between  the  fitted  surface  and  the  basic  mecha 
nism  of  the  system.  Biometrics,  11:287,  1955. 

15.  Brownlee,    K.   A.    Statistical  Theory  and  Methodology  in  Science  and  Engi 
neering.  John  Wiley  and  Sons,  Inc.,  New  York,  I960. 

16.  Budne,  T.  A.  Random  balance:  Part  I — The  missing  statistical  link  in  fact 
finding   techniques,    Part  II — The  techniques   of  analysis,    Part   III — Case 
histories.   Industrial  Quality  Control,   15   (No's.   10-1 1-12)  :5~10,    11-16,    16- 
19,  April,  May,  and  June  1959. 

37.  .  The    application    of    random    balance     designs,     Technometrics,     1 

(No.  2):139-55,  May,  1959. 

18.  Carroll,  M.  B.,  and  Dykstra,  O.,  Jr.  Application  of  fractional  factorials  in  a 
food    research    laboratory.    Experimental   Designs   in   Industry.    (Editor:   V. 
Chew).  John  Wiley  and  Sons,  Inc.,  New  York,  1958. 

19.  Chapin,    F.   S.   Experimental  Designs  in  Sociological  Research.   Harper  and 
Brothers  Publishers,  New  York,  1947. 

20.  Chew,  V.  (editor)  Experimental  Designs  in  Industry*  John  Wiley  and  Sons, 
Inc.,  New  York,   1958. 

21.  .  Basic     experimental    designs.     Experimental    Designs    in    Industry. 

(Editor:  V.  Chew).  John  Wiley  and  Sons,  Inc.,  New  York,  1958. 

22.  Cochran,  W.  G.,  and  Cox,  G.  M.  Experimental  Designs.  Second  Ed.  John 
Wiley  and  Sons,  Inc.,  New  York,  1957. 

23.  Connor,    W.   S.   Experiences   with  incomplete  block  designs.    Experimental 
Designs  in  Industry.   (Editor:  V.   Chew).  John  Wiley  and  Sons,   Inc.,   New 
York,  1958. 

24.   9  and  Young,    S.   Fractional  Factorial  Designs  for  Experiments   With 

Factors   at    Two  and    Three   Levels.   National  Bureau  of   Standards   Applied 
Mathematics  Series  58,  U.S.  Govt.  Print.  Of!.,  Washington,  IXC.,  Sept.  1, 
1961. 

25.   ,  and  Zelen,   M.  Fractional  Factorial  Experiment  Designs  for  Factors 

at  Three  Levels.  National  Bureau  of  Standards  Applied  Mathematics  Series 
54,  U.S.  Govt.  Print.  Off.,  Washington,  D.C.,  May  1,  1959. 

26.  Cox,  D.  R.  Planning  of  Experiments.  John  Wiley  and  Sons,  Inc.,  New  York, 
1958. 

27.  Daniel,   C.  Fractional  replication  in  industrial  research.   Proceedings  of  the 
Third  Berkeley  Symposium  on  Mathematical  Statistics  and  Probability.  Vol.  5 
(Editor:  J.  Neyman).  University  of  California  Press,  Berkeley,  Calif.,  1956. 

28.  Davies,   O.  L.   (editor)    The  Design  and  Analysis  of  Industrial  Experiments. 
Second  Ed.  Oliver  and  Boyd,  Edinburgh,  1956. 

29.  Davies,    O.   L.,  and   Hay,    W.   A.   The   construction  and  uses   of  fractional 
factorial  designs  in  industrial  research.  Biometrics,  6:233—49,  1950. 

30.  DeBaun,  R.  M.,   and  Schneider,  A.  M.  Experiences  with  response  surface 
designs.  Experimental  Designs  in  Industry.  (Editor:  V,  Chew).  John  Wiley 
and  Sons,  Inc.,  New  York,  1958. 

31.  Federer,  W.  T.  Experimental  Design.  Macmillan  Co.,  New  York,  1955. 

32.  Finney,    D.    J.    The   fractional   replication   of  factorial   experiments.    Ann. 
Eugenics,  12:291-301,  1945. 

33.  Fisher,  R.  A.  Statistical  Methods  for  Research  Workers.  Tenth  Ed.  Oliver  and 
Boyd,  Edinburgh,  1946. 

34.   .  The  Design  of  Experiments.  Fourth  Ed.  Oliver  and  Boyd,  Edinburgh, 

1947. 

35    f  an<i  Yates,  F.  Statistical  Tables  for  Biological,  Agricultural  and  Me&i- 

cal  Research.  Third  Ed.  Oliver  and  Boyd,  Edinburgh,  1948. 
36.   Horton,  W.  H.  Experiences  with  fractional  factorials.  Experimental  Designs 


436  CHAPTER    13,    OTHER    DESIGNS 

in  Industry.    (Editor:   V.    Chew).   John   Wiley   and   Sons,   Inc.,    New  York, 
1958. 

37.  Hunter,   J.   S.   Determination  of  optimum  operating  conditions  by  experi 
mental    methods,    Part    II-1-2-3,    models    and    methods.    Industrial    Quality 
Control,    15    (No's.   6-7-8):    16-24,   7-15,    and    6-14,    Dec.,    1958,   Jan.,    and 
Feb.,   1959. 

38.  Kempthorne,  O.  The  Design  and  Analysis  of  Experiments.  John  Wiley  and 
Sons,  Inc.,  New  York,  1952. 

39.  National  Bureau  of  Standards.  Fractional  Factorial  Experiment  Designs  for 
Factors  at  Two  Levels.  National  Bureau  of  Standards  Applied   Mathematics 
Series  48,  U.  S.  Govt.  Print.  Off.,  Washington,  D.C.,  April  15,  1957. 

40.  Plackett,  R.  L.,  and  Burman,  J.  P.  The  design  of  optimum  multi-factor  experi 
ments.  Biometrika,  33:305-25,  1946. 

41.  Quenouille,   M.  H.   The  Design  and  Analysis  of  Experiment.   Charles  Griffin 
and  Co.,  Ltd.,  London,  1953. 

42.  Rao,  C.  R.  General  methods  of  analysis  for  incomplete  block  designs.  Jour. 
Amer.  Stat.  Assn.,  42:541,  1947. 

43.  Satterthwaite,     F.    E.     Random    balance    experimentation.     Technometrics, 
I  (No.  2):lll-37;  May,  1959. 

44.  Snedecor,  G.  W.  Statistical  Methods.  Fifth  Ed.  The  Iowa  State  University 
Press,  Ames,  1956. 

45.  Youden,  W.  J.,  Kempthorne,  O.,  Tukey,  J.  W.,  Box,  G.  E.  P.,  and  Hunter, 
J.  S.  Discussion  of  the  papers  of  Messrs.  Satterthwaite  and  Budne  (including 
authors7   responses    to   discussion).    Technometrics,    1    (No.   2):157— 93,    May, 
1959. 


CH  APTE  R    14 

ANALYSIS  OF  COVARIANCE 

IN  PRECEDING  CHAPTERS  great  emphasis  has  been  placed  on  two  very 
important  techniques,  namely,  regression  analysis  and  analysis  of  vari 
ance.  Further,  in  certain  Sections  (11.11,  12.11,  and  13.6)  these  two 
techniques  were  combined  to  handle  particular  problems  associated 
with  the  exploration  of  response  curves  or  surfaces.  In  the  present 
chapter  we  shall  investigate  another  blending  of  these  two  fundamental 
tools.  This  new  technique,  known  as  analysis  of  covariance,  is  one 
which  has  proved  very  useful  in  many  areas  of  research. 

14.1     USES  OF  COVARIANCE  ANALYSIS 

Before  discussing  the  actual  methods  of  covariance  analysis,  let  us 
give  one  or  two  examples  of  situations  in  which  the  technique  may  be 
profitably  employed.  These  examples  should,  of  course,  indicate  to  the 
reader  the  nature  of  the  combination  of  the  ideas  of  regression  and 
analysis  of  variance.  For  the  first  example,  consider  a  case  where  the 
researcher  is  interested  in  the  effects  of  various  rations  on  the  weights 
of  hogs.  If  a  randomized  complete  block  design  is  utilized  and  the  final 
weights,  F,  of  the  animals  after  a  specified  number  of  days  of  feeding 
are  analyzed,  the  differences  among  the  effects  of  the  various  rations 
may  or  may  not  be  significant.  In  either  case,  however,  the  good  re 
searcher  will  think  more  about  the  conduct  of  the  experiment  before 
drawing  any  conclusions  from  the  analysis  of  variance  implied  in  the 
preceding  sentence.  He  might  say  to  himself,  "If  the  experimental  ani 
mals  varied  greatly  with  respect  to  their  initial  weights  at  the  time  the 
experiment  was  started,  how  do  we  know  that  differences  among  final 
weights  reflect  ration  effects  rather  than  just  varying  initial  weights? 
Calling  the  initial  weights  X,  he  might  adjust  the  F-values  according 
to  the  associated  X-values  and  then  analyze  and  interpret  the  experi 
mental  data.  The  method  by  which  this  is  carried  out  is  known  as  Co- 
variance  analysis. 

Another  example  is  the  following:  When  dealing  with  an  experiment 
to  compare  several  methods  of  teaching  statistics  in  which  the  criterion 
is  to  be  the  final  score,  Y,  obtained  by  the  students,  all  of  whom  take 
the  same  examination,  final  judgment  concerning  the  various  methods 
of  teaching  should  not  be  rendered  until  the  I.Q.  ratings,  X,  of  the  in 
dividual  students  have  been  examined  and  the  necessary  corrections 
(adjustments)  made.  Many  other  examples  could  be  given,  and  the 
reader  is  asked  to  formulate  some  in  his  own  field  of  interest  as  an  aid 
in  better  appreciating  the  techniques  to  be  presented. 

£4373 


438  CHAPTER    14,    ANALYSIS    OF    COVAR1ANCE 

14.2     ASSUMPTIONS  UNDERLYING  ANALYSES  OF 
COVARIANCE 

As  would  be  expected,  the  assumptions  one  makes  "when  performing 
a  covariance  analysis  are  similar  to  those  required  for  linear  regression 
and  analysis  of  variance.  Thus  we  find  the  usual  assumptions  of  inde 
pendence,  normality,  homogeneous  variances,  fixed  X's,  etc.  To  be 
more  specific,  we  give  the  mathematical  models  associated  'with  some 
of  the  more  common  designs  when  a  covariance  analysis  is  contem 
plated. 

Completely  randomized  design 


YH  -  f  Hr  Tt  +  pXv  +  €<,-;  i  =   1,  -  -  •  ,  k        (14.1) 

j  =   1,  -  -  -  ,  n 
Randomized  complete  block  design 

+  *«;  i  =  1,  •  •  •  ,r       (14.2) 


Latin  square  design 

Yijk  =  £  +  Pi  +  yj  +  rk  +  0Xvm  +  e»/(*o;      i  =  1,  -  -  -  ,  m      (14.3) 

j  =  1,  -  -  -  ,  m 
k  =  1,  •  •  •  ,  m 

Two-factor  factorial  in  a  randomized  complete  block  design 

Y-iik  =  £  +  Pi  +  <*j  +  Vk  +  (&v}jk  +P  Xijk  +  eijk',     i  =  1?  -  •  •  ,  r    (14.  4) 

j  =  1,  -  •   •  ,  a 
k  =  1,  -  •  •  ,  c. 

In  practice  it  is  more  customary  to  express  these  equations  in  terms 
of  deviations  of  the  X  variable  from  its  mean.  When  this  is  done,  the 
equations  appear  as 

YV  =  M  +  T*  +  0(Xt,-  -  X)  +  €,y;  i  =   1,   -   -   -  ,  k         (14.5) 


j  =  i,  .  .  .  ,  n 
-  M  +  Pi  +  TJ  +  /3(Jr<y  -  X}  +  €,y;  i  =   1,  -   -  -  ,  r         (14.6) 


F,-yJb    =    M    +    Pi   +    Vj    +    Tk 

+  /3(-y<yc*)  -  3)  +  €</»);          i  =  1,  •   •  •  ,  m        (14,7) 

y  =  i,  -  -  -  ,  m 

k  =   1,  •   •  •  ,  ra 
and 


14.3       COMPLETELY    RANDOMIZED    DESIGN  439 

y^jk  =  M  +  pi  +  dj  H-  yk  +  (coOy* 

+  /3(Xv*  -  X)  +  €</*;  i  =   1,  -  -   -  ,  r       (14.8) 

y  =  i,  •  •  •  ,  a 

z>  =   1     ...      r 

**  •*•»  ?    °? 

respectively,  where  /j,=*g-\-p'X,  and  3s  is  the  arithmetic  mean  of  all  the 
X's.  The  reason  for  writing  the  equations  in  this  last  form  is  that,  by 
so  doing,  we  simplify  the  algebra  of  the  mathematical  solution.  Conse 
quently.,  it  is  in  this  form  that  you  will  find  the  model  presented  in 
most  texts.  The  usual  assumptions  are  made  concerning  the  various 
terms  in  each  of  the  preceding  equations,  and  the  reader  is  advised  to 
read  again  the  appropriate  sections  in  Chapters  8,  11,  12,  and  13  to 
refresh  his  memory  on  such  matters. 

Although  not  mentioned  above,  there  is  another  item  which  is  com 
monly  considered  as  a  necessary  assumption  for  a  valid  analysis  of 
covariance,  namely,  that  the  concomitant  variable,  X,  should  not 
be  affected  by  the  treatments.  That  is,  the  treatments  which  have  been 
applied  to  the  experimental  units  so  that  we  may  observe  and  judge 
their  effects  on  the  Y  variable  should  not  influence  the  observed  values 
of  -XT.  However,  this  is  too  restrictive  an  assumption.  Even  though  the 
treatments  do  affect  the  JXT-values,  a  covariance  analysis  may  be 
profitably  employed  if  proper  care  is  exercised  in  the  interpretation  of 
the  experimental  results.  It  is  clear,  then,  that  the  inferences  which 
may  be  made  are  different  in  the  two  cases,  depending  upon  whether 
or  not  the  X  variable  is  affected  by  the  treatments.  The  researcher  is 
therefore  cautioned  to  be  extremely  careful  when  dealing  with  the  inter 
pretation  of  covariance  analyses.  Let  us  now  consider  some  special 
cases  so  that  the  reader  will  not  only  gain  practice  in  the  interpretation 
of  data  amenable  to  an  analysis  of  covariance  but  also  become  familiar 
with  the  details  of  the  computational  procedure. 

14.3   COMPLETELY   RANDOMIZED    DESICN 

As  our  first  example  of  a  covariance  analysis  we  shall  make  use  of  a 
completely  randomized  design.  Before  giving  a  numerical  example,  let 
us  examine  the  problem  in  general  form.  Assuming  that  we  have  t 
treatments,  or  groups,  and  that  there  are  n^  observations  on  each  of  X 
and  Y  in  the  ith  group,  it  is  customary  to  proceed  as  follows:  Calculate 
the  following  sums  of  squares  and  products  and  then  complete  Table 
14.1  as  indicated.  (NOTE:  There  is  a  great  similarity  between  Table 
14.1  and  Table  8.22.) 


x2  =  corrected  total  sum  of  squares  for  X 

/        *          "^ 

\  ;=i  y-i 


y-y-^       ^=A^     '  (14*9) 

=  2^  2^  xa  — 
1=1  y==i 


Pi 

faO 


p 


i 

o 


a 

o 


a1 


g 

o 


ft 
M 


O 
U 


OJ 

<u 

o 


PQ 


-0 

X-  v 

a 

| 

w 

3 

4) 

T 

4J 

c5 

£ 

•»—  i 

.S 

3 

\ 

3 

.2 

CO 

l/Nv 

^ 

fl 

1 

^^ 

^? 

"o 

a 

0 

^" 

^"s* 

I 

&) 

'§ 

K5 
CO 

5 

I 

II 

a 

8 

| 

go 

T—  t 
i 

T 

CM 
1 

1 

S 

&1 

s 

3 

^ 

5 

.2 

(^f3-! 

*»r  IT 

«r/j"jf 

J^ 

"a 

.2 

W 

^'* 

'* 

S 

"? 

4_» 

53 
ft 

W 

$ 

H 
H 

^ 

"S 

^" 

A 

"co 

^ 

d. 

'gl 

Wc4 

1^ 

*^  [J 

•i—l 

w 

i 

1! 

f  J"^ 

1 

T 

ii 

^  + 

0 

8. 

W 

CO 

CO 

co 

s 

b 

w~ 

^ 

W 

T3 

£ 

# 

W 

II 

-*2 

ft                          A 

II 

|^ 

I 

c| 

y 

o 

fl 

w 

a 

^ 

+ 

*0 

Cu 

^5 

^ 

03 

tn 
<U 

w 

I) 

S 

« 

1 

s£        4 

II 

*? 

I 

a 
.S 

CO 

T3 

*0 

a 

J 

1 
a 

1 

1 

•M 

CO 

•Jt 

w 

f 

1 

JJ  - 

(U 

E^            ^ 

co 

I 

A 

* 

«4H 

T-l 

*s 

CO 

Sg 

T 

I 

a1 

rt 

OJ^ 

;* 

£ 

0 

*  fe 

^fi 

JL      -W-1 

-Wit 

d 

C|,§ 

r-H 

b£ 

|H 

iT.P 

1     
1 

4 

•* 

H 

Source  of 
Variation 

Among  treatments 

Among  experimen 
tal  units  treated 
alike  (within 
treatments)  

Among  treatments 
+ 
Within  treatments 
(  =  total).  

Difference  for  testir 

*  The  symbols  Sx 
other  tables  which 

14.3       COMPLETELY    RANDOMIZED    DESIGN  441 

xy  =  corrected  total  sum  of  products  for  X  and  Y 


t        ni  \     /       t        m 

2;  2:**)  (2:  2: 

—  i  ,-i          /  \  .-i  y-i 


x  r  —  i  ,-i          /  \  .-i  y-i          /  (14.  10) 

=  2^ 


2^  ni 
y2  =  corrected  total  sum  of  squares  for  F 


F~.  _     (14-  u> 


t=-l 
.  =  treatment  sum  of  squares  for 


(ni  \2  /       *         ni 

T^  jc    \  I  T^  V  x 

Z-*t   ^  13  J  \      Z-*i    2-*t    ^  *} 

-2.     ^      '    -      -1  '71 


i—  i 
/  =  treatment  sum  of  products  for  X  and  Y 


yy 

2  t         ni  2 


(14.13) 


Tyy  =  treatment  sum  of  squares  for  Y 

\  2 

(14.14, 


=  experimental  error  sum  of  squares  for  X 
2  —   T 

—    -L  xx 


=  experimental  error  sum  of  products  for  X  and  F 


Eyy  =  experimental  error  sum  of  squares  for  F 


442  CHAPTER    14,    ANALYSIS    OF    COVAR1ANCE 

The  proper  F-ratio  for  testing  the  hypothesis  that  there  are  no  dif 
ferences  among  the  true  effects  of  the  t  treatments  on  the  Y  variable 
after  adjusting  for  the  effect  of  the  X  variable  is 


F  « 


A* 
z;,.,-  *- 
i=l 

with  degrees  of  freedom  v\  =  t  —  1  and 


It  is  customary,  in  addition  to  performing  the  jP-test  just  indicated,  to 
present  a  table  of  adjusted  treatment  means  as  an  aid  in  the  interpre 
tation  of  the  experimental  results.  The  adjusted  means  may  be  found 
using  the  formula 


adj.   Yt  =    Y;  -  b(Xt  -  X);  i  =  1,  .  -  -  ,  t,  (14,  19) 

where  6  is  the  regression  coefficient  calculated  from  the  experimental 
error  sums  of  squares  and  products,  that  is,  'b  =  Eaxu/EaiX.  The  estimated 
variance  of  an  adjusted  treatment  mean  is  given  by 


V  (adj.  F,)  =  s^E  \-  +  (^~T)21  (14  .  20) 

Lni  Exx        J 

and  the  standard  error  of  an  adjusted  treatment  mean  by 

f\ 

«  s    y 


(adj.   F,)  «  s    y  --  1  --  ~  --        (14.21) 
n*  Exas 

The  estimated  variance  of  the  difference  between  two  adjusted  treat 
ment  means  is,  of  course,  given  by 

+  (Xi  ~  ^°T     (14.  22) 


V  (adj.  F,  -  adj.  Fy)  -  4  \-  +  — 

L^i         ny 


It  should  be  clear  that  the  regression  coefficient,  /?,  in  Equation  (14.5) 
has  been  assumed  to  be  nonzero.  If  such  were  not  the  case,  the  intro 
duction  of  the  concomitant  variable,  X,  into  the  calculations  would  be 
an  unnecessary  complication.  Sometimes  the  researcher  will  wish  to 
check  on  this  assumption.  That  is,  he  will  consider  the  hypothesis 
H:&  =  Q,  rather  than  the  assumption  £=^0.  When  this  is  done,  he  will 
be  interested  in  testing  the  validity  of  H.  The  proper  F-ratio  is 


SE 


(14.23) 


74.3       COMPLETELY    RANDOMIZED    DESIGN 


443 


which  has  degrees  of  freedom  v-L  =  l.  and 


TABLE   14.2-Gains  In  Weight  (F)  and  Initial  Weights  (X}  of 
Pigs  in  a  Feeding  Trial 


Treatment 

1 

2 

3 

4 

X 

F 

X 

F 

X 

F 

X 

F 

30 

165 

24 

180 

34 

156 

41 

201 

27 

170 

31 

169 

32 

189 

32 

173 

20 

130 

20 

171 

35 

138 

30 

200 

21 

156 

26 

161 

35 

190 

35 

193 

33 

167 

20 

180 

30 

160 

28 

142 

29 

151 

25 

170 

29 

172 

36 

189 

Total 

160 

939 

146 

1031 

195 

1005 

202 

1098 

TABLE    14.3-Analysis  of  Covariance  for  Data  in  Table  14.2 


Source  of 
Variation 

Degrees 
of 
Freedom 

Sum  of  Squares  and  Products 

Deviations  About  Regression 

2> 

lL,xy 

Z^2 

^v  £>»• 

Degrees 

of 
Freedom 

Mean 
Square 

^     2> 

Among  treat 
ments  

3 
20 

365.46 
361.50 

451.21 
496.83 

2163.13 
5937.83 

Among    ani 
mals  treat 
ed  alike.  .  . 

Total.  . 

5255.01 

19 

276.58 

23 

726.96 

948.04 

8100.96 

6864  .  61 
1609.60 

22 
3 

536.53 

Difference  for  testing  among  adjusted  treatment 
means 

Example  14.1 

Given  the  data  of  Table  14.2,  the  following  calculations  were  made 
and  the  results  reported  in  Table  14.3. 

]L>2  =  Sxx  =  TX*  +  Exx  =  (30)2  +  .  -  -  +  (36)2  -  ^—^  =  726.96 


•-£*!,=  (30)  (165) 


24 

4-  (36)  (189)  - 


(703)  (4073) 
24 


948.04 


444 


CHAPTER    14,    ANALYSIS    OF    COVARIANCE 


Tyy  +  Evy  =  (165)* 


h  (189)2  - 


(4073) 2 


24 


8100.96 


(160)2  +  (146)2  +  (195)2  +  (202)2       (703)2 


365.46 


•*•  xy    === 


T        — 
-*  vy  — 


6  24 

(160)  (939)  +  (146)  (1031)  +  (195)  (1005)  +  (202)  (1098) 


__  (703)  (4073) 
"""  24 

(939)2+  (103 1)2 


=  451.21 

(1005) 2 


(1098) 2       (4073) 2 


24 


2163.13 


=  S**  -  Txx  =  361.50 
=  S*v  —  Txv  =  496.83 

^  *^yv  —    •*•  yy  ==   59o/.0v5. 

Carrying  out  the  F-test  outlined  in  Equation  (14.18),  we  obtain 
F  =  536. 53/276. 58  =  1.94  with  degrees  of  freedom  ?i  =  3  and  j>2  =  19.  This 
is  not  significant  at  the  5  per  cent  level,  and  thus  we  are  unable  to  reject 
the  hypothesis  of  no  differences  among  the  true  effects  of  the  4  treat 
ments  on  the  gain  in  weight  of  pigs  after  adjusting  for  the  varying 
initial  weights  of  the  experimental  animals.  Incidentally,  in  this  case 
the  same  decision  would  have  been  reached  had  no  adjustment  been 
made  for  the  concomitant  variable.  However,  in  many  instances  the 
conclusions  may  change  considerably  depending  on  whether  or  not  the 
covariance  technique  is  used,  and  thus  the  researcher  should  always 
see  if  it  is  applicable  to  the  problem  at  hand.  The  adjusted  treatment 
means  are  presented  in  Table  14.4. 

TABLE   14.4-Calculation  of  Adjusted  Treatment  Means  From 

Data  of  Table  14.2 

(jr  =  29.29,  7=169.71,  6  =  496.83/361.50=1.374) 


Treatment 


1 

2 

3 

4 

Xi  

26.67 

24.33 

32.50 

33.67 

Xi—X  

—      2.62 

—      4.96 

3.21 

4.38 

b(Xi—  T)  

—      3.60 

—      6.82 

4.41 

6.02 

Yi  

156.50 

171.83 

167.50 

183  00 

adj.  ~Yi  

160.10 

178.65 

163  09 

176  98 

Standard  error  of  adj. 
Tf  

7.17 

8.06 

7.35 

7.80 

14.4      RANDOMIZED  COMPLETE   BLOCK   DESIGN 

When  our  data  conform  to  a  randomized  complete  block  design,  the 
appropriate  mathematical  model  is  as  given  in  Equation  (14.6).  The 
analysis  to  be  performed  is  given  in  Table  14.5,  the  quantities  Rxx, 
T**,  Ex*,  RVV)  Tyy,  and  Eyy  being  obtained  as  in  any  randomized  com- 


o 

7 

T-l 

^ 

J 

^—  i 

S3 

^*, 

I 

cr 

<^*s 

fl 

CO 

J 

i^ 

*5? 

d 

1 

^ 

ft 

1 

^r* 

1 

fl 

-^ 

.2 

J 

;r 

-S 

to 

" 

Co 

3 

a 

*t^ 

**~* 

o 

o3 

£ 

+J 

M 

0 

T-H 

IS* 

"S 

4> 

J^ 

1 

0 

£ 

^-( 

T3 

s 

*s 

to 

2; 

7 

s 

.2 

Si 

8 

s 

| 

I 

"s 

o 

rrt 

1 
1 

s 

V. 

i 

1 

f—  j 

"a 

«d 

^ 

4 

„ 

ce 

N 

2 
.a 

y 

1 
i» 

II 

^c|^ 

's 

i 

II 

^T  + 

1 

w 

co" 

«o 

to 

O 

U 

S 

o3 

4-> 

+ 

ft 
M 

CO 

w 

ft 

£ 

1 

S  g       S 

II 

8 

0 

^          ^ 

CO 

d 

j-i 
PH 

^ 

*»H 

T3 

tq 

ce 
o 

§ 

to 

If 

4- 

5k 
H 

s 

O 

MH 
0 

W3 

fi 

ci 

1 

w 

«  4 

0 

.1 

8 

g 

^ 

*0 

c 

G 

| 

4 

d 

M 

<^ 

C/3 

e^ 

-f- 

-M 

g 

w 

1 

CO 

P 

3 

B?  EH'        &? 

J 

f 

1 

,—  „ 

bo 

•^ 

T-i 

*1 

»  g 

1* 

1 

5 

'  •§ 

^*, 

cJ 

w 

i 

4       O 

'  v 

1—  ( 

W) 

1 

"H    •«—  (                     | 

1 

d 

< 

< 

?* 

ii        i 

I 

CO 
0 

^ 

*+ 
C 

3  g 

:  «j3 

CO        I 

2 

C 

c/ 

33 

II 

1  HI  g 

eatment 
+  error. 

8 

§ 
£ 

f^^^f^  ** 

k* 

H 

S 

446  CHAPTER    14,    ANALYSIS    OF    COVAR1ANCE 

plete  block  design.  The  sums  of  products  Rxyj  Txy,  and  Exy  are  found 
using  the  following  equations: 

=  corrected  total  sum  of  products  for  X  and  Y 

(14.  24) 


F<,  ±  F* 


=  replicate  (block)  sum  of  products  for  X  and  Y 

±  F,A 

^-^/    •*•  *J   I 

(14.25) 
t  rt 

,=  treatment  sum  of  products  for  X  and  Y 

3 1  \   -  ,    -   „      **  / 

(14.26) 

(14.27) 

The  proper  F-ratio  for  testing  the  hypothesis  of  no  differences  among 
the  true  effects  of  the  t  treatments  on  the  F  variable  after  adjusting  for 
the  effect  of  the  X  variable  is 


r  rt 

=  experimental  error  sum  of  products  for  X  and 


(ST+E  - 


(  } 


As  in  the  case  of  a  completely  randomized  design,  we  will  wish  to 
calculate  the  adjusted  treatment  means  and,  possibly,  to  test  the  hy 
pothesis  that  0  of  Equation  (14.6)  is  0.  The  calculation  of  the  adjusted 
treatment  means  is  easily  carried  out  using 

adj.   Y,,  =    F.y  -  b(X.,  -2);          j  =  1,  -  .  -  ,  t,         (14.29) 

where  &  is  the  regression  coefficient  computed  from  the  experimental 
error  sums  of  squares  and  products,  that  is,  b  =  E^/E^.  The  estimated 
variance  of  an  adjusted  treatment  mean  is 


V  (adj.  F.,-)  =  4  f-1  +  V-*-  ^1  (14.  30) 

L  r  Exx         J 

and  the  standard  error  of  an  adjusted  treatment  mean  is 


Jadi.  ?.,  =  VF(adj 


j.  F.y)  =  SE\/~  +  (X\     ^  •        (14.31) 


14.4       RANDOMIZED    COMPLETE    BLOCK    DESIGN 


447 


The  estimated  variance  of  the  difference  between  two  adjusted  treat 
ment  means  is  _ 


V  (adj.  F.y  -  adj.  TV)  =  4    -  + 


To  test  the  hypothesis  that  13  equals  0,  we  calculate 


F  = 


(14.32) 


(14.33) 


which  has  degrees  of  freedom  v±=  1  and  v2  =  (r—  1) (t—  1)  — - 1. 


TABLE   14.6- Yields  for  3  Varieties  of  a  Certain   Crop  in  a  Randomized 
Complete  Block  Design  With  4  Blocks 

(^  ==  yield  of  a  plot  in  a  preliminary  year  under  uniformity 
trial  conditions;  Y  — yield  on  the  same  plot  in  the 
experimental  year  when  the  3  varieties  were  used) 


Varieties 

"Rlorlr 

Block 

A 

B 

C 

Totals 

1  

X 

54 

51 

57 

162 

Y 

64 

65 

72 

201 

2  

X 

62 

64 

60 

186 

Y 

68 

69 

70 

207 

3 

X 

51 

47 

46 

144 

Y 

54 

60 

57 

171 

4 

X 

53 

50 

41 

144 

Y 

62 

66 

61 

189 

Variety.  .  .  . 

X 

220 

212 

204 

636 

Totals  .  . 

Y 

248 

260 

260 

768 

Reproduced  from  Table  7  in  John  Wishart,  Field  Trials  II:  The  Analysis  of  Covariance, 
Tech.  Comm.  No.  15,  Commonwealth  Bureau  of  Plant  Breeding  and  Genetics,  School  of 
Agriculture,  Cambridge,  England,  May,  1950.  With  permission  of  author  and  publishers. 

Example  14.2 

Consider  the  data  of  Table  14.6.  These  data  have  been  examined  in 
considerable  detail  by  Wishart  (17);  we  shall,  however,  consider  them 
from  a  more  limited  point  of  view,  which  will  be  sufficient  for  our  pur 
poses.  The  required  calculations  are 


396 


12 
(162)2  4-  (186)2  +  (144)2  4-  (144) 2       (636) 2 


12 


448 


CHAPTER    14,    ANALYSIS    OF    COVARIANCE 

(220)2  +  (212)2  +  (204)2       (636)2 


12 


32 


514  —  396  —  32  =  86 
+  ----  h  (61)2  __ 


324 


(201)2  +  (207)2  +  (171)2  +  (189)2       (768)  * 
-  — 

(248)2  +  (260)  2  +  (260)  2       (768)  2 


252 


__ 

-  24 


-  324  -  252  —  24  =  48 

-  (54)  (64)  -i  ----  +  (41)  (61)  - 


286 


(162)  (201)  +  (186)  (207)  +  (144)  (171)  +  (144)  (189)        (636)  (768) 


(220)  (248)  +  (2 12)  (260)  +  (204)  (260)        (636)  (768) 

4  ~~  12 

=  286  —  264  —  (—24)  =  46, 
and  these  are  summarized  in  Table  14.7. 


12 

—  24 


264 


TABLE   14.7-Analysis  of  Covariance  for  Data  of  Table  14.6 


Source  of 
Variation 

Degrees 
of 
Free 
dom 

Sum  of  Squares 
and  Products 

Deviations  About 
Regression 

2>2 

][>:y 

Zy 

IZy*         ^ 
-(2>30VZ> 

Degrees 
of  Free 
dom 

Mean 
Square 

Replicates  (blocks)  . 
Treatments      (varie 
ties) 

3 

2 
6 

396 

32 
86 

264 

-24 
46 

252 

24 
48 

Experimental  error. 

23.4 

5 

4.68 

Treatments  +  error  . 

8 

118 

22 

72 

67.9 

44.5 

7 
2 

22.25 

Difference  for  testing  among  adjusted  variety  means  .  . 

Before  testing  the  hypothesis  of  no  differences  among  the  true  effects 
of  adjusted  varieties,  let  us  test  the  hypothesis  that  the  true  regression 
coefficient,  0,  is  0.  After  all,  it  is  more  reasonable  to  examine  this  point 
first,  for  unless  we  can  reject  such  a  hypothesis,  that  is,  unless  we  can 
safely  conclude  that  /3  5^0,  the  decision  to  perform  a  regular  analysis  of 
covariance  is  questionable.  Accordingly,  we  compute  F  —  [(46)2/48]/4.68 
=  9.42  with  degrees  of  freedom  *>i  =  l  and  v*=* 5.  Since  F  =  9. 42  >Ff 95(1,5) 
=  6.61,  we  may  reasonably  assume  that  /J  is  not  0  and  thus  be  justified 
in  performing  a  covariance  analysis. 

We  shall  now  examine  the  variety  differences.  First,  we  note  that  an 
ordinary  analysis  of  variance  would  give  rise  to  F  =  (24/2)/(48/6) 
=  12/8  =  1.5  with  degrees  of  freedom  j>i  =  2  and  j>2  =  6,  and  this  would 
not  permit  us  to  reject  the  hypothesis  of  no  differences  among  the  true 


14.5       LATIN    SQUARE    DESIGN  449 

effects  of  the  3  varieties.  Let  us  next  observe  what  effect,  if  any,  the  per 
formance  of  a  co variance  analysis  will  have  on  our  inferences.  To  test 
the  hypothesis  of  no  differences  among  the  true  effects  of  the  3  varieties 
after  adjusting  for  the  effect  of  the  natural  fertility  differences  from  plot 
to  plot,  as  measured  by  the  uniformity  trial,  we  compute  F  =  22. 25/4. 68 
=  4.75  with  degrees  of  freedom  v\  =  2  and  v2  =  5.  Since  F=4.75  lies 
between  F.90(2,5)  =  3.78  and  F. 95(2,5)  =5.79,  the  variety  differences  would 
not  ordinarily  be  called  statistically  significant.  However,  the  reader  is 
cautioned  again  that  the  choice  of  the  significance  level  is  quite  arbi 
trary  and  the  above  results,  therefore,  may  well  indicate  differences  that 
are  of  real  importance.  Apart  from  this,  we  note  that  the  adjustment  of 
the  yields  for  the  unequal  fertilities  of  the  experimental  plots  has  caused 
the  resulting  F-value  to  approach  more  closely  what  is  conventionally 
thought  of  as  a  critical  value.  This  should  suggest  to  the  researcher  that 
quite  likely  the  fertility  differences  among  the  plots  are  tending  to 
obscure  the  true  differences  among  varieties.  If  the  experiment  were 
performed  again  with  more  replication,  significant  results  might  be 
obtained. 

14.5      LATIN   SQUARE   DESIGN 

The  performance  of  an  analysis  of  covariance  on  data  resulting 
from  a  Latin  square  design  introduces  no  new  concepts.  Thus,  we  shall 
proceed  immediately  to  outline  the  computational  technique  and 
indicate  the  appropriate  test  procedures.  The  only  calculations  besides 
those  specified  for  an  ordinary  analysis  of  variance  in  a  Latin  square 
are  the  sums  of  products.  These  are  found  as  follows: 

=  corrected  total  sum  of  products  for  X  and  Y 

_"*     *** 

2: 2: 


m        m 


=  row  sum  of  products  for  X  and  F 


m  (14.35) 


column  sum  of  products  for  X  and  F 

m       /     m  \      /     rn 

z:(  i:x-«c*))(  z: 

/— 1     \    z-»l  /     \  i-1 


(14.36) 


45O  CHAPTER    14,    ANALYSIS    OF    COVARIANCE 


m        m 


TW  =  treatment  sum  of  products  for  X  and  Y 

m  /     jm        m  x      /      m        rn  \ 

Z  («r*)  G,rfc)       (  Z  Z  *<,<*,  )  (  Z  Z  Ftfob,  )        (14  .  37) 

k=t  \    t—l    jr'=l  /     \    i=l    j=l  / 

m  m* 

E*y  =   5>y  ~  ^-v  ~"  C^  —  r^,  (14.38) 

where 

xT^  =  sum  of  all  X  observations  associated  with  treatment  k       (14.39) 
yTk  ==  sum  of  all  Y  observations  associated  with  treatment  k.     (14.40) 

The  results  of  these  calculations  are  presented  in  Table  14.8.  The  test 
of  H  :  /3  =  0  is  given  by 

F  =  _  E**"/Exx  _  =    E^/Exx  (14  41) 

SE/[(m  -  l)(m  -  2)  -  1]  ^  U    ^ 


with  degrees   of   freedom   j>i=l   and   v^—  (m—~  l)(m  —  2)  —  1.    To    test 
among  adjusted  treatment  means  we  compute 


=  =         + 

^/[(«  -!)(«-  2)  -  1]  4  ' 

with  degrees  of  freedom  ^i  =  ?7^  —  1  and  ^2  =  (m  —  1)  (m  —  2)  —  1  . 
The  adjusted  treatment  means  may  be  found  using 


adj.  F..0b)  -  7..(Jfc)  -  SC^.c*)  -  ^)  (14.43) 

where  6  is  the  regression  coefficient  associated  with  experimental  error, 
that  is,  b=Exv/Exx.  The  estimated  variance  of  an  adjusted  treatment 
mean  is 


V  (adj.  7..w)  -  4  |^—  +  v     "w           '     \ ,  (14.44) 

and  the  estimated  variance  of  the  difference  between  two  adjusted 
treatment  means  is 

f'Cadj.  "F..C*)  —  adj.  7..(fc/))  =  4    —    ^ — "(&/)      L  (14.45) 


o 


cr 

CO 

es 


1 

cr 
to 


<i> 
tf 


c« 
o 


§ 

X 


(U 

Q 


.*.        A 


Sf 


a 


cu 

o 

00 


j 

R 

o 

W 


+ 


w 


w 


&| 

p^ 


"s  § 

4)  *£5 

S'C 

o  ,rt 

CO  > 


ft 


e 


I  1   I 


I 

03 
W) 
fi 


S 


§ 

£ 


W 


452  CHAPTER    14,    ANALYSIS    OF    COVARIANCE 

14.6     TWO-FACTOR   FACTORIAL   IN   A   RANDOMIZED 
COMPLETE   BLOCK  DESIGN 

The  performance  of  a  covariance  analysis  when  the  treatments  are 
of  a  factorial  nature  follows  the  pattern  established  in  the  preceding 
sections.  The  only  refinement  that  occurs  is  one  which  is  naturally 
anticipated  once  we  know  we  are  dealing  with  a  factorial  setup:  We 
are  now  able  to  test  among  adjusted  means  for  each  of  the  factors  and 
for  all  the  interactions.  The  appropriate  calculations,  in  addition  to 
those  outlined  earlier  are  given  below: 

=  corrected  total  sum  of  products  for  X  and  Y 

r         a          b 

=  23  23  23  XjjkYijk 

a=l    y=l    fc=l 

(14.46) 

r         a         b  \      /      r         a         b 


(r         a         b  \      /      r         a         b 

z  z  z  *«*)  (  z  z  z  Yij 
i=-l    y=l    £=1  /     \    i=l    y—1    A;«*l 


rab 
=  replicate  sum  of  products  for  X  and  Y 

b  /      a          b 


T        /      a          b  \     /      a          b 

z(  sz;*«0(5:i:r«* 

i=*l    \  y=l    A=l  /     \  /«!    A:=l 


ab 

(14.47) 

(r          a          &  N.       /      r          a          b 

Z  Z  Z  x«*  )  (  Z  Z  Z  F«» 
4^.1  y=i  <fe—  =1  /   \  a-«i  y«i  A—I 


rab 
=  sum  of  products  for  X  and  Y  for  the  a  X  b  table 


(r         a,         b  \      /      r         a    '     b 

Z  Z  Z  x«*)  (  Z  Z  Z 
x=l   y=»i    &*=•!  /    \   i,-,!   y»*i    fc=l 


a,       /      r          b  \    /      r          b 

Z  (  Z  Z  x«*)  (  Z  Z  F 

y—1    \   i=l  A— 1  /    \  t— 1  A=-X 


(r          a          b  \.       /      r          a          b 

Z  Z  Z  xv)  (  Z  Z  Z  r< 
^=1  y—i  A«=I  /    \  i=i  y=-i  jfc=-i 


(14.48) 


(14.49) 

a,        /      r          b  \     /      r          b 


r5  (14.50) 


raft 


14.6      TWO-FACTOR   FACTORIAL  453 


J>        /       r          a  v       /      r          a, 

z;  ( z  z;  *«»)  ( z:  i: 

fc=l   \  i«l  J—l  /  \  i=l  JW1 


ra 


(14.51) 

(r          g          5  \       /      r          a,          b 

z:  z:  z;  *«»  )  (  z;  z:  z  r 
1=1  y«=i  jb^i  /    \  twi  y=i  &=i 

rob 
Bxy.  (14.52) 

These  calculations  are  summarized  in  Table  14.9.  Following  are  the 
appropriate  ^-ratios  for  testing  the  hypotheses:  (1)  no  differences 
among  the  true  effects  of  the  levels  of  factor  a  on  the  variable  Y  after 
adjusting  for  the  concomitant  variable  X,  (2)  no  differences  among  the 
true  effects  of  the  levels  of  factor  &  on  the  variable  Y  after  adjusting 
for  the  concomitant  variable  X,  and  (3)  no  interaction  between  factor 
a  and  factor  6  as  they  affect  the  variable  Y  after  adjusting  for  the  effect 
of  the  concomitant  variable  X,  respectively, 

-  —  1) 

'  /•.,    ,      CTQ\ 

'    (14'53) 


. 


and 

=  (^H^-  =      M, 


The  test  of  H:@  =  Q  is  performed  by  calculating 

F  =  — 2^ — —  (14.56) 

with  degrees  of  freedom  n=l  and  j>2=  (r—  l)(ab  — -1)  —  1. 

The  appropriate  standard  errors  for  the  various  mean  effects  are 
found  from  the  following  estimated  variances : 


A-effect  V  (adj.  T.y.)  -  ^~  +        "  <14' 57) 

B-effect  V  (adj.  T..*)  =  ^  f—  +  (^"* _"  ^H  (14.  58) 

Lra^  J^x*         J 

f  (adj.  7^)  =  *Z  f-  +  (X^~X)1 '  <14' 59) 

L   ^  -^sx  -I 


8 

ft 


I 


a 

o 

"B 


-1 

O 
•+-> 
O 


1 


I 


I 


O 

-o 


w 
1? 

w 


sf 


» 

Co* 


I 

•§ 


I  I 

-Cl 


co"1 


T-H 
I 


I 

J 


O) 
CD 
O 


PQ 


1 

PH 

"& 


cr^ 

CO 


w 


w 


w 


I 


« 


4 


«?* 

II  fc? 

<*+ 


• 


CCj      « 


•T 


i 

•8 


i — I 
I 


| 


«> 
1 


bO 


4J, 


1   + 


§3  a 

"  a 


a 

, 


s 

-|  s 


^4+erro 


Diff 


fference 


"S 


a* 

o 

a 


14.6       TWO-FACTOR    FACTORIAL 


455 


Example  14.3 

As  an  example  of  a  covariance  analysis  in  a  randomized  complete 
block  design  where  the  treatments  are  of  a  factorial  nature,  consider 
the  data  of  Table  14.10,  These  data  were  originally  examined  by 
Wishart  (17),  and  the  interested  reader  is  referred  to  his  study  for  a  more 
detailed  discussion  than  we  shall  give  here. 

In  discussing  the  data  of  Table  14.10,  we  shall  consider  the  5  pens  as 
5  replicates,  and  thus  we  have  a  3X2  factorial  in  a  randomized  complete 
block  design.  Following  the  calculational  procedure  outlined  in  Table 
14.9,  we  arrive  at  the  results  presented  in  Table  14.11. 

To  test  H:{3  =  0,  we  calculate 


F  = 


(39.367)  V442.93 
0.2534 


13.81 


with   degrees   of  freedom  PI  =  1    and  v2=19,    and  this  is  significant  at 
a.  =  .01.  The  various  treatment  effects  may  also  be   tested  for  signifi- 


TABLE   14.10-Initial  Weights  and  Gains  in  Weight  of  Young  Pigs 
in  a  Comparative  Feeding  Trial 

(X  =  initial  weight  in  pounds;  Y  —  gain  in  weight  in  pounds} 


Pen 

Feeding  Treatments 

Totals 

A 

B 

C 

Male 

Female 

Male 

Female 

Male 

Female 

I 

X 
Y 

38 
9.52 

48 
9.94 

39 
8.51 

48 
10.00 

48 
9.11 

48 
9.75 

269 
56.83 

II 

X 
Y 

35 
8.21 

32 
9.48 

38 
9.95 

32 
9.24 

37 
8.50 

28 
8.66 

202 
54.04 

III 

X 
Y 

41 
9.32 

35 
9.32 

46 
8.43 

41 
9.34 

42 
8.90 

33 
7.63 

238 
52.94 

IV 

X 
Y 

48 
10.56 

46 
10.90 

40 
8.86 

46 
9.68 

42 
9.51 

50 
10.37 

272 
59.88 

V 

X 
Y 

43 
10.42 

32 

8.82 

40 
9.20 

37 
9.67 

40 
8.76 

30 
8.57 

222 
55.44 

Totals 

X 
Y 

205 
48.03 

193 
48.46 

203 
44.95 

204 
47.93 

209 
44.78 

189 
44.98 

1203 
279.13 

Reproduced  from  Table  11  in  John  Wishart,  Field  Trials  II:  The  Analysis  of  Covariance, 
Tech.  Comm.  No.  15,  Commonwealth.  Bureau  of  Plant  Breeding  and  Genetics,  School  of 
Agriculture,  Cambridge,  England,  May,  1950.  With  permission  of  author  and  publishers. 


456  CHAPTER    14,  ANALYSIS   OF   COVARIANCE 

TABLE   14. 11- Analysis  of  Covariance  for  the  Data  of  Table  14.10 


Source  of 
Variation 

Degrees 
of 
Free 
dom 

Sum  of  Squares  and  Products 

Deviations  About  Regression 

2>      I>3>      Z?2 

(2>y)2 

Y%* 

Degrees 

of  Free 
dom 

Mean 
Square 

-    v 

Replicates  (pens)  .  . 
Treatments 
Food 

4 

2 
1 
2 
20 

605.87       39.905     4.8518 

5,40     -0.147     2.2686 
32.03     -3.730     0.4344 
22.47         3.112     0.4761 
442.93       39.367     8.3144 

Sex 

Food  X  sex 

Experimental  error 

4.8155 

19 

0.2534 

Food  ~  f~  error 

22 

448.33       39.220  10.5830 

7.1520 

21 

Difference  for  testing  among 

adjusted  food  means  

2.3365 

2 

1  .  16825 

Sex-j-error  

21 

474.96       35.637     8.7488 

6.0749 

20 

Difference  for  testing  among 

adjusted  sex  means 

1.2594 

1 

1.2594 

(Food  X  sex) 
-f-  (error)  .  

22 

465.40       42.479     8.7905 

4.9133 

21 

Difference  for  testing  among  adjusted  foodXsex  effects  .  .  . 

0.0978 

2 

0.0489 

cance,  the  appropriate  variance  ratios  being 

1.16825 


Food:  F  = 


Sex: 


Food  X  Sex:     F 


0.2534 
1.2594 
0.2534  : 
0.0489 
0.2534  : 


4.61 


•  4.97 


0.19 


where  the  degrees  of  freedom  are  as  given  in  Table  14.11.  These  F-ratios 
(and  the  corresponding  inferences)  should  be  compared  with  those 
resulting  from  an  analysis  of  variance  on  the  gains  in  weight  taking  no 
account  of  the  varying  initial  weights.  Such  comparisons  will  aid  the 
reader  in  understanding  the  principles  of  covariance  analyses  and,  in 
our  example,  will  help  to  explain  the  effect  of  initial  weights  on  weight 
gains  subject  to  the  chosen  experimental  conditions.  A  table  of  adjusted 
treatment  means,  together  with  the  appropriate  standard  errors, 
should  also  be  presented  to  make  the  analysis  complete. 

COVARIANCE  WHEN   THE  X  VARIABLE   IS 
AFFECTED  BY  THE  TREATMENTS 

When  the  treatments  being  employed  in  the  experiment  are  such 
that  they  have  an  appreciable  effect  on  the  concomitant  variate,  X, 
as  well  as  on  the  Y  variate,  the  researcher  should  proceed  with  caution. 
Computationally,  each  step  is  carried  through  as  before,  but  the  final 


14.7 


14.8      MULTIPLE    COVARIANCE  457 

inferences  must  take  account  of  the  effect  which  the  treatments  have 
had  on  the  concomitant  variate.  For  example,  if  the  concomitant 
variate  in  a  feeding  experiment  had  been  "amount  of  feed  consumed" 
rather  than  the  "initial  weight/'  it  is  quite  possible  that  the  different 
treatments  (feeds)  "would  have  a  significant  effect  on  the  food  con 
sumption.  Thus  any  covariance  analysis  of  gain  in  weight  should  take 
cognizance  of  the  weight-producing  effects  of  the  different  feeds  due  to 
increased  (or  decreased)  consumption  apart  from  any  nutritional  dif 
ferences  among  the  feeds.  Other  examples  of  covariance  analyses  in 
volving  similar  problems  are  available  in  the  literature  and  should  be 
studied  critically  by  those  who  desire  a  fuller  understanding  of  covari 
ance  techniques.  The  reader  is  especially  referred  to  Cochran  and 
Cox  (5)  and  Bartlett  (2)  for  a  discussion  of  this  type  of  problem. 

14.8      MULTIPLE  COVARIANCE 

As  might  be  anticipated,  procedures  are  available  for  performing 
covariance  analyses  when  data  are  collected  on  two  or  more  concomit 
ant  variables.  These  procedures  will  follow  the  same  pattern  that  has 
been  used  in  the  preceding  sections  of  this  chapter,  the  only  change 
being  that  the  sum  of  squares  due  to  regression  is  calculated  in  accord 
ance  witli  the  principles  outlined  in  Section  8.15.  Rather  than  specify 
an  analytical  approach  for  the  general  case,  let  us  be  content  with  a 
numerical  example  to  illustrate  the  ideas  involved. 

Example  14.4 

Crampton  and  Hopkins  (8)  studied  the  effects  of  initial  weight  and 
food  consumption  on  the  gaining  ability  of  pigs  when  given  different 
feeds.  The  data  are  presented  in  Table  14.12. 

To  carry  out  a  multiple  covariance  analysis,  the  first  step  is  to  find  the 
various  sums  of  squares  and  products  for  treatments  and  for  error  and, 
hence,  for  "treatments+error."  These  are  determined  to  be 

^2x2  =  28,404.9 
Sr,*,  =  90,792.3 
>*,*,  «  119,197.2 

^x,  —  2187.8 
£,„,  =  264<5.2 


T*i*i-  509.2 

£*i*i=  368.4 

Tyy    =            5741.7 
Eyy    =     10,405.5 

&I!BI  =  877.6 

TXlV  =  1172.2 
JSclV  =  1001.8 

Syy    -     16,147.2 

TXlky  =  11,596.5 
E*rt  =  24,508.7 

2173.8       SXjty  =  36,105.2        S^x*  —  4834.0 

where 

Y  =  final  weight 
X±  —  initial  weight 
Xz  —  feed  eaten. 

The  next  step  is  to  calculate  the  partial  regression  coefficients  as 
sociated  with  the  multiple  regression  equation  so  that  we  may  obtain 
the  sum  of  squares  "due  to  regression"  and  thus,  by  subtraction  from 
the  corrected  total  sum  of  squares,  the  sum  of  squares  of  the  deviations 
about  regression.  The  required  partial  regression  coefficients  may  be 


5" 

ll 

£ 

^<Qoe5S^°SS)^ 

CO 

t^ 

•I'| 

^•Ncs^^-es-.^-. 

s 

^ 

R 

^ 

fe: 

^ 

.| 

£ 

Oco^Sooo^M^ 

u^ 

b 

g 

o 

^ 

^ 

1 

00 

§ 

|l 

*"5 

^SBSssSs^S 

0 

i—  r 
i—  i 

^ 

*"• 

ti 

3 

^ 

CO 

£ 

it 

^ 

3gS33$383S 

S3 

Ht 

be 

M     1 

rt 

Q 

g 

a\ 

•rf 

"*§ 

1.1 

fc^ 

§l^s§§SilS 

cs 

g 

0) 

pD     > 

W) 

fe 

CS 

c 

^3 

native 

i—  i 

i 

O 

^ 

«§Sg^H52§ 

S 

itive  Fee 

CO 

CU 

S 

I 

11 

^ 

SggSisSISi 

§ 

1 

o 

1 

u 

11 

- 

^tccs^cst^ooo^vo 

« 

O 

.S 

•a  's 

^N 

*o 

.J9 

ff 

t/D 

fbO 

'S  ^ 

N. 

^ 

f^ 

.1  .g 

£ 

§§^B^S§SS§ 

ON 

c 

v 

*s 

ON 

•5 

£ 

H 

,s 

^ 

•*vocNOOr*5OOesesio 

•rH 

.S 

.2 

M 

Cj 

N 

vO^oor^u^O'Vor—   -^*io 

NO 

rt 

"c 

o 

.2 

ronsump 

Treatme 

11 

" 

SSSiiSSiiS 

S 

rtial  Regress 

o 

i—  i  ^ 

^' 

%*sss"°a"sq 

ss 

«s 

P-i 

*o 

4) 

4j 

00 

"S 

T-j 

P 

,~ 

Sooo^SoJoo- 

§ 

•5 

o 

0* 

»o 

S  e 
|l 

.§ 

HH 
1—  I 

1 

£ 

"SSSSSSSSS 

r- 

G 

O 

*H 

O     3 

o 

u 
1 

_-  c 

vo 

*£ 

IS* 

CJ    _OJ 

* 

illllsSSIl 

00* 
t— 
vO 

S3   c 

f  S 

'S 

"53 

11 

^ 

aasjsaasssa 

i 

Jfl 

M  3 
&  « 

W3 

(N 

T—  1 

1—  1   *•* 

*°. 

^'^ 

.  fl 

^T 

a    W 

^t 

ONr-OOO^i^i^oooNOO 

00 

1—  v  tO 

i—l 

S    fe 

^H     *H      »-l      CS      i-t                 ^(»-(^H^-t 

vH 

•XJ  "S 

w 

"0 

!l 

rVN. 

I-H 

'I 

£ 

^inSSS^^^^^SSo 

VO 

IS. 

HH 

O 

O 

a   * 

•^ 

S 

i  n 

EH 

i 

"2  1 

M 

t-esvOOv^oOt->ro^ro 
\OvOvOvot^-iOiCvOvOvO 

i 

0^ 

sC- 

.    r^ 

"3  5 

ON 

M  g 

.-J 

O^H^Hfor-.-^OO>OOO 

iO 

6  "• 

"S  "^ 

^N 

rcesc^f^cscscstscscs 

N 

g    vT 

M     ^ 

,h!    aj 

•y-e 

CJ 

CO 

§1 

| 

^CSro^^V5^00C>0 

I 

•5      PH 
1^ 

fS 

10 

> 

14.8       MULTIPLE    COVARlANCE 


459 


found  using  the  methods  of  Chapter  8.  Since  we  are  dealing  with  a 
case  involving  only  two  independent  variables,  it  is  simpler  to  solve  the 
following  equations: 

Error 

36SAbuE  +      2646.2&2J?  =      1001.8 
2646.2&1J?  +  90,792.3^2^7  =  24,508,7 


Treatment  +  error 
87 
4834.0&UJP+JB)  4-  119,197.2&2<T+*> 


2173.8 
36,105.2 


giving 


.9868, 


=  .2412, 


«  1.0410  and 


.2607. 


Thus,  the  sums  of  squares  due  to  regression  (and,  hence,  the  sum  of 
squares  of  the  deviations  about  regression)  for  "error"  and  for  "treat- 
ment  +  error,"  respectively,  are  determined  to  be: 

Error 

S.S.  due  to  regression  =  (.9868)  (1001.8)  +  (.2412)  (24,508.7) 

=  6900.07 
S.S.  of  deviations  about  regression  =»  10,405.5  —  6900.07 

=  3505.43 
Treatment  +  error 

S.S.  due  to  regression  =  (1.0410)  (2173.8)  +  (.2  607)  (3  6,  105.  2) 

=  11,675.45 

5\.S.  of  deviations  about  regression  =  16,147.2  —  11,675.45 

=  4471.75. 

These  results  are  then  presented  in  Table  14.13.  Notice  that  this  time 
we  have  '  lost'7  two  degrees  of  freedom  rather  than  one  degree  of  freedom 
as  in  a  simple  covariance.  If  there  had  been  k  independent  (concomitant) 
variables,  we  would  have  "lost"  k  degrees  of  freedom.  The  reader  will 
note  that  this  is  in  agreement  with  the  procedures  outlined  in  Chapter  8. 
The  F-test  is  then  performed  as  before,  and  we  obtain 
F  =  241,  58/103.  10  =  2.  34  with  degrees  of  freedom  vi  =  4  and  i>2  =  34, 
which  is  not  significant  at  a.  =  .05.  This  should  be  compared  with  the 

TABLE   14.  13-  Abbreviated  Analysis  of  Covariance  for  Data  of  Table  14.12 


Source  of 
Variation 

Degrees 
of  Free 
dom 

I> 

S.S.  Due  to 
Regression 

S.S.  Devia 
tions  About 
Regression 

Degrees 
of  Free 
dom 

Mean 
Square 

Treatments  (T)  .  . 
Error  CE)  

4 
36 

5741.7 
10,405.5 

6900.07 

3505.43 

34 

103  .  10 

T-f-72 

40 

16,147.2 

11,675.45 

4471.75 
966.32 

38 
4 

241.58 

Difference  for  testing  among  adjusted  treatment  means  .  . 

460 


CHAPTER    14,   ANALYSIS    OF   COVARlANCC 


results  obtainable  from  an  analysis  of  variance  of  the  final  weights,  and 
the  appropriate  conclusions  drawn.  Adjusted  treatment  means  and  their 
standard  errors  may  be  calculated  by  combining  methods  indicated  in 
Section  8.25  and  earlier  sections  of  this  chapter. 

Problems 

14,1        An  experiment  using  a  randomized  complete  block  design  gave  the 
following  corrected  sums  of  squares  and  products: 


Degrees 

of  Free 

Source  of  Variation 

dom 

2>2 

Z>:v 

Z?2 

b 

Replicates 

5 

200 

600 

4000 

Treatments  

5 

100 

200 

2500 

2 

Experimental  error.  .  . 

25 

300 

1200 

7500 

4 

(a)   Based  on  the  experimental  error  sum  of  squares  and  products, 

is  the  regression  of  Y  on  X  significant  at  a:  =  .05? 
(6)    Are  the  differences  among  the  treatment  means  for  Y  adjusted  for 

variation  attributed  to  X  significant  at  OL  =  .05? 
(c)    What  conclusions  do  you  draw  from  the  above  data  about  the 

effects  of  treatments?  Make  any  additional  computations  that 

you  consider  necessary. 
14.2        Given  the  following  data: 


Source  of  Variation 

Degrees  of 
Freedom 

X>2 

Z>;y 

i:?2 

Replicates  

4 

100 

140 

400 

Xreatments  

10 

100 

100 

900 

Experimental  error  

40 

400 

900 

2500 

(a)   What  conclusions  may  be  drawn  about  the  effect  of  treatments 

on  F? 
(6)    Test  the  regression  coefficient  based  on  experimental  error  for 

significance  at  the  5  per  cent  level. 

14.3  Ten  lines  of  soybeans  were  compared  in  randomized  complete  blocks 
with  4  replications.  The  differences  in  yield,  Y,  were  not  significant, 
but  it  was  observed  that  the  incidence  of  an  infestation,  X,  differed 
among  the  varieties.  Following  is  the  table  of  sums  of  squares  and 
products: 


Source 

Degrees 

of 

of  Free 

yjlx* 

y^xv 

>  ^-y2 

Variation 

dom 

Lines  

9 

4684 

—  532 

112 

Error  

27 

3317 

—  65O 

216 

PROBLEMS 


461 


Test  the  hypothesis  that  the  yields  adjusted  for  infestation  do  not 
differ  in  the  sampled  populations.  What  fraction  of  X)2/2  f°r  lines  is 
unexplained  by  the  regression? 

j.4.4  For  an  experiment  involving  9  soil  sterilization  treatments,  the  effect 
on  the  number  of  seedling  alfalfa  plants  (X)  and  on  the  green  weight 
of  plants  at  3  weeks  (F)  is  summarized  by  the  following  sums  of 
squares  and  products: 


Source  of 

Degrees 
of  Free 

Variation 

dom 

!>2 

T,*y 

Z?2 

Replicates 

5 

4 

16 

96 

Treatments  

8 

16 

32 

80 

Error  

40 

20 

40 

160 

Complete  the  analysis,  making  appropriate  tests  to  indicate  the 
reason  for  your  conclusions.  If  the  mean  of  the  X9 s  is  15  and  the 
mean  of  the  Y's  is  25,  give  the  regression  equation  for  error. 
14.5  A  study  of  eastern  Iowa  farms  included  one  group  of  tenants  who 
were  not  related  to  their  landlords,  and  another  group  of  tenants, 
each  of  whom  was  related  to  his  landlord.  It  was  assumed  that  soil 
improvements  would  be  more  generally  undertaken  when  landlord 
and  tenant  were  related.  Hence,  value  of  crops  should  be  greater  in 
those  situations.  An  analysis  of  variance  was  undertaken  to  examine 
this  hypothesis.  Since  size  of  farm  could  confuse  the  comparison,  the 
size  of  farm  was  introduced  as  a  covariate.  The  following  table  was 
prepared : 

COVARIANCE  ANALYSIS  or  VALUE  or  CROPS  ON  SIZE  OP  FARM  TOR 
EASTERN  IOWA  FARMS  WITH  LANDLORD  AND  TENANT  RELATED 
AND  LANDLORD  AND  TENANT  NOT  RELATED 


Source  of  Variation 

Degrees 
of  Free 
dom 

I>* 

JLxy 

I>* 

Total  

59 

125000 

33000 

36600 

Sub-areas  (replicates)  

4 

20000 

14010 

13600 

Bet.  grps   of  tenants  

1 

61000 

13260 

4200 

Interaction  , 

4 

4000 

—  13270 

6100 

"Within  subclasses  

50 

40000 

19000 

12700 

14.6 


Value  of  crops  has  been  coded  for  this  analysis. 

(a)   Is  the  acceptance  of  the  hypothesis,  H  (no  difference  between 

groups  of  tenants)  changed  by  the  introduction  of  farm  size  as  the 

covariate? 

(6)    Is  the  error  regression  significant? 
A  sample  of  farms  was  taken  in  the  eastern  livestock  area  of  Iowa  for 


462 


CHAPTER    14,   ANALYSIS    OF    COVARIANCE 


the  purpose  of  studying  certain  types  of  farm  lease  arrangements. 
For  this  problem  we  are  taking  a  portion  of  the  data  to  study  the  dif 
ference  in  "gross  value  of  crops"  produced  on  two  groups  of  cash- 
rented  farms:  (1)  farms  for  which  landlord  and  tenant  are  related, 
and  (2)  farms  for  which  landlord  and  tenant  are  not  related.  The 
variates  measured  are  value  of  crops  produced  (F)  and  size  of  farm 
(X).  The  data  presented  are  given  for  3  hypothetical  blocks.  In 
practice,  these  blocks  might  be  strata,  e.g.,  different  counties,  differ 
ent  soil  areas,  type-of-farming  areas,  or  groups  of  farms  enumerated 
by  different  enumerators,  for  instance,  3  agricultural  economics  stu 
dents  for  the  3  blocks  of  our  example.  The  data  are  presented  in  the 
table  following.  Perform  an  analysis  of  covariance  on  these  data. 

FARM  DATA  FROM  EASTERN  LIVESTOCK  AREA,  IOWA,  FOR  COVARIANCE 

ANALYSIS  OF  VALUE  OF  CROPS  AND  SIZE  OF  FARM  GROUPS  To  BE 

COMPARED:  FARMS  WITH  LANDLORD  AND  TENANT  RELATED 

AND  LANDLORD  AND  TENANT  NOT  RELATED* 


Block  I 

Block  II 

Block  III 

Farm 

Y 

X 

Farm 

Y 

X 

Farm 

Y 

X 

No. 

Related 

No. 

Related 

No. 

Related 

22 

6399 

160 

27 

2490 

90 

17 

4489 

120 

13 

8456 

320 

24 

5349 

154 

25 

10026 

245 

20 

8453 

200 

11 

5518 

160 

1 

5659 

160 

8 

4891 

160 

34 

10417 

234 

26 

5475 

160 

21 

3491 

120 

38 

4278 

120 

4 

11382 

320 

Not  Related 

Not 

Related 

Not 

Related 

31 

6944 

160 

13 

4936 

160 

20 

5731 

160 

30 

6971 

160 

1 

7376 

200 

15 

6787 

173 

11 

4053 

120 

19 

6216 

160 

7 

5814 

134 

6 

8767 

280 

32 

10313 

240 

5 

9607 

239 

16 

6765 

160 

28 

5124 

120 

25 

9817 

320 

*  Source:  Agricultural  Economics  Dept.,  Iowa  State  College,  1951. 

14.7  (a)    What  are  the  assumptions  behind  a  covariance  analysis? 

(6)    In  the  process  of  analyzing  data  by  a  covariance  analysis,  what 

tests  of  significance  are  made? 
(c)    Explain  the  interpretation  or  inferences  and  the  course  of  action 

indicated  when  each  of  the  above  tests  is  significant;  when  each 

is  nonsignificant. 

14.8  The  following  is  an  experiment  involving  randomized  complete  blocks 
with  4  replications.  Eleven  lines  of  soybeans  were  planted.  The  data 
are  as  follows: 

X i  =  maturity,  measured  in  days  later  than  the  Hawkeye  variety 
Xz  =  lodging,  measured  on  a  scale  from  0  to  5 

Y   =  infection  by  stem,  canker  measured  as  a  percentage  of  stalks 
infected. 


PROBLEMS 


463 


Replicate  1 

Replicate  2 

Replicate  3 

Replicate  4 

Line 

JSTi      X^         Y 

Xi      X*         Y 

JSTi      X2         Y 

-Yi      X2         F 

Lincoln    

9     3.0     19  3 

10     20     29.2 

12     3  0       10 

925        6.4 

A7-6102    

10     3.0     10.1 

10     20     34.7 

9     20      14  0 

930       5.6 

A7-6323    .... 

10     25      13   1 

9      15     59  3 

12     2  5        11 

10     25       8.1 

A7-6520   

8     2.0     15   6 

5     20     49  0 

8     20     17  4 

6     20     11   7 

A7-6905  

12     2.5       4.3 

11      10     48.2 

13     3  0       63 

10     2  5       67 

C-739            ...    . 

4     2.0     25   2 

2      15     36  5 

2     20     23  4 

1     20     12  9 

C-776      

3      1.5     67.6 

4     10     79  3 

6     2  0     13   6 

2      15     39  4 

H-6150  . 

7     2.0     35   1 

8     2  0     4O  0 

7     20     24  7 

720       48 

L6-8477 

8     20      14  0 

8      15     30  2 

10     1   5        72 

720       89 

L7-1287 

925        3.3 

9     20     35  8 

13     3  0       11 

930       20 

BEIV   Sp 

10     3   5        31 

10     3  0       96 

11     3  0        10 

10     3  5       01 

The  principal  objective  is  to  learn  whether  maturity  or  lodging  is 
more  closely  related  to  infection.  Determine  this  from  the  error 
multiple  regression.  Test  the  hypothesis  of  no  differences  among 
adjusted  mean  infection  for  the  varieties. 

14.9  Discuss  the  use  of  covarianee  analysis.  What  factors  must  be  con 
sidered  in  interpreting  the  results  of  the  analysis? 

14.10  The  data  for  this  problem  consist  of  54  pairs  of  observations  on  the 
calories  consumed  (Y)  on  one  particular  day  by  a  respondent,  and  her 
age  (X) .  The  respondents  were  adult  Iowa  women  over  the  age  of  30 
who  were  interviewed  to  obtain  information  on  nutrition  and  health. 
About   1000   women  were   so   enumerated  for  this  survey,    and  our 
group  of  54  is  a  subgroup  from  the  total,  which  was  taken  so  as  to 
make  numbers  in  the  subclasses  equal. 

Among  the  items  observed  for  each  respondent  in  addition  to 
caloric  intake  and  age  was  place  of  residence  (zone)  and  income  class. 
These  are  listed  as 


Zone  1 — open  country 
Zone  2 — rural  place 
Zone  3 — urban 


Income  Group 


1 
2 
3 
4 
5 
6 


0-$  999 
1000-  1499 
1500-  1999 
2000-  2999 
3000-  3999 
over  4000 


Education,  height,  weight,  national  origin,  marital  status,  family 
composition,  and  many  other  factors  were  recorded  for  each  re 
spondent. 

The  nutritionists  studying  these  data  are  interested  in  determining 
how  food  intake  and  health  are  related  to  these  other  observed  fac 
tors.  A  few  relevant  hypotheses  could  be  advanced.  Preliminary 
analysis  consisted  of  preparing  tables  of  means  for  several  classifica 
tions  of  the  total  sample  and  graphical  analysis  (plotting  on  scatter- 
grams  of  a  subsample  of  60  stratified  by  age).  A  number  of  nutritive 
factors  exhibited  an  apparent  negative  regression  on  age.  Age  thus 
seemed  a  useful  covariate.  Other  factors,  such  as  education,  height, 
and  weight,  seemed  to  indicate  no  relation  to  nutritive  intake. 


464 


CHAPTER    14,   ANALYSIS    OF    CO  VARIANCE 


With  this  background  we  shall  use  these  data  to  undertake  an 
analysis  of  covariance  for  the  purpose  of  testing  hypotheses  about 
zone  and  income  group  effects  after  taking  account  of  the  regression 
on  age.  The  data  table  gives  the  54  pairs  of  observations  with  the 
sums,  sums  of  squares,  and  sums  of  products.  Both  zone  and  income 
group  may  be  considered  as  fixed  effects, 
(a)    Prepare  the  analysis  of  covariance  table. 
(6)    Find  the  error  regression  of  calories  on  age. 

(c)  Test  the  hypotheses  that  zone  and  income  effects  are  =  0  (sepa 
rately,  of  course)* 

(d)  It  would  also  be  of  interest  to  test  for  interaction  of  zone  and  in 
come  group.  What  do  you  conclude  on  this  point? 

(e)  The  regression  of  calories  on  age  may  not  be  homogeneous  over 
the  zones.  Indicate  by  a  schematic  analysis  of  variance  of  regres 
sion  how  you  would  examine  these  regressions. 


Zone 

1 

Zone 

2 

Zone 

3 

Income  Group 

Y 

X 

F 

X 

F 

X 

1  

1911 

46 

1318 

80 

1127 

74 

1560 

66 

1541 

67 

1509 

71 

2639 

38 

1350 

73 

1756 

60 

2  

1034 

50 

1559 

58 

1054 

83 

2096 

33 

1260 

74 

2238 

47 

1356 

44 

1772 

44 

1599 

71 

3  

2130 

35 

2027 

32 

1479 

56 

1878 

45 

1414 

51 

1837 

40 

1152 

59 

1526 

34 

1437 

66 

4  

1297 

68 

1938 

33 

2136 

31 

2093 

43 

1551 

40 

1765 

56 

2035 

59 

1450 

39 

1056 

70 

5  

2189 

33 

1183 

54 

1156 

47 

2078 

36 

1967 

36 

2660 

43 

1905 

38 

1452 

53 

1474 

50 

6  

1156 

57 

2599 

35 

1015 

63 

1809 

52 

2355 

64 

2555 

34 

1997 

44 

1932 

79 

1436 

54 

166416926 
—  CT     156053200 


-CT 


=  10363726 


4573454 
4773496 

—  200042 


—  CT 


157356 
146016 

*=  11340 


REFERENCES    AND    FURTHER    READING  465 

References  and  Further  Reading 

1.  Anderson,  R.  L,.,  and  Bancroft,  T.  A.  Statistical  Theory  in  Research.  McGraw- 
Hill  Book  Company,  Inc.,  New  York,  1952. 

2.  Bartlett,  M.  S.  A  note  on  the  analysis  of  covariance.  Jour.  Agr.  Sci.,  26:448, 
1936. 

3.  Cochran,   W.   G.   Analysis  of   covariance.   Mimeo  Series  No.  6,   Institute  of 
Statistics,  University  of  North.  Carolina,  Chapel  Hill,  1949. 

4    u   Analysis   of   covariance:   its   nature   and  uses.   Biometrics,    13    (No. 

3):261-81,  Sept.,  1957. 

5.   ,  and  Cox,  G.  M.  Experimental  Designs.  Second  Ed.  John  Wiley  and 

Sons,  Inc.,   New  York,    1957. 

6.  Coons,   Irma.  The  analysis  of  covariance  as  a  missing  plot  technique.  Bio 
metrics,  13  (No.  3)  :387-405,  Sept.,  1957. 

7.  Cox,  D.  R.  Planning  of  Experiments.  John  Wiley  and  Sons,  Inc.,  New  York, 
1958. 

8.  Crampton,   E.   W.,   and   Hopkins,   J.   W.   The  use  of  the   method  of  partial 
regression  in  the  analysis  of  comparative  feeding  trial  data.  Part  II.  Jour. 
Nutrition,  8:329,   1934. 

9.  Federer,  W.  T.  Experimental  Design.  Macmillan  Co.,  New  York,  1955. 

10. Variance    and    covariance    analyses    for    unbalanced    classifications. 

Biometrics,  13  (No.  3)  :333-62,  Sept.,  1957. 

11.  Finney,    D.  J.   Stratification,   balance,   and  covariance.   Biometrics,    13    (No. 
3):373-86,  Sept.,  1957. 

12.  Kempthorne,   O.   The  Design  and  Analysis  of  Experiments.  John  Wiley  and 
Sons,  Inc.,  New  York,   1952. 

13.  Quenouille,   M.  H.   The  Design  and  Analysis  of  Experiment.  Charles  Griffin 
and  Co.,  Ltd.,  London,  1953. 

14.  Smith,     H.     Fairfield.     Interpretation     of    adjusted    treatment    means     and 
regressions  in  analysis  of  covariance.  Biometrics,  13  (No.  3)  :282— 308,  Sept., 
1957. 

15.  Truitt,  J.  T.,  and  Smith,  H.  Fairfield.  Adjustment  by  covariance  and  conse 
quent  tests  of  significance  in  split-plot  experiments.  Biometrics,  12   (No.  1): 
23-39,  Mar.,  1956. 

16.  Wilkinson,    G.    N.    The   analysis   of   covariance   with  incomplete   data.   Bio 
metrics,  13  (No.  3):363-72,  Sept.,  1957. 

17.  Wishart,  J.  Field  trials  II:  The  analysis  of  covariance.   Tech.   Comm.  No.  15, 
Commonwealth  Bureau  of  Plant  Breeding  and  Genetics,  Cambridge,   Eng 
land,  May,  1950. 

18.  Zelen,    M.    The   analysis   of   covariance  for  incomplete   block   designs.   Bio 
metrics,  13  (No.  3):309-32,  Sept.,  1957, 


CH  APTE  R    15 

DISTRIBUTION-FREE  METHODS 

IN  PRECEDING  CHAPTERS  the  emphasis  has  been  on  those  statistical 
techniques  which  assume  the  sampled  populations  to  be  of  known  form. 
However,  because  the  analyst  is  not  always  certain  of  the  validity  of 
such  assumptions  and/or  because  not  all  statistical  techniques  are 
robust  (i.e.,  insensitive  to  departures  from  such  assumptions),  much 
work  has  been  done  in  recent  years  to  devise  procedures  which  are  free 
of  these  restrictions.  These  new  techniques,  referred  to  as  distribution- 
free  methods -,1  will  be  the  subject  of  this  chapter. 

Although  most  distribution-free  methods  have  been  developed  only 
recently,  the  literature  in  this  area  is  already  quite  extensive.  Further, 
it  continues  to  grow  every  day.  Consequently,  it  will  be  impossible  to 
do  more  than  mention  a  few  of  the  more  popular  and  useful  methods 
in  this  book.  Those  persons  who  wish  to  delve  deeper  into  this  area  of 
statistics  are  encouraged  to  consult  the  references  listed  at  the  end  of 
this  chapter. 

15.1  DISTRIBUTION-FREE  METHODS   INCLUDED    IN 
PREVIOUS  CHAPTERS 

Four  widely  used  distribution-free  methods  have  already  been  intro 
duced  in  earlier  chapters.  These  are:  (1)  Tchebycheff's  inequality  dis 
cussed  in  Section  5.3,  (2)  the  distribution-free  tolerance  limits  referred 
to  in  Section  6.13,  (3)  the  chi-square  goodness  of  fit  test  described  in 
Section  7.15,  and  (4)  the  measures  of  rank  correlation  described  in 
Section  9.11.  Because  these  methods  were  discussed  in  the  afore 
mentioned  sections,  it  would  be  superfluous  to  repeat  their  descriptions 
at  this  time.  It  is  recommended,  however,  that  the  indicated  sections 
be  reread  in  the  present  context.  Let  us  now  proceed  to  the  study  of 
some  additional  distribution-free  methods  that  have  been  found  useful 
in  a  variety  of  situations. 

15.2  THE  SIGN  TEST 

In  many  experimental  situations,  the  investigator  wishes  to  compare 
the  effects  of  two  treatments.  When  the  data  occur  in  pairs,  one  mem 
ber  of  the  pair  being  associated  with  treatment  A  and  the  other  with 
treatment  B,  one  test  of  wide  applicability  is  the  sign  test.  Using  in 
equality  signs  to  denote  the  relationship  between  the  members  of  a 

1  Many  authors  refer  to  distribution-free  methods  as  nonparametric  methods 
and,  although  the  expressions  are  not  strictly  equivalent,  they  have  been,  and 
probably  will  continue  to  be,  used  interchangeably. 

[466] 


15.2       THE    SIGN    TEST  467 

pair,  whether  the  comparison  be  qualitative  or  quantitative.2  the  sign 
test  proceeds  as  follows: 

(1)  Examine  each  of  the  pairs  (Xiy  Y£. 

(2)  If  Xi>  Yi?  assign  a  plus  sign;  if  X* <  Yif  assign  a  minus  sign;  if 
Xi  =  Yt,  discard  the  pair. 

(3)  Denote  the  number  of  pairs  remaining,  that  is,  the  number  of 
pairs  resulting  in  either  a  plus  or  minus  sign,  by  n. 

(4)  Denote  by  r  the  number  of  times  the  less  frequent  sign  occurs. 

(5)  To  test  the  hypothesis  of  no  difference  between  the  effects  of 
the  two  treatments,  compare  r  with  the  critical  values  tabu 
lated  in  Appendix  13. 

(6)  If  the  observed  value  of  r  is  less  than  or  equal  to  the  tabulated 
value  for  the  chosen  significance  level,  the  hypothesis  is  re 
jected;  otherwise,  it  is  not  rejected. 

Before  giving  numerical  illustrations,  it  is  appropriate  that  attention 
be  called  to  different  hypotheses  that  can  be  tested  in  the  manner 
indicated  above.  Some  of  these  are: 

(1)  Each  difference  X*  —  F;,  has  a  probability  distribution  (which 
need  not  be  the  same  for  all  differences)  with  median  equal  to 
0,  that  is,  H:P{Xi>  Y,}  =  0.5  for  all  i. 

(2)  If  the  underlying  distributions  are  assumed  to  be  symmetric, 
the  sign  test  may  be  used  to  test  the  hypothesis  flr:Mjr,  =  My./. 

(3)  If  it  can  be  assumed  that  the  underlying  distributions'  differ 
only  in  their  means,  then  a  test  of  Hip,x,=pY.  is  equivalent 
to  testing  the  hypothesis  that  the  probability  distributions  of 
each  pair  are  the  same. 

(4)  Questions  such  as: 

(a)  Is  A  better  than  B  by  P  per  cent? 
and 

(b)  Is  A  better  than  B  by  U  units? 

may  also  be  studied  by  applying  the  sign  test  to  the  differences 
D  =  A  -  (l+P/100) B  and  D  =  A  —  (B+  C7),  respectively. 

Example  15.1 

Consider  once  again  the  data  given  in  Table  7.6  and  discussed  in 
Example  7.21.  We  note  that  ™  =  15  and  r  =  4.  Assuming  <*  =  0.05,  it  is 
seen  that  the  hypothesis  of  equal  hardness  indications  by  the  two'  steel 
balls  cannot  be  rejected  since  the  critical  value  of  r  tabulated  in  Appen 
dix  13  was  3.  You  will  note  that  this  is  the  opposite  decision  to  that 
reached  in  Example  7,21.  The  reason  for  this  is  that,  when  normality 
can  be  assumed,  the  sign  test  is  less  efficient  (that  is,  less  sensitive)  than 
"Student's"  J-test. 

2  If  measurements  are  recorded,  then  X<  >  Yt  will  signify  that,  in  the  ith  pair, 
treatment  A  resulted  in  a  higher  reading  than  treatment  B.  If  no  measurements 
are  available,  then  Xi  >  Yi  will  signify  that,  in  the  fcth  pair,  treatment  A  resulted 
in  something  larger  than  Cor  better  than  or  preferred  over)  treatment  B. 


468  CHAPTER    15,    DISTRIBUTION-FREE    METHODS 

Example  15.2 

In  a  marketing  study,  two  brands  of  lemonade  were  compared.  Each 
of  50  judges  tasted  two  samples,  one  of  brand  A  and  one  of  brand  B, 
with  the  following  results:  35  preferred  brand  A,  10  preferred  brand  J3, 
and  5  could  not  tell  the  difference.  Thus,  n  =  45  and  r  =  10.  Assuming 
CK=0.01,  we  reject  the  hypothesis  of  equal  preference  (since  r  =  10  <13 
=  critical  value)  and  conclude  that  brand  A  is  preferred. 

15.3     THE  SIGNED   RANK  TEST 

The  sign  test  described  in  Section  15.2  was  simple  to  apply.  However, 
when  measurement  data  have  been  obtained,  it  is  not  the  most  efficient 
distribution-free  test  available.  A  better  test,  sometimes  referred  to  as 
the  Wilcoxon  signed  rank  test  and  at  other  times  more  simply  as  the 
signed  rank  test,  is  one  which  takes  account  of  the  magnitude  of  the 
observed  differences.  It  proceeds  as  follows: 

(1)  Rank  the  differences  without  regard  to  sign,  that  is,  rank  the 
absolute  values  of  the  differences.  (The  smallest  difference  is 
given  rank  1  and  ties  are  assigned  average  ranks.) 

(2)  Assign  to  each  rank  the  sign  of  the  observed  difference. 

(3)  Obtain  the  sum  of  the  negative  ranks  and  the  sum  of  the 
positive  ranks. 

(4)  Denote  by  T  the  absolute  value  of  the  smaller  of  the  two 
sums  of  ranks  found  in  the  previous  step. 

(5)  To  test  the  hypothesis  of  no  difference  between  the  effects  of 
the  two  treatments,  compare  T  with  the  critical  values  tabu 
lated  in  Appendix  14. 

(6)  If  the  observed  value  of  T  is  less  than  or  equal  to  the  tabulated 
value  for  the  chosen  significance  level,  the  hypothesis  is  re 
jected;  otherwise,  it  is  not  rejected. 

Before  giving  numerical  illustrations,  we  should  note  that  the  signed 
rank  test  is  also  applicable  in  the  following  situations : 

(1)  To  test  the  hypothesis  that  the  median  of  a  population  is 
equal  to  some  specified  value,  say  M0. 

(2)  To  test  the  hypothesis  that  the  median  of  a  population  of  dif 
ferences  is  equal  to  some  specified  value,  say  M0. 

It  should  be  clear,  of  course,  that  in  Case  1  the  basic  variable  is 
\X  —  MO\,  while  in  Case  2  it  is  \(X—Y)—MQ\.  Apart  from  this 
obvious  transformation,  the  procedure  is  exactly  as  specified  above. 

Example  15.3 

Consider  again  the  data  of  Table  7.6.  These  are  reproduced  in  Table 
15.1  for  your  convenience.  Applying  the  procedure  for  the  signed  rank 
test,  it  is  seen  that  77  =  18.5.  Assuming  a:  =  0.05  and  consulting  Appendix 
14,  it  may  be  verified  that  T=18.5  <Tc  =  25,  and  therefore,  the  hy 
pothesis  of  equal  treatment  effects  is  rejected.  The  reader  should  com 
pare  this  result  with  those  reached  in  Examples  7.21  and  15.1. 


15.3       THE    SIGNED    RANK   TEST 

TABLE  15.1-Data  Obtained  in  a  Brinell  Hardness  Test 


469 


Sample 
Number 

Differences 
(£>) 

Rank  of  |Z>| 

Signed  Rank 

Positive 

Negative 

1    

22 
2 
4 
12 
11 
15 
28 
—   5 
S 
4 
—    1 
—  10 
—   2 
25 
7 

13 
2.5 
4.5 
11 
10 
12 
15 
6 
8 
4.5 
1 
9 
2.5 
14 
7 

13 
2.5 
4.5 
11 
10 
12 
15 

2        .... 

3 

4     

5 

6 

7 

8   

—    6 

9 

8 
4.5 

10 

11 

—    1 
—    9 
-   2.5 

12 

13 

14 

14 
7 

15 

Total 

101.5 

—  18.5 

Example  15.4 

Given  the  data  of  Table  15.2,  test  the  hypothesis  that  the  population 
median  equals  12.  It  is  easily  seen  that  the  calculations,  also  shown  in 
Table  15.2  for  convenience,  lead  to  T=6  which,  for  n  =  8  and  oj=0.05, 
tells  us  we  are  unable  to  reject  the  stated  hypothesis. 


TABLE   15.2-Hypothetical  Data  To  Illustrate  the  Procedure 
of  the  Signed  Rank  Test 


"O  rt  -nTi-    s\f 

Signed 

Rank 

Ob  servations 

tx) 

X—  Mo 

jtvanK.  01 

\X-M*\ 

Positive 

Negative 

12   55 

0  55 

3 

3 

14  62 

2   62 

8 

8 

12   93 

0  93 

4 

4 

12  46 

0.46 

2 

2 

11   95 

—  0  05 

1 

—  1 

14  55 

2   55 

7 

7 

13   11 

1    11 

6 

6 

10  90 

—  1.10 

5 

—  5 

Total 

30 

—  6 

470  CHAPTER    15,    DISTRIBUTION-FREE    METHODS 

15.4     THE   RUN  TEST 

Among  other  things,  the  theory  of  runs  may  be  used  to  test  the 
following  two  hypotheses: 

(1)  The  observations  have  been  drawn  at  random  from  a  single 
population. 

(2)  Two  random  samples  come  from  populations  having  the  same 
distribution. 

Because  the  mathematics  of  the  theory  of  runs  is  quite  involved,  we 
shall  do  no  more  than  sketch  the  approach.  Those  persons  who  need  to 
know  the  details  are  advised  to  consult  the  references  at  the  end  of 
the  chapter. 

Case  1 

(a)  List  the  observations  in  the  order  in  which  they  were  obtained, 
that  is,  in  their  order  of  occurrence. 

(b)  Determine  the  sample  median. 

(c)  Denote  observations  below  the  median  by  minus  signs  and 
observations  above  the  median  by  plus  signs. 

(d)  Denote  the  number  of  minus  signs  by  HI  and  the  number  of 
plus  signs  by  n^ 

(e)  Count  the  number  of  runs3  and  denote  this  number  by  r. 

(f)  If  r  is  less  than  or  equal  to  the  critical  value  tabulated  in 
Appendix  15  (Table  1)  or  greater  than  or  equal  to  the  critical 
value  tabulated  in  Appendix  15  (Table  2),  the  hypothesis  is 
rejected  at  the  5  per  cent  significance  level. 

Case  2 

(a)  List  the  r&i+n2  observations  from  the  two  samples  in  order  of 
magnitude,  that  is,  arrange  them  in  one  sequence  according  to 
their  values. 

(b)  Denoting  observations  from  one  population  by  x's  and  obser 
vations  from  the  other  population  by  j/'s,  count  the  number 
of  runs. 

(c)  Denote  the  observed  number  of  runs  by  r. 

(d)  If  r  is  less  than  or  equal  to  the  critical  value  tabulated  in 
Appendix  15  (Table  1),  the  hypothesis  is  rejected  at  the  5  per 
cent  significance  level, 

Example  15.5 

Suppose  a  manufacturing  process  is  turning  out  washers,  and  the 
characteristic  of  interest  is  the  outside  diameter.  In  the  first  40  washers 
tested,  there  were  16  runs  above  and  below  the  sample  median.  Noting 
that  ni  =  n2  =  20,  we  refer  to  Appendix  15  and  find  that  rI  =  14<16 
<28==ru-.  Thus,  at  the  5  per  cent  significance  level  we  are  unable  to  re 
ject  the  hypothesis  that  the  40  observations  constitute  a  random  sample 
from  a  single  population. 

3  In  terms  of  our  symbols,  a  run  is  a  sequence  of  signs  of  the  same  kind  bounded 
by  signs  of  the  other  kind. 


15.5       THE    KOLMOGOROV-SMIRNOV   TEST 


471 


Example  15.6 

Consider  the  data  of  Table  15.3.  Listing  according  to  ranks,  we  have: 
B,  AAAAAA,  B,  A,  B,  AA,  BBB,  A,  BBB}  A,  B}  A.  That  is  we  have 
r  =  12  runs.  In  addition,  7^  =  12  and  72^  =  10.  Reference  to  Appendix  15 
(Table  1)  tells  us  that,  since  r=12>rc  =  7?  we  are  unable  to  reject  the 
hypothesis  that  the  two  random  samples  came  from  populations  having 
the  same  distribution.  In  other  words,  using  the  run  test  and  operating 
at  the  5  per  cent  significance  level,  we  are  unable  to  reject  the  hypothesis 
that  the  two  lines  are  producing  equivalent  product. 

TABLE  15.3-Outside    Diameters    of    Washers    Produced 

by  Two  Different  Production  Lines  (Figures  in 

Parentheses  Are  the  Ranks) 


Line  A 

Line  B 

1.63  (6) 

1.65  (8) 

1  .  68  (9) 

1.69  (10) 

1  .  59  (4) 

1.72  (13) 

1.64  (7) 

1.91  (21) 

1.70  (11) 

1.74  (14) 

1.58  (3) 

1.75  (15) 

1.62  (5) 

1.55  (1) 

1.71  (12) 

1.86  (17) 

1.57  (2) 

1.87  (18) 

1.84  (16) 

1.88  (19) 

1.90  (20) 

1.96  (22) 

15.5     THE   KOLMOGOROV-SMIRNOV  TEST  OF 
GOODNESS  OF   FIT 

An  alternative  to  the  chi~square  goodness  of  fit  test  described  in 
Section  7.15  is  provided  by  the  Kolmogorov-Smirnov  test  to  be  de 
scribed  here.  Since  the  Kolmogorov-Smirnov  test  is  more  powerful 
than  the  ehi-square  test,  its  use  is  to  be  encouraged.  It  proceeds  as 
f  ollows : 

(1)  Let  F(x)  be  the  completely  specified  theoretical  cumulative 
distribution  function  under  the  null  hypothesis. 

(2)  Let  Sn(x)  be  the  sample  c.d.f .  based  on  n  observations.  For  any 
observed  x,  Sn(x)  =k/n  where  k  is  the  number  of  observations 
less  than  or  equal  to  x. 

(3)  Determine  the  maximtim  deviation,  D,  defined  by 

D  =  max   [  F(x)  —  Sn(x)  \  . 

(4)  If,  for  the  chosen  significance  level,  the  observed  value  of  D  is 
greater   than    or    equal    to    the    critical    value   tabulated    in 
Appendix  16,  the  hypothesis  will  be  rejected. 


472 


CHAPTER    15,    DISTRIBUTION-FREE    METHODS 


Example  15.7 

To  illustrate  the  Kolmogorov-Smirnov  test  of  goodness  of  fit,  we  shall 
apply  It  to  the  data  of  Table  7.11.  These  data  are  reproduced  here  as 
Table  15.4.  To  test  the  hypothesis  that  the  data  constitute  a  random 
sample  from  a  Poisson  population  with  a  mean  of  10.44,  calculations 
are  carried  out  as  shown  in  Table  15.4.  The  values  of  F(x)  were  found 
by  consulting  Appendix  2  and  using  A  =  10. 5.  (TTOTE:  If  a  more  precise 
evaluation  is  needed,  F(x)  should  be  determined  using  A  =  10. 44.  The 
approximate  value,  10.5,  was  used  since  A  =  10.44  would  require  interpo 
lation  in  Appendix  2.)  Since  £>  =  max  \F(x)  —  Sn (x)  \  =0.013  <  1.63 
/  VS75£  —  0.027,  the  hypothesis  may  not  be  rejected  at  the  1  per  cent 
significance  level.  The  reader  should  compare  this  result  with  that 
obtained  in  Example  7.27. 

TABLE   15. 4- Application  of  the  Kolmogorov-Smirnov  Goodness  of  Fit 
Test  to  the  Number  of  Busy  Senders  in  a  Telephone  Exchange 


Number 
Busy 

Observed 
Frequency 

Observed 
Curn  illative 
Frequency 

Relative 
Cumulative 
Frequency 
Sn(x) 

Expected 
Relative 
Cumulative 
Frequency 
FW 

\F(x*)-Sn(^\ 

0  

0 

0 

0 

0 

0 

1  

5 

5 

0.001 

o 

0.001 

2  

14 

19 

O.005 

0.002 

0.003 

3  

24 

43 

0.011 

0.007 

0.004 

4  

57 

100 

0.027 

0.021 

0.006 

5  

111 

211 

0.056 

0.050 

0.006 

6  

197 

408 

0.109 

0.102 

0  007 

7 

278 

686 

0   183 

0   179 

0  004 

8     ... 

378 

1064 

0.283 

0  279 

0  004 

9  

418 

1482 

0.395 

0  397 

0  002 

10 

461 

1943 

0.518 

0  521 

0  003 

11.  .  .    . 

433 

2376 

0.633 

0  639 

0  006 

12  

413 

2789 

0.743 

0.742 

0  001 

13  

358 

3147 

0.838 

0.825 

0  013 

14  

219 

3366 

0.897 

0.888 

0.009 

15  

145 

3511 

0.935 

0  932 

0  003 

16  

109 

3620 

0.964 

0.960 

0  004 

17 

57 

3677 

0  979 

0  978 

0  001 

18.    ., 

43 

3720 

0  991 

0  988 

0  003 

19  

16 

3736 

0  995 

0  994 

0  001 

20  

7 

3743 

0.997 

0.997 

o 

21  

8 

3751 

0.999 

0.999 

0 

22  

3 

3754 

1.000 

0.999 

O.OO1 

Data  Source:  Thornton  C.  Fry,  JProbability  and  Its  Engineering  Uses.  D.  Van  No  strand 
Company,  Inc.,  New  York,  1928,  p.  295. 


PROBLEMS  473 

15-6      MEDIAN   TESTS 

The  procedures  to  be  described  in  this  section  are  of  value  when  test 
ing  the  following  hypotheses; 

(1)  That   k  random   samples  were   drawn  from   identically   dis 
tributed  populations,  fc>2. 

(2)  That,  in  a  one-factor  experiment,  the  k  levels  of  the  factor 
have  the  same  effect. 

(3)  That,  in  a  two-factor  experiment,  (a)  the  a  levels  of  factor  a 
have  the  same  effect,  (b)  the  b  levels  of  factor  6  have  the  same 
effect,   and   (c)   there  is  no  true  interaction  between  factors 
a  and  6. 

For  our  purposes,  it  is  deemed  sufficient  to  concentrate  on  Case  1. 
Those  persons  wishing  to  investigate  Cases  2  and  3  are  referred  to 
Brown  and  Mood  (4,  5)  and  Mood  (22). 

If  k  random  samples  consisting  of  n\,  -  •  •  ,  nk  observations,  respec 
tively,  are  available,  determine  the  numbers  of  observations  in  each 
sample  that  are  above  and  below  the  median  of  the  combined  samples. 
These  data  may  then  be  analyzed  as  a  2Xfc  contingency  table  in  the 
manner  specified  in  Section  7.17. 

Example  15.8 

Consider  the  two  samples  in  Table  15.3.  Examination  of  these  data 
leads  to  the  2X2  contingency  table  shown  in  Table  15.5.  Using  Equa 
tion  (7.27),  we  obtain 

X2  -  22(|  (4) (3)  -  (8) (7)  |    -  11)Y(12)(10)(11)(11)  -  2.07. 

Since  x2  =  2.07  <X295(1)  =3.84,  we  are  unable  to  reject  the  hypothesis 
that  the  two  random  samples  were  drawn  from  identically  distributed 
populations. 

TABLE   15.5-Contingency  Table  Formed  From  the  Data 

of  Table  15.3 


Line  A 

Line  B 

Total 

Above  median  

4 

7 

11 

Below  median     .      ... 

8 

3 

11 

Total 

12 

10 

22 

Problems 

15.1      Apply  the  method  described  in  Section  15.2  to  the  following  problems: 

(a)  7.29 
(6)  7.32 
(c)  7.33 

Discuss  any  differences  between  the  method  used  here  and  that  used 
in  Chapter  7. 


474  CHAPTER    1  5,  ^DISTRIBUTION-FREE    METHODS 

15.2  Apply  the  method  described  in  Section  15.3  to  the  following  problems: 

(a)  7.29 
(6)  7.32 
(c)  7.33 

Discuss  your  results  with  reference  to  those  obtained  in  Problem  15.1 
and  in  Chapter  7. 

15.3  Apply  the  method  described  in  Section  15.3  to  the  following  problems: 

(a)   7.1 
(6)   7.2(d) 
(c)    7.3 

State  any  changes  in  the  assumptions  and  the  wording  of  the  hy 
potheses  that  you  make.  Discuss  the  implications  of  these  changes. 
Compare  the  results  of  using  distribution-free  tests  with  those  ob 
tained  using  parametric  tests  in  Chapter  7. 

15.4  Apply  the  method  described  in  Section  15.4  to  the  data  given  in  the 
following  problems: 

(a)  6.1  (i)    7.30 

(6)  6.5  (/)    7.31 

(c)  6.20  (fc)    7.33 

(d)  6.22  (Z)    7.35 

(e)  6.23  (m)  7.36 
Of)  7.1  (n)   7.37 
(0)  7.2  (o)    7.39 
(A)  7.3  (p)   7.40 

In  each  case,  state  the  hypothesis  being  tested  and  specify  any  as 
sumptions  you  make. 

15.5  Apply  the  method  described  in  Section  15.5  to  check  on  the  assump 
tion  of  normality  in  the  following  problems: 

(a)    6.5 

(6)    7.1 

15.6  Apply  the  method  described  in  Section  15.6  to  the  data  given  in  the 
following  problems: 

(a)   7.30  (gr)  7.40 

(6)    7.31  (/O  11.9 

(c)  7.35  (0  11.10 

(d)  7.36  (j)  11.11 
(<0    7.37  (Jfe)  11.12 
CO    7.39 

Compare  the  conclusions  reached  here  with  those  reached  in  the  earlier 
chapters.  Discuss  any  discrepancies. 

References  and  Further  Reading 

1.  Birnbaum.,  Z.  W.  Numerical  tabulation  of  the  distribution  of  Kolmogorov's 
statistic  for  finite  sample  values.  Jour.  Amer.  Stat.  Assn.,  47:425—41,  1952. 

2.  .  Distribution-free  tests  of  fit  for  continuous  distribution  functions. 

Ann.  Math.  Stat.,  24:1-8,  1953. 


REFERENCES    AND    FURTHER    READING  475 

3.  Blum,  J.  R.,  and  Fattu,   N.  A.   Nonparametric  methods.  Rev.  Educ.  Res., 
24:467-87,   1954. 

4.  Brown,   G.  W.,  and  Mood,  A.  M.  Homogeneity  of  several  samples.  Amer. 
Stat.,  2:22,  1948. 

5.   ,  and .  On  median  tests  for  linear  hypotheses.   Proceedings  of 

the  Second  Berkeley  Symposium  on  Mathematical  Statistics  and  Probability, 
pp.  159—66,  University  of  California  Press,  Berkeley,  Calif.,   1951. 

6.  Brownlee,  K.  A.  Statistical  Theory  and  Methodology  in  Science  and  Engineer 
ing.  John  Wiley  and  Sons,  Inc.,  New  York,  1960. 

7.  Brunk,    H.    D.   An   Introduction  to   Mathematical   Statistics.    Ginn   and    Co., 
New  York,  1960. 

8.  Cochran,  W.  G.  The  x2  test  of  goodness  of  fit.  Ann.  Math.  Stat.,  23:315-45, 
1952. 

9.  .  Some  methods  for  strengthening  the  common  x2  tests.  Biometrics, 

10:417-51,  1954. 

10.  Dixon,  W.  J.  Power  under  normality  of  several  non-parametric  tests.  Ann. 
Math.  Stat.,  25:610-14,  1954. 

11.  ,  and  Massey,   F.  J.,  Jr.  Introduction  to  Statistical  Analysis.  Second 

Ed.  McGraw-Hill  Book  Company,  Inc.,  New  York,  1957. 

12.  ,  and  Mood,  A.  M.  The  statistical  sign  test.  Jour.  Amer.  Stat.  Assn., 

41:557-66,  1946. 

13.  Festinger,  L.  The  significance  of  differences  between  means  without  refer 
ence  to  the  frequency  distribution  function.  Psychometrika,   11:97,   1946. 

14.  Fraser,  D.  A.  S.  Nonparametric  Methods  in  Statistics.  John  Wiley  and  Sons, 
Inc.,  New  York,  1957. 

15.  Kendall,   M.   G.   Rank  Correlation   Methods.    Charles    Griffin  and  Company 
Limited,  London,  1948. 

16.  Kruskal,  W.  H.  Historical  notes  on  the  Wilcoxon  unpaired  two-sample  test. 
Jour.  Amer.  Stat.  Assn.,  52:356-60,  1957. 

17.   ,  and  Wallis,  W.  A.  Use  of  ranks  in  one-criterion  variance  analysis. 

Jour.  Amer.  Stat.  Assn.,  47:583,  1952, 

18.  Mann,  H.  B.,  and  Whitney,  ID.  H.  On  a  test  of  whether  one  of  two  random 
variables  is  stochastically  larger  than  the   other.   Ann.   Math.   Stat.,   18:50, 
1947. 

19.  Massey,  F.  J.,  Jr.  The  Kolmogorov-Smirnov  test  for  goodness  of  fit.  Jour. 
Amer.  Stat.  Assn.,  46:68-78,  1951. 

20.   .  The  distribution  of  the   maximum  deviation   between  two   sample 

cumulative  step  functions.  Ann.  Math.  Stat,,  22:125—28,  1951. 

21.  Mood,  A.  M.  The  distribution  theory  of  runs.  Ann.  Math.  Stat.,  11:367—92, 
1940. 

22.   .  Introduction  to  the  Theory  of  Statistics.  McGraw-Hill  Book  Company, 

Inc.,  New  York,  1950. 

23.  Moses,   L.  E.   Non-parametric  statistics  for  psychological  research.   Psych. 
Bui.,  49:122-43,  1952. 

24.  Olmstead,  P.  S.,  and  Tukey,  J.  W.  A  corner  test  for  association.  Ann.  Math. 
Stat.,  18:495-513,  1947. 

25.  Savage,  I.  H,  Bibliography  of  nonparametric  statistics  and  related  topics. 
Jour.  Amer.  Stat.  Assn.,  48:844-906,   1953. 

26.  SchefTe*,    H.   Statistical  inference  in   the  non-parametric   case.   Ann.    Math. 
Stat.,   14:305-32,   1943. 

27.  Siegel,  S.  Nonparametric  Statistics  for  the  Behavioral  Sciences.  McGraw-Hill 
Book  Company,  Inc.,  New  York,  1956. 

28.  Smirnov,  N.  V.  Table  for  estimating  the  goodness  of  fit  of  empirical  distri 
butions.  Ann.  Math.  Stat.,  19:279-81,  1948. 

29.  Smith,   K.   Distribution-free  statistical  methods  and  the  concept  of  power 
efficiency.    Research   Methods  in  the  Behavioral  Sciences   (edited  by  L.  Fes 
tinger  and  D.  Katz).  Dryden  Press,  New  York,  1953. 


476  CHAPTER    15,    DISTRIBUTION-FREE    METHODS 

30.  Swed,  Frieda  S.,  and  Eisenhart,  C.  Tables  for  testing  randomness  of  group 
ing  in  a  sequence  of  alternatives.  Ann.  Math.  Stat.,  14:83-86,  1943. 

31.  Wilcoxon,    F.    Individual    comparisons    by    ranking    methods.    Biometrics, 
1:80-83,  1945. 

32.  .  Probability  tables  for  individual  comparisons  by  ranking  methods. 

Biometrics,  3:119-22,  1947. 

33.  .  Some  Rapid  Approximate  Statistical  Procedures.  American  Cyanamid 

Co,,  Stanford,  Conn.,  1949. 

34.  Wilks,  S.  S.  Order  statistics.  Bui.  Amer.  Math.  Soc.,  54:6-50,  1948. 


CHAPTER     16 

STATISTICAL  QUALITY  CONTROL 

SINCE  THE  EARLY  1940's  the  use  of  statistical  methods  in  industry  has 
been  on  the  upswing.  This  increased  use  of  statistics  has  been  par 
ticularly  noticeable  in  two  areas:  (1)  research  and  development  and  (2) 
reliability  and  quality  control.  Because  the  major  part  of  this  book 
has  been  concerned  with  those  statistical  methods  that  have  proved 
most  valuable  in  research  and  development,  this  chapter  will  be 
devoted  to  a  presentation  of  two  special  techniques  especially  useful 
for  controlling  quality  and  reliability.  (NOTE:  This  is  not  to  imply 
that  these  methods  are  of  no  value  to  those  persons  located  in  research 
and  development  organizations;  it  is  only  a  statement  of  relative 
importance.) 

The  two  techniques  not  discussed  in  the  preceding  chapters  and 
which  are  especially  useful  in  controlling  and  improving  quality  and 
reliability  are  control  charts  and  acceptance  sampling  plans.  Because 
these  two  techniques  are  discussed  at  great  length  in  books  devoted 
entirely  to  the  subject  of  statistical  quality  control,  only  a  brief  out 
line  of  each  will  be  given.  However,  the  material  to  be  presented  will 
be  sufficient  to  acquaint  the  reader  with  the  basic  concepts.  Those 
persons  in  need  of  more  detailed  explanations  are  referred  to  the  publi 
cations  listed  at  the  end  of  the  chapter. 

1 6.1       CONTRO  L«  CHARTS 

Although  control  charts  may  prove  useful  in  many  situations  they 
are  most  commonly  employed  in  the  analysis  and  control  of  production 
processes.  For  this  reason,  discussion  of  these  charts  will  be  in  terms 
perhaps  more  familiar  to  the  engineer  than  to  the  research  worker. 

It  has  long  been  recognized  that  some  variation  is  inevitable  in  any 
repetitive  process.  For  example,  Grant  (16)  states: 

Measured  quality  of  manufactured  product  is  always  subject  to  a  certain 
amount  of  variation  as  a  result  of  chance.  Some  stable  "system  of  chance 
causes'3  is  inherent  in  any  particular  scheme  of  production  and  inspection. 
Variation  within  this  stable  pattern  is  inevitable.  The  reasons  for  variation 
outside  this  stable  pattern  may  be  discovered  and  corrected.  .  1 

Taking  Grant  at  his  word,  what  we  seek,  then,  are  tests  for  detecting 
unnatural  patterns  in  the  plotted  data. 

The  control  chart,  as  conceived  and  developed  by  Shewhart  (21),  is 
"a  simple  pictorial  device  for  detecting  unnatural  patterns  of  variation 
in  data  resulting  from  repetitive  processes.  That  is,  the  control  chart 

1  E.  L.  Grant,  Statistical  Quality  Control,  Second  Edition,  McGraw-Hill  Book 
Company,  Inc.,  KTew  York,  1952,  p.  3. 

C477J 


478  CHAPTER    16,    STATISTICAL    QUALITY    CONTROL 

provides  criteria  for  detecting  lack  of  statistical  control.  (NOTE: 
When  a  process  is  operating  under  a  constant  system  of  chance  causes, 
it  is  said  to  be  in  statistical  control.)  Rather  than  go  into  details  con 
cerning  the  theory  underlying  control  charts,  we  shall  be  content  with: 
(1)  indicating  the  proper  procedures  for  constructing  the  charts,  (2) 
stating  criteria  to  be  used  for  indicating  unnatural  patterns  of  vari 
ation,  and  (3)  giving  numerical  examples  of  the  four  most  commonly 
used  charts. 

Basically,  all  control  charts  appear  as  in  Figure  16.1.  The  sample 
points  are,  of  course,  plotted  in  a  sequential  manner,  that  is,  as  ob- 


UPPER    CONTROL     LIMIT 


o 

1 

/^-^        /         \  I  \ 

CENTER 


o: 
LJ 

g 

ID 

O 

I— 

CO 

§ 

o: 
O 


\ 


LINE 


-LOWER    CONTROL     LIMIT 


1  2  3          A  5  6  7  8  9  1O         11          12 

SAMPLE      NUMBER 

FIG-    16.1 -—Illustration    of  the  general  appearance   of  a   control   chart. 

tained.  The  plotted  points  are  joined  together  solely  as  an  aid  to  the 
visual  interpretation. 

ISTow,  what  tests  should  be  employed  to  detect  unnatural  patterns  in 
the  plotted  data?  Depending  on  the  degree  of  effectiveness  desired, 
there  are  many  criteria  that  might  be  used.  However,  since  our  sole 
purpose  is  to  indicate  the  basic  nature  of  the  control  chart  technique, 
only  the  most  common  tests  will  be  mentioned.  As  mentioned  earlier, 
those  persons  wishing  to  investigate  more  sophisticated  tests  should 
consult  the  references  at  the  end  of  the  chapter. 

The  most  common  tests  for  unnatural  patterns  are  tests  for  insta 
bility,  that  is,  tests  for  determining  if  the  cause  system  is  changing.  As 
commonly  employed,  they  refer  to  the  A,  B,  and  C  zones  shown  in 
Figure  16.2  With  reference  to  these  zones,  the  observed  pattern  of 
variation  is  said  to  be  unnatural,  or  the  process  is  said  to  be  "out  of 


16.1        CONTROL    CHARTS  479 

control/'  if  any  one  or  more  of  the  following  events  occurs:2 

(1)  A  single  point  falls  outside  of  the  control  limit;  i.e.,  beyond 
zone  A. 

(2)  Two  out  of  three  successive  points  fall  in  zone  A  or  beyond. 

(3)  Four  out  of  five  successive  points  fall  in  zone  B  or  beyond. 

(4)  Eight  points  in  succession  fall  in  zone  C  or  beyond. 

It  should  be  noted  that  the  above  tests  apply  to  both  halves  of  a  control 
chart  but  they  are  applied  separately  to  each  half,  not  to  the  two  halves 
in  combination. 

CONTROL    LIMIT 


ZONE    A 


ZONE    B 


ZONE   C 


^  -  CENTER     LINE 

FIG.   16.2—  Diagram  defining  the  A,  B,  and  C  zones  used  in  control 

chart  analyses.  (Each  of  the  zones.  A,  B,  and  C,  constitutes 

one-third    of   the    area    between    the    center    line 

and    the   control    limit.) 

Before  presenting  numerical  illustrations  of  control  chart  applica 
tions,  it  is  necessary  that  the  four  most  commonly  used  charts  be  intro 
duced  and  that  formulas  be  given  for  the  calculation  of  the  center  lines 
and  control  limits.  The  charts  most  often  encountered  are:  (1)  the  ~X 
chart,  (2)  the  R  chart,  (3)  the  p  chart,  and  (4)  the  c  chart.  The  first  two 
of  these  charts  deal  with  measurement  data  while  the  last  two  deal 
with  attribute  (enumeration)  data.  The  pertinent  assumptions  and  the 
formulas  for  the  control  limits  are  specified  in  Table  16.1. 

Example  16.1 

Consider  the  data  of  Table  16.2.  It  may  be  verified  that  If  —  T^SVfc 
=  213,20/20  =  10.66  and   R=*  J2R/k  =  31.  8/20  =  1.59.   Then,   using  the 


2  The  tests  described  here  may  be  used  when  the  two  control  limits  are  reason 
ably  symmetrically  located  with  respect  to  the  center  line.  If  the  limits  are  de 
cidedly  asymmetric,  the  tests  should  be  modified  as  described  in  (33). 


48O 


CHAPTER    16,   STATISTICAL   QUALITY    CONTROL 


TABLE  16.1— Assumptions  and  Formulas  for  the  Most  Commonly  Used 

Control  Charts 


Chart 

Assumed 
Distribution 

Center 
Line 

Upper  Control 
Limit  (UCL) 

Lower  Control 
Limit  (LCL) 

~x  

Normal 

*X 

X+A^R 

X—Aolt 

R  

Normal 

R 

D*R 

D*R 

Binomial 

P 

p+3\/p(l—p)/n 

p  —  3Vp(l  —  p)/n 


c  

Poisson 

~c 

Z-hS-x/f 

"c  —  3\/zF 

The  constants  A2,  Z>3y  and  D4  are  given  in  Appendix  8,  while  the  quantities  X,  R,  $, 
and  c  are  calculated  from  the  sample  data  as  shown  in  the  numerical  examples  which 
follow. 

formulas  given  in  Table  16.1,  we  see  that,  for  the  3T  chart,  UCL  =  10. 66 
+  (0.58) (1.59)  =  11,58  and  LCL  =  10.66-  (0.58) (1.59)  =9.74.  Similarly, 
for  the  R  chart,  UCL  =  (2.11)(1.59)  =3.35  and  LCL  -  (0)(1.59)  =0.  The 
resulting  charts  are  shown  in  Figure  16.3.  The  tests  for  unnatural 
patterns  have  been  applied  to  the  X  chart  and  the  potential  trouble 

TABLE  16.2-Coded  Values  of  the  Crushing  Strengths  of  Concrete  Blocks 


Sample 

Number 

Individual 

Values 

Mean 
(3) 

Range 

X,             X2             X,             X4             X5 

1  

11.1 

9.6 
9.7 
10.1 
12.4 

10.1 
11.0 
11.2 
10.6 
8.3 

10.6 
10.8 
10.7 
11.3 
11.4 

10.1 
10.7 
11.9 
10.8 
12.4 

9.4 
10.8 
10.0 
8.4 
10.0 

10.2 
11.5 
10.0 
10.4 
10.2 

9.9 
10.2 
10.7 
11.4 
11.2 

10.1 
12.8 
11.9 
12.1 
11.1 

11.2 
10.1 
10.0 
10.2 
10.7 

10.2 
11.8 
10.9 
10.5 
9.8 

10.7 
10.5 
10.8 
10.4 
11.4 

9.7 
11.2 
11.6 
11.8 
10.8 

10.4 
10.8 
9.8 
9.4 
10.1 

11.2 
11.0 
11.2 
10.5 
9.5 

10.2 
8.4 
8.6 
10.6 
10.1 

9.8 
11.2 
12.4 
9.4 
11.0 

10.1 
11.0 
10.4 
11.0 
11.3 

10.1 
11.3 
11.0 
10.9 
9.8 

11.4 
9.9 
11.4 
11.1 
11.6 

10.5 
11.3 
11.4 
11.6 
11.9 

10.44 
10.46 
9.98 
9.82 
10.90 

10.36 
11.32 
10.86 
10.58 
9.52 

10.56 
9,96 
10.44 
10.96 
11.14 

10.04 
11.44 
11.84 
11.14 
11.44 

1.8 
1.4 
0.7 
2.6 

2.4 

1.1 
0.8 
1.2 
0.5 
1.9 

1.5 

2.4 
2.8 
1.0 
1.5 

0.8 
2.1 
1.0 

2.7 
1.6 

2  

3  

4 

5  

6  

7  

8   

9.    . 

10   . 

11,  

12  

13  

14  

15 

16.,..  

17  

18 

19     

20  

Average 

10.66 

1.59 

16.1       CONTROL    CHARTS 


481 


10  - 


12345 


1O 
SAMPLE    NUMBER 


20 


2345 


10 
SAMPLE     NUMBER 


15 


2O 


FIG.    16.3-Control  charts  for  the  data   of  Table    16.2. 

spots  tagged  in  the  customary  manner.  It  is  suggested  that  the  reader 
consider  the  application  of  these  tests  to  the  R  chart. 

Example  16.2 

Consider  the  data  of  Table  16.3.  Here  we  are  dealing  with  enumera 
tion  data,  namely,  the  number  of  defective  fuses  in  samples  of  size  50 
taken  at  random  times  during  the  production  process.  It  is  easily  verified 
that  p=  ^Cp/fc^1-68/40^0-042-  Usin£  the  formulas  specified  in  Table 
16.1,  we  obtain 

0.127 


UCL  =  0.042  +  3VC0.042)  (0.958) /SO  = 


and 


LCL  =  0.042  -  3V(0.042)(0.958)/50  =  -  0.043. 


(NOTE:  Since  the  formula  leads  to  a  negative  value  for  the  lower  control 
limit  and  because  the  fraction  defective  is  a  nonnegative  quantity,  the 
lower  control  limit  is  arbitrarily  set  at  0.  This,  of  course,  makes  the 
control  limits  asymmetric  with  respect  to  the  center  line,  The  tests  for 


TABLE  16.3-Number  of  Defective  Fuses  in 
Random  Samples  of  Size  50 


Sample  Number 

Number  of 
Defectives 

Fraction 
Defective  (p) 

1  

2 

0.04 

2  

1 

0.02 

3.  . 

2 

0.04 

4  

0 

0.00 

5  

2 

0.04 

6  

3 

0.06 

7  

4 

0.08 

8  

2 

0.04 

9.                            . 

0 

0.00 

10  

3 

0.06 

11  

0 

0.00 

12  

1 

0.02 

13  

2 

0.04 

14   

2 

0.04 

15  

3 

0.06 

16  

5 

0.10 

17    

1 

0.02 

18  

2 

0.04 

19  

3 

0.06 

20    

1 

0.02 

21 

1 

0  02 

22.    .  .. 

1 

0.02 

23  

4 

0.08 

24  

2 

0.04 

25  

2 

0  04 

26  

4 

0.08 

27  

1 

0.02 

28  

3 

0.06 

29  

3 

0.06 

30  

2 

0.04 

31  

3 

0.06 

32  

6 

0.12 

33  

2 

0.04 

34  

3 

0.06 

35  

2 

0.04 

36  

3 

0.06 

37  

1 

0.02 

38  

0 

0.00 

39  

2 

0.04 

40  

0 

0.00 

Average 

0.042 

16.1        CONTROL    CHARTS  483 


.=  0.127 


.  p=tO.O-42 

T7T7 — \  /\   A'     V  \     /  "  \/    v — *   v  \    X 

O.O2 

"  ....          ..          r-  »^ . 1 , . , , . . . . . , j . . .         _j  j.    ...     m.  .  _  f  i  m  ._      ..    _j  I 

12345  10  15  20  25  3O  35  AO 

SAMPLE     NUMBER 

FIG.    16.4— Control   chart  for  the  data   of  Table    16.3, 

unnatural  patterns  specified  earlier  have,  therefore,  not  been  applied. 
The  application  of  the  modified  tests  is  left  as  an  exercise  for  the  reader. 
The  results  of  the  preliminary  analysis  are  presented  in  Figure  16.4.) 

Example  16.3 

Consider  now  a  situation  in  which  the  characteristic  of  interest  is  the 
number  of  defects  per  unit.  In  such  a  case,  a  Poisson  distribution  would 
undoubtedly  be  assumed  and  a  c  chart  would  be  appropriate.  Given  the 
data  in  Table  16.4,  it  is  easily  verified  that  c=  T^c/fc  =  144/24  =  6. 
Using  the  formulas  specified  in  Table  16.1,  we  obtain  UCL==6  +  3VB 
=  13.35  and  LCL  =  6  —  3v^=  —1.35.  (NOTE:  As  in  Example  16.2,  the 
negative  lower  control  limit  will  arbitrarily  be  changed  to  0  since  c  is, 
by  definition,  a  nonnegative  quantity.)  The  results  of  the  analysis  are 
plotted  in  Figure  16.5. 


12345  1O  15  2O  25 

SAMPLE    NUMBER 

FIG.   16.5— Control  chart  for  the  data  of  Table  16.4. 


484 


CHAPTER    16,    STATISTICAL    QUALITY    CONTROL 


TABLE   16.4r-Number  of  Defects  Observed  in  a  Welded  Seam  (Each  Count 
Taken  on  a  Single  Seam;  the  Welder  produced  8  Seams  per  Hour) 


Sample 
Number 

Date 

Time  of 
Sample 

Number  of 
Defects  (c) 

1  

July  18 

8:00  A.M. 

2 

2  

9:05  A.M. 

4 

3  

10:10  A.M. 

7 

4  

11  :00  A.M. 

3 

5  

12  :30  P.M. 

1 

6  

1  :35  P.M. 

4 

7  

2:20  P.M. 

8 

8  

3:30  P.M. 

9 

9  

July  19 

8:10  A.M. 

5 

10  

9:00  A.M. 

3 

11  

10:05  A.M. 

7 

12  

11:15  A.M. 

11 

13  

12:25  P.M. 

6 

14  

1  :30  P.M. 

4 

15 

2  :30  P.M. 

9 

16 

3:40  P.M. 

9 

17  

July  20 

8:00  A.M. 

6 

18  

8:55  A.M. 

4 

19  

10:00  A.M. 

3 

20 

11:10  A.M. 

9 

21  ... 

12:25  P.M. 

7 

22  

1  :30  P.M. 

4 

23  

2:20  P.M. 

7 

24  

3:30  P.M. 

12 

Total 

144 

Source:  E.  L.  Grant,  Statistical  Quality  Control,  Second  Edition,  McGraw-Hill  Book 
Company,  Inc.,  New  York,  1952,  p.  33.  By  permission  of  the  author  and  publishers. 

Examination  of  Figure  16.5  will  reveal  that  some  modifications  have 
been  made  in  the  usual  method  of  presentation,  namely,  the  various 
half-days  have  been  separated  (i.e.,  not  connected)  to  emphasize  the 
breaks  between  work  periods.  Now,  it  will  be  noted  that  two  conclusions 
are  obvious:  (1)  no  points  plot  above  the  upper  control  limit,  and  (2) 
essentially  the  same  pattern  appears  in  each  half-day.  Study  of  the 
recurring  pattern  suggests  the  existence  of  a  fatigue  factor.  [NOTE:  For 
further  discussion  of  this  example,  consult  Grant  (16,  pp.  32—35),] 

Before  leaving  the  topic  of  control  charts,  some  additional  remarks 
are  necessary.  These  will,  however,  be  very  brief  and  in  no  particular 
order. 
1.   The  primary  reason  for  using  control  charts  is  to  provide  a  signal 

that  some  action  is  desirable. 


16.2       ACCEPTANCE   SAMPLING    PLANS  485 

2.  Before  control  limits  may  be  calculated  with,  any  assurance  of  their 
being  reliable,  at  least  20  subgroups  (samples)  should  be  available. 

3.  Before  the  control  limits,  calculated  from  past  production  records, 
are  used  to  monitor  future  production,  the  process  should  be  in 
control. 

4.  If  a  process  is  in  control  and  a  point  plots  outside  the  control  limits, 
the  taking  of  action  may  be  viewed  as  committing  a  Type  I  error. 

5.  The  control  chart  is  not  a  panacea  for  all  production  problems;  it  is 
only  another  useful  tool. 

16.2      ACCEPTANCE  SAMPLING   PLANS 

Acceptance  sampling,  or  sampling  inspection,  is  of  two  types :  lot-by- 
lot  sampling  and  sampling  of  continuous  production.  In  this  brief  ex 
posure  to  the  concepts  and  procedures  of  acceptance  sampling,  only  the 
first  of  these  two  types  will  be  discussed.  Persons  desiring  information 
on  continuous  sampling  plans  are  referred  to  the  publications  listed  at 
the  end  of  the  chapter.  In  addition  to  the  distinction  between  lot-by- 
lot  and  continuous  sampling,  it  is  customary  to  classify  sampling  plans 
as  either  attributes  or  variables  plans.  Attributes  plans  refer  to  those 
cases  in  which  eacli  item  is  classified  simply  as  eitlier  defective  or  non- 
defective  ;  variables  plans  refer  to  those  cases  in  which  a  measurement 
is  taken  and  recorded  (numerically)  on  each  item  inspected.  In  this 
section  only  attributes  plans  will  be  considered.  There  is  one  other  way 
in  which  acceptance  sampling  plans  may  be  classified,  namely,  as 
single,  double,  or  multiple  (including  sequential}  plans.  These  categories 
refer,  of  course,  to  the  number  of  samples  selected  and  will  become 
clear  as  the  exposition  continues. 

An  attributes  single  sampling  plan  which  operates  on  a  lot-by-lot  basis 
is  completely  defined  by  three  numbers:  the  lot  size,  AT;  the  sample 
size,  n;  and  the  acceptance  number,  a.3  Such  a  plan  operates  as  follows: 

(1)  A  single  sample  of  n  items  is  selected,  by  chance,  from  a  lot  of 
N  items. 

(2)  Each  item  in  the  sample  is  then  classified  as  either  defective 
or  nondefective. 

(3)  If  the  number  of  defective  items  in  the  sample  does  not  exceed 
a,  the  lot  is  accepted. 

(4)  If  the  number  of  defective  items  in  the  sample  exceeds  a,  the 
lot  is  rejected. 

As  with  all  statistical  procedures,  the  risks  associated  with  decisions 
(inferences)  resulting  from  sampling  inspection  must  be  assessed.  The 
customary  manner  of  presenting  these  risks  is  by  means  of  a  graph  of 
the  OC  function  of  the  sampling  plan,  that  is,  by  plotting  the  prob 
ability  of  accepting  the  lot  as  a  function  of  the  fraction  defective  of  the 
lot.  The  protection  afforded  by  various  sampling  plans  may  then  be 
compared  by  examining  their  OC  curves. 

3  In  many  publications,  the  acceptance  number  is  denoted  by  c.  However,  I 
prefer  a  for  two  reasons:  it  stands  for  the  word  acceptance  and  it  permits  an 
easier  extension  to  double  and  multiple  sampling. 


486  CHAPTER    16,    STATISTICAL   QUALITY    CONTROL 

For  the  single  sampling  plan  specified  previously,  the  OC  function  is 


=          C(D,  d}-C(N  -  D,n-  d}/C(N,  n)  (16. 


where  D  represents  the  number  of  defective  items  in  the  lot  and  d 
represents  the  number  of  defective  items  in  the  sample.  Equation  (16.1) 
may,  of  course,  be  evaluated  for  Z)  =  0,  1,  -  -  *  ,  N  and  the  results 
plotted  as  a  series  of  ordinates  erected  at  the  corresponding  values  of 
p  =  D/N,  namely:  0,  1/-ZV,  2/N,  •  •  •  ,  1.  It  is  clear,  however,  that  these 
calculations  may  prove  onerous  unless  a  high-speed  computer  is  avail 
able.  Fortunately,  helpful  tables  of  the  hypergeometric  function  have 
been  provided  by  Lieberman  and  Owen  (19).  In  addition,  extensive 
catalogs  of  OC  curves  for  lot-by-lot,  single  sampling  (by  attributes) 
plans  have  been  prepared  by  Wiesen  (34)  and  Clark  and  Koopmans  (9) . 
If  none  of  these  publications  is  readily  available  and  access  to  a  high 
speed  computer  is  impossible,  the  OC  function  may  be  approximated  by 

C(n,  d)pd(l  -  p^~d  (16.2) 

d=Q 

or 


Pace  S*  2L,  e-^\d/d\  (16.3) 

where  X  =  up.  (NOTE :  The  reader  is  referred  to  Chapter  5  for  discussion 
of  the  accuracy  and  relevancy  of  these  approximations.) 

Persons  familiar  with  the  presentation  of  OC  curves  for  sampling 
plans,  or  those  individuals  who  consult  the  references  given  in  the 
preceding  paragraph,  will  realize  that  it  is  not  customary  to  plot  OC 
functions  as  a  series  of  ordinates  (as  stated  in  the  preceding  paragraph) 
but  to  show  smooth  OC  curves.  That  is,  the  functions  are  plotted  as 
though  p  were  a  continuous  parameter.  For  lot-by-lot  plans,  this 
minor  "tampering  with  the  truth"  is  not  serious  and  the  resulting  gain 
in  ease  of  presentation  far  outweighs  the  inaccuracy  of  the  graph. 
Consequently,  OC  curves  usually  appear  as  in  Figure  16.6. 

Rather  than  plot  the  entire  OC  function,  the  practitioner  frequently 
contents  himself  with  calculating  two  or  three  points  on  the  curve.  The 
three  points  most  commonly  determined  are: 

Pi  =  the  value  of  p  for  which  Pacc  =  0.95, 

p%  =  the  value  of  p  for  which  Pacc  =  0.50,  and 

Pz  =  the  value  of  p  for  which  Pacc  =  0.10. 

These  three  values  of  p  (that  is,  pi,  p2,  and  z>3)  are  usually  referred  to  as 
the  acceptable  quality  level  (AQL),  the  indifference  quality }  and  the 


16.2       ACCEPTANCE    SAMPLING    PLANS 


487 


PRODUCER'S    RISK 


:ONSUMERIS     RISK 


AQL  RQL 

p=LOT    FRACTION    DEFECTIVE 

FIG.    1 6.6— Illustration   of  the  general   appearance  of  an   OC   curve, 

(To    aid    in    understanding    the    concepts,    the   AQL,    RQL,    consumer's 

risk,    and    producer's    risk    are    shown    in    this    figure.) 

rejectable  quality  level  (RQL),4  respectively.  The  AQL  and  RQL  points, 
as  well  as  the  associated  expressions,  consumer's  risk  and  producer's 
risk  have  been  shown  on  Figure  16.6.  (NOTE  :  If,  in  the  definitions  of  pa 
and  j>3,  the  values  0.95  and  0.10  are  replaced  by  1  —  a.  and  0,  respec 
tively,  the  reader  will  immediately  see  the  close  connection  between 
the  ideas  of  this  section  and  those  discussed  in  Chapter  7.  Incidentally, 
the  same  remark  may  be  made  here  as  there,  namely,  the  values  as 
signed  to  a,  and  0  are  arbitrary;  the  use  of  «  =  0.05  and  £  =  0.10  is  only 
a  matter  of  custom.) 

One  other  common  way  of  presenting  the  performance  ability  of 
acceptance  sampling  plans  is  to  calculate  and  plot  the  average  outgoing 
quality  (AOQ)  function.  This  function,  restricted  to  cases  in  which  the 
testing  is  nondestructive,  depends  on  the  assumption  that  all  rejected 
lots  are  submitted  to  100  per  cent  inspection,  with  all  defective  items 
being  removed  and  replaced  by  nondefective  items.  Under  this  as 
sumption,  the  average  outgoing  quality  is  determined  to  be 


AOQ 


(16.4) 


4  Historically,  the  rejectable  quality  level  (RQL)  has  been  known  as  the  lot 
tolerance  per  cent  defective  (LTPD).  However,  in  recent  years,  the  expressions 
rejectable  quality  level  and  objectionable  quality  level  (OQL)  have  been  suggested, 
and  I  believe  that  rejectable  quality  level  will  soon  be  universally  adopted.  For 
this  reason,  it  is  used  in  this  book. 


488 


CHAPTER    16,   STATISTICAL   QUALITY    CONTROL 


where  Pacc  is  defined  by  Equation  (16.1).  The  graph  of  a  typical  AOQ 
function  is  shown  in  Figure  16.7.  The  maximum  value  of  the  average 
outgoing  quality  is  known  as  the  average  outgoing  quality  limit  (AOQL). 
From  a  practical  point  of  view,  the  AOQL  is,  perhaps,  the  most  im 
portant  descriptive  measure  associated  with  any  acceptance  sampling 
plan.  [NOTE:  AOQ  curves  are  also  included  in  the  catalogs  of  Wiesen 
(34)  and  Clark  and  Koopnaans  (9) .  ] 

AOQ 


AOQL 


O 


P*LOT      FRACTION      DEFECTIVE 
FIG.   16.7—  Illustration  of  the  general  appearance  of  an  AOQ  curve. 

Let  us  now  consider  double  sampling  plans.  An  attributes  double 
sampling  plan  which  operates  on  a  lot-by-lot  basis  is  completely  defined 
by  six  numbers:  the  lot  size,  N;  the  size  of  the  first  sample,  ni;  the 
acceptance  number  associated  with  the  first  sample,  a\\  the  rejection 
number  associated  with  the  first  sample,  rij  the  size  of  the  second 
sample,  n^*  and  the  acceptance  number  associated  with  the  combined 
samples,  a2.  Such  a  plan  operates  as  follows: 

(1)  A  sample  of  n±  items  is  selected,  by  chance,  from  a  lot  of  N 
items. 

(2)  Each  item  in  the  sample  is  then  classified  as  either  defective  or 
nondef  ective . 

(3)  If  the  number  of  defective  items  in  the  sample  does  not  exceed 
ai,  the  lot  is  accepted. 

(4)  If  the  number  of  defective  items  in  the  sample  equals  or  ex 
ceeds  ri,  the  lot  is  rejected. 

(5)  If  the  number  of  defective  items  in  the  sample  exceeds  ai  but 
is  less  than  ri,  a  second  sample  (of  n2  items)  is  selected,  by 
chance,  from  the  remainder  of  the  lot. 

(6)  If  the  number  of  defective  items  in  the  two  samples  combined 
does  not  exceed  a2,  the  lot  is  accepted. 

(7)  If  the  number  of  defective  items  in  the  two  samples  combined 
exceeds  a*,  the  lot  is  rejected. 

Rather  than  go  into  the  same  detail  as  was  done  for  single  sampling, 
only  the  bare  essentials  will  be  presented.  The  OC  function  is 


16.2       ACCEPTANCE    SAMPLING    PLANS  489 

r-l 

C(D,  dd-C(N  -  D,  m  -  dd/C(N,  «0  +      2^     C(D,  d$ 


•C(N  —  D,n±- 

d2=0 

•C(N  —  ni—  D  +  di,  n*  —  d$/C(N  —  nly  n%)  (16.5) 

where  di  represents  the  number  of  defective  items  in  the  first  sample 
and  d2  represents  the  number  of  defective  items  in  the  second  sample. 
Subject  to  the  usual  restrictions,  this  may  be  approximated  by 


C(n2?  J2)^(l  —  p)-2-^2  (16.6) 

^2=0 

or 

ai  ri~1  az~di 

Paoc^  2  «-XlxtV^il+     12     «-XlxtV^iI    S   «r-x*xtV^2l         (16.7) 

di=0  d1=a14-l  ^2=0 

where  \i  =  n^p  and  \2=:n2p.  As  before,  the  average  outgoing  quality  is 
given  by 

AOQ  =  p-Pacc  (16.8) 

where  Pocc  is  defined  by  Equation  (16.5).  In  addition,  since  the  total 
sample  size  is  now  a  variable,  it  is  also  possible  to  assess  the  relative 
"costs"  of  sampling  plans  by  comparing  their  expected  sample  sizes. 
To  summarize,  it  is  customary  to  calculate  the  average  sample  number 
(ASN).  Proceeding  on  the  assumption  that  every  item  in  each  sample 
is  inspected,  the  average  sample  number  for  a  double  sampling  plan  is 
given  by 


ASN  =  m  +  n2         3     C(D,dJ-C(N  —  Z>,  »i  —  d^/C(N,n^.      (16.9) 

<Zl*=ai+l 

The  graph  of  a  typical  ASN  function  for  a  double  sampling  plan  is 
shown  in  Figure  16.8,  where  it  is  compared  with  the  constant  sample 
size  of  an  equivalent  (in  terms  of  the  OC  function)  single  sampling 
plan. 

Before  proceeding  to  multiple  sampling  plans,  it  seems  desirable  to 
list  the  advantages  and  disadvantages  of  double  sampling  plans  rela 
tive  to  single  sampling  plans.  These  are: 

Advantages 

(1)   They  have  the  psychological  advantage  of  giving  doubtful 
lots  a  second  chance. 


490 


CHAPTER    16,    STATISTICAL    QUALITY    CONTROL 


(2)    On  the  average,  they  have  (for  the  two  extremes  of  good  and 
bad  quality)  the  advantage  of  requiring  fewer  inspections. 

Disadvantages 

(1)  They  are  said  to  be  more  difficult  to  administer.  (Personally, 
I  do  not  subscribe  to  this  point  of  view.) 

(2)  The  inspection  load  is  variable. 

(3)  The  maximum  number  of  inspections  can   (for  intermediate 
quality)  exceed  that  for  comparable  single  sampling  plans. 


ASN 


DOUBLE 
SAMPLING 


n 


SINGLE 
SAMPLING 


0  p  =  LOT    FRACTION      DEFECTIVE 

FIG.    1  6.8— Illustration  of  the  general  appearance  of  ASN  functions 

for    single    and    double    sampling    plans    that    exhibit 

essentially   the    same    OC    curve. 

Since  there  is  little,  if  anything,  new  in  multiple  sampling  plans  that 
has  not  been  discussed  in  connection  with  single  and  double  sampling 
plans,  the  following  remarks  will  be  very  brief.  If  one  extends  the  con 
cepts  and  procedures  of  double  sampling  to  k  >  2  samples,  it  may  easily 
be  seen  that  such  plans  are  completely  defined  by  the  numbers:  AT, 
ni,  •  -  •  ,  nk,  ai,  -  -  -  ,  a*,  ri,  •  -  -  ,  rk  =  a,k+l'  Proceeding  in  a  manner 
analogous  to  that  followed  for  double  sampling  plans,  one  may  obtain 
OC,  AOQ,  and  ASN  functions.  [NOTES:  (1)  The  calculations  will, 
however,  be  quite  involved  and  lengthy.  (2)  If  each  n,-=  1  and  if  k — >  oo  9 
the  multiple  plan  is  usually  referred  to  as  a  sequential  plan.]  For  those 
persons  desiring  more  details  on  multiple  and  sequential  acceptance 
sampling  plans,  the  appropriate  references  at  the  end  of  the  chapter 
are  recommended. 

Example  16.4 

Suppose  that  lots  of  size  100  are  submitted  for  acceptance.  A  sample 
of  size  2  is  drawn  from  a  lot  and  the  lot  is  accepted  if  both  the  sample 
items  are  nondefective;  if  one  or  both  are  defective,  the  lot  is  rejected. 
The  OC  function  for  this  plan  is 

Pace  =  CCA  0)-C(100  -  D,  2)/CC100,  2) 
and  this  may  be  evaluated  for  Z>  =  0,  1,  •  -  -  ,  100. 


PROBLEMS  491 

Example  16.5 

With  reference  to  Example  16.4,  it  is  easily  verified   that   the  AOQ 
function  is 

AOQ  =  p-P^c  -  [Z>/100][C(D,0)-C(100  -  D,  2)/C(1003  2)] 
where,  once  again,  this  may  be  evaluated  for  Z>  =  0,  1,  •  *  •  ,  100. 

Problems 

16.1  Define  and  discuss  each  of  the  following  terms:  (a)  quality,  (b)  con 
trol,  (c)  quality  control,  (d)  statistical  quality  control. 

16.2  Why  should  both  the  ~5c  chart  and  the  R  chart  be  used  when  dealing 
with  measurement  (as  opposed  to  attributes)  data? 

16.3  A  control  chart  for  ~X  has  the  upper  control  limit  (UCL)  and  the 
lower  control  limit   (LCL)    equal  to  24.23   and  23.99,   respectively. 
The  calculations  were  based  on  samples  of  size  four.  If  the  engineering 
specifications  had  been  given  as  24.10  ±0.20,  what  percentage  of  the 
sample  means  would  you  expect  to  plot  out  of  control?  State  any 
assumptions  you  find  it  necessary  to  make. 

16.4  Given  the  following  data,   construct  and  interpret  the  appropriate 
control  charts. 

SAMPLE  MEANS  AND  RANGES  (w— 10)  FOR  LENGTH 

OF  LIFE  OF  LIGHT  Bulges 
(CODED  DATA) __ 

Sample  No.  Mean  (5T)  Range  (R) 


1 

69.4 

45 

2 

63.4 

48 

3 

55.0 

72 

4 

64.0 

48 

5 

57.4 

36 

6 

82.0 

81 

7 

85.0 

78 

8 

33.4 

42 

9 

46.0 

69 

10 

112.4 

84 

11 

93.8 

48 

12 

95.6 

75 

13 

117.8 

51 

14 

113.6 

84 

15 

74.8 

54 

16 

80.8 

45 

17 

71.8 

57 

18 

53.2 

75 

19 

74.8 

48 

20 

59.2 

63 

21 

65.8 

129 

22 

109.6 

42 

23 

44.2 

51 

24 

73.6 

51 

25 

51.4 

27 

Total  1848.0  1503 


492  CHAPTER    16,    STATISTICAL    QUALITY    CONTROL 

16.5  Given  the  following  sample  data,  construct  control  charts  for  the 
sample  mean  and  the  sample  range,  that  is,  X  and  R  charts,  and 
interpret  the  results. 


Sample  

1 

2 

3 

4 

5 

6 

7 

8 

9 

Individual  values.  . 

+  2 

0 

+3 

+  1 

—  1 

—  1 

0 

0 

+  2 

—  1 

0 

+  1 

-3 

0 

0 

0 

+  2 

0 

+  1 

—  3 

+4 

0 

+  1 

—  1 

+  3 

—  2 

0 

0 

+2 

—  2 

+  1 

0 

—  2 

—  1 

—  «—  ^ 

+  1 

0 

—  3 

—  1 

—  1 

+1 

—  3 

+  3 

0 

+  1 

Sample    

10 

11 

12 

13 

14 

15 

16 

17 

18 

Individual  values.  . 

0 

+  1 

0 

—  2 

+  2 

+  2 

0 

0 

—  2 

+  3 

—2 

0 

—  3 

—  1 

—  2 

+4 

0 

—  1 

+2 

+  1 

—  2 

+  2 

0 

+  1 

—  4 

—  1 

+2 

+  1 

+  1 

—  2 

—  1 

—  2 

—  1 

0 

0 

—  1 

+  1 

+3 

0 

—  5 

—  1 

0 

+3 

+  2 

0 

Sample      

19 

20 

21 

22 

23 

24 

25 

Individual  values  .  . 

—  2 

+2 

0 

0 

—  1 

0 

+  1 

0 

—  2 

—  1 

+  2 

—  1 

+  1 

—  1 

—  1 

-1 

—  1 

+  1 

+  2 

+  1 

—  4 

—  1 

+  2 

—  1 

+  1 

_  i 

—  3 

—  1 

0 

—  1 

+  1 

+  1 

+  1 

—  2 

+  1 

PROBLEMS 


493 


16.6        Construct  and  interpret  the  appropriate  control  charts  for  the  fol 
lowing  data. 


Group  No. 

(-)      (V      (*>      (d)      (e) 

X 

R 

1 

.831 

.829 

.836 

.840 

.826 

.8324 

.014 

2 

.834 

.826 

.831 

.831 

.831 

.8306 

.008 

3 

.836 

.826 

.831 

.822 

.816 

.8262 

.020 

4 

.833 

.831 

.835 

.831 

.833 

.8326 

.004 

5 

.830 

.831 

.831 

.833 

.820 

.8290 

.013 

6 

.829 

.828 

.828 

.832 

.841 

.8316 

.013 

7 

.835 

.833 

.829 

.830 

.841 

.8336 

.012 

8 

.818 

.838 

.835 

.834 

.830 

.8310 

.020 

9 

.841 

.831 

.831 

.833 

.832 

.8336 

.010 

10 

.832 

.828 

.836 

.832 

.825 

.8306 

.011 

11 

.831 

.838 

.844 

.827 

.826 

.8332 

.018 

12 

.831 

.826 

.828 

.832 

.827 

.8288 

.006 

13 

.838 

.822 

.835 

.830 

.830 

.8310 

.016 

14 

.815 

.832 

.831 

.831 

.838 

.8294 

.023 

15 

.831 

.833 

.831 

.834 

.832 

.8322 

.003 

16 

.830 

.819 

.819 

.844 

.832 

.8288 

.025 

17 

.826 

.839 

.842 

.835 

.830 

.8344 

.016 

18 

.813 

.833 

.819 

.834 

.836 

.8270 

.023 

19 

.832 

.831 

.825 

.831 

.850 

.8338 

.025 

20 

.831 

.838 

.833 

.831 

.833 

.8332 

.007 

21 

.823 

.830 

.832 

.835 

.835 

.8310 

.012 

22 

.835 

.829 

.834 

.826 

,828 

.8304 

.009 

23 

.833 

.836 

.831 

.832 

.832 

.8328 

.005 

24 

.826 

.835 

.842 

.832 

.831 

.8332 

.016 

25 

.833 

.823 

.816 

.831 

.838 

.8282 

.022 

26 

.829 

.830 

.830 

.833 

.831 

.8306 

.004 

27 

.850 

.834 

.827 

.831 

.835 

.8354 

.023 

28 

.835 

.846 

.829 

.833 

.822 

.8330 

.024 

29 

.831 

.832 

,834 

.826 

.833 

.8312 

.008 

494  CHAPTER    16,    STATISTICAL    QUALITY    CONTROL 

16.7  Using  the  data  below,  calculate  limits  and  plot  the  ~X  and  R  charts. 
Apply  the  standard  tests  for  unnatural  patterns  and  discuss  the 
results.  (NOTE:  The  sample  size  is  7^  =  5.) 

X  R  X  R 


1 

1.444 

.09 

26 

1.424 

.05 

2 

1.427 

.08 

27 

1.434 

.05 

3 

1.464 

.08 

28 

1.414 

.09 

4 

1.455 

.08 

29 

1.406 

.07 

5 

1.462 

.10 

30 

1.418 

.14 

6 

1.448 

.05 

31 

1.438 

.09 

7 

1.454 

.04 

32 

1.416 

.07 

8 

1.446 

.08 

33 

1.419 

.06 

9 

1.437 

.12 

34 

1.406 

.08 

10 

1.471 

.11 

35 

1.428 

.06 

11 

1.438 

.09 

36 

1.430 

.06 

12 

1.438 

.05 

37 

1.421 

.07 

13 

1.415 

.12 

38 

1.434 

.07 

14 

1.428 

.12 

39 

1.408 

.05 

15 

1.425 

.08 

40 

1.414 

.08 

16 

1.440 

.09 

41 

1.410 

.03 

17 

1.430 

,05 

42 

1.406 

.06 

18 

1.457 

.09 

43 

1  .405 

.07 

19 

1.444 

.09 

44 

1.419 

.10 

20 

1.432 

.05 

45 

1.410 

.07 

21 

1.438 

.05 

46 

1.420 

.05 

22 

1.404 

.10 

47 

1.414 

.05 

23 

1.409 

.05 

48 

1.426 

.11 

24 

1.400 

.07 

49 

1.386 

.06 

25 

1.425 

.09 

50 

1.387 

.08 

PROBLEMS 


495 


16.8  Following  are  the  number  of  defective  piston  rings  inspected  in  23 
daily  samples  of  100.  Calculate  the  control  limits  for  the  type  of 
control  chart  that  should  be  used  with  these  data.  Interpret  the 
results. 


Date 

Number  Defective 

1    

9 

2  

5 

3.           

10 

4  

10 

5                    .... 

13 

8  

10 

9  

13 

10        

2 

11  

1 

12      .            

3 

15  

2 

16  

2 

Date 

Number  Defective 

17...           .          .     . 

3 

18 

6 

19  

3 

22 

3 

23  

3 

24 

0 

25  

5 

26  

5 

29 

3 

30  

2 

31                      .  .  .     . 

4 

496  CHAPTER    16,    STATISTICAL   QUALITY    CONTROL 

16.9        Given  the  following  data,  construct  the  appropriate  control  chart. 
Interpret  the  results. 

NTJMBEU  or  DEFECTIVES  IN  SAMPLES  or  SIZE  100 


Sample  No. 

Number  Defective  (cf) 

1  

3 

2       .              .... 

1 

3  

4 

4  

4 

5. 

4 

6  

6 

7  

5 

8 

5 

9  

2 

10  

4 

11.  . 

3 

12  

4 

13  

4 

14.  . 

3 

15 

5 

16  

8 

17  

2 

18 

3 

19  

5 

20  

4 

21  

3 

22  

4 

23  

6 

24  

4 

25  

3 

26  

5 

27  

4 

28  

7 

29  

6 

30  

5 

31  

5 

32  

6 

33  

4 

34  

9 

35  

6 

36  

4 

37  

3 

38  

1 

39  

3 

40  ;    .     . 

1 

PROBLEMS 


497 


16.10 


At  a  certain  point  in  the  assembly  process,  TV  sets  are  subjected  to 
a  critical  inspection.  The  following  data  resulted  from  the  inspection 
of  25  randomly  selected  sets.  Plot  and  interpret  the  appropriate  con 
trol  chart. 


Set  No. 

Number  of  Defects 
per  Set 

Set  No. 

Number  of  Defects 
per  Set 

1  

12 

14 

7 

2  

11 

15 

4 

3  

13 

16 

9 

4  

17 

17 

15 

5  

7 

18 

12 

6  

10 

19 

11 

7  

9 

20 

9 

8  

8 

21 

4 

9  

13 

22 

4 

10  

15 

23 

10 

11  

18 

24 

12 

12  

9 

25 

4 

13  

14 

16.11  Phonograph  records  were  selected  at  random  times  from  a  production 
line,  Given  the  following  data,  construct  and  interpret  the  appropri 
ate  chart. 


Record  No, 

Number  of  Defects 
per  Record 

Record  No. 

Number  of  Defects 
per  Record 

1.  .    .              .... 

1 

16  

20 

2 

1 

17  

1 

3    

3 

18  

6 

4.  .  .       .           .... 

7 

19  

12 

5 

8 

20  

4 

6 

1 

21  

5 

7   .                    

2 

22  

1 

8 

6 

23  

8 

9 

1 

24  

7 

10   . 

1 

25  

9 

11 

10 

26  ,  

2 

12 

5 

27  

3 

13 

0 

28  

14 

14 

19 

29  

6 

15 

16 

30  

8 

16.12      (a)   Plot  the  OC  curve  for  Example  16.4. 
(6)    Determine  the  AQL  and  RQL  points. 


498  CHAPTER    16,    STATISTICAL   QUALITY    CONTROL 

16.13  (a)    Plot  the  AOQ  curve  for  Example  16.5. 
(6)    Determine  the  AOQL. 

16.14  Discuss  the  basic  purpose  of  acceptance  sampling  and  the  general 
methods  by  which  this  purpose  is  approached. 

16.15  Plot,  on  the  same  graph,  the  following  OC  curves: 
(a)    7^  =  500,  n  =  5Q,  a  =  0 

(6)    AT 
(c)    AT 

16.16  Plot,  on  the  same  graph,  the  following  OC  curves: 
(a)    AT  =  500,  ^  =  25,  a  =  l 

(6)    AT  =  500,  7i  =  50,  a  =  l 
(c)    N  =  500,  n  =  100,  a  =  1 

16.17  For  the  double  sampling  plan  specified  by  AT —  500,  ni  =  50,  n2  =  100, 
ai  =  5,  n  =  14,  and  a2  =  13,  determine  and  plot  the  OC,  AOQ,  and  ASN 
functions.  Find  the  value  of  the  AOQL. 

16.18  The  following  truncated  sequential  sampling  plan  is  proposed: 
(a)   If  the  first  item  is  defective,  reject  the  lot; 

(6)  If  the  first  item  is  nondefective  and  there  is  no  more  than  one 
defective  up  to  and  including  the  tenth  item,  accept  the  lot; 

(c)  If  the  first  item  is  nondefective  and  there  are  two  defectives 
prior  to  or  including  the  tenth  item,  reject  the  lot. 

Determine  the  OC  and  ASN  functions. 

References  and  Further  Reading 

1.  American  Society  for  Testing  Materials.  ASTM  Manual  on  Quality  Control 
of  Materials.  Special  Technical  Publication  15-C,  Philadelphia,  Pa.,  1951. 

2.  American  Standards  Association,  Inc.,  Guide  for  Quality  Control.  Zl.  1-1958, 
New  York,  1958. 

3     m    Control  Chart  Method  of  Analyzing  Data.  Zl. 2-1958.  New  York,  1958. 

4.  .  Control    Chart    Method    of    Controlling    Quality    During    Production. 

Zl.3-1958,  New  York,  1958. 

5.  Bazovsky,  I.  Reliability  Theory  and  Practice.  Prentice-Hall,  Inc.,  Englewood 
Cliffs,  N.J.,  1961. 

6.  Bowker,  A.  H.,  and  Lieberman,   G.  J.  Engineering  Statistics.  Prentice-Hall, 
Inc.,  Englewood  Cliffs,  N.J.,  1959. 

7.  Burr,   I.  W.  Engineering  Statistics  and  Quality  Control.  McGraw-Hill  Book 
Company,  Inc.,  New  York,  1953. 

8.  Chorafas,    D.    N.   Statistical  Processes  and  Reliability  Engineering.   D.    Van 
Nostrand  Company,  Inc.,  New  York,  1960. 

9.  Clark,  C.  R.,  and  Koopmans,  L.  H.  Graphs  of  the  Hyper  geometric  O.C.  and 
A.O.Q.  Functions  for  Lot  Sizes  10  to  %&5.  Sandia  Corporation  Monograph 
SCR-121,  Sandia  Corp.,  Albuquerque,  N.  Mex.,  Sept.,  1959. 

10.  Crow,  E.  L.,   Davis,  F.  A.,  and  Maxfield,   M.  W.  Statistics  Manual.  Dover 
Publications,  Inc.,  New  York,  1960. 

11.  Dodge,  H.  F.  A  sampling  inspection  plan  for  continuous  production.  Ann. 
Math.  Stat.,  14:264,  1943. 

12.  ,  and  Romig,   H.  G.  Sampling  Inspection  Tables:  Single  and  Double 

Sampling.  Second  Ed.  John  Wiley  and  Sons,  Inc.,  New  York,  1959. 

13.  Dummer,  G.  W.  A.,  and  Griffin,  N.  Electronic  Equipment  Reliability.  John 
Wiley  and  Sons,  Inc.,  New  York,  1960. 

14.  Duncan,  A.  J.  Quality  Control  and  Industrial  Statistics.  Revised  Ed.  Richard 
D.  Irwin,  Inc.,  Homewood,  111.,  1959. 

15.  Feigenbaum,    A.   V.    Total   Quality    Control:    Engineering  and   Management. 
McGraw-Hill  Book  Company,  Inc.,  New  York,  1961. 


REFERENCES    AND    FURTHER    READING  499 


;  £c.   S^?^   ^^    ^^   **'    McGraw-Hill   Book 
' 


f          '  J'r"-  Mc^fee>'N-  J.,  Ryerson,  C.  M.,  and  Zwerling,  S.  (editors) 
ew  York    1Q™™0   ^   SeC°nd   Ed'   Instit^   of  Radio   Engineers,   Inc.," 


20"    il  °7d'  ?•  K;.'  an£  Lipow,  M.  Reliability:  Management,  Methods  and  Mathe- 

matics   Prentice-Hall,  Inc.,  Englewood  Cliffs,  N.J  ,  1962 
21.    Shewhart     W     A.    Economic    Control    of   Quality    of  Manufactured   Product. 

JJ.  Van  Nostrand  Company,  Inc.,  New  York,  1931 
22'       .      Q     •  Statistical  Method  from  the  Viewpoint  of  Quality  Control.  The  Gradu- 

ate  fachool,  The  Department  of  Agriculture,  Washington,  D.C.,  1939. 

lor,nin^:,  NetWY?rk,Z'T9eIl'.  ^"^  °f  Statistical  ^ethode.  John  Wiley  and 
24.  Statistical  Research  Group,  Columbia  University.  Selected  Techniques  of 
Stato*ttcalA.nalyata.  (Edited  by  C.  Eisenhart,  M.  W.  Hastay,  and  W.  A. 
Walhs).  McGraw-Hill  Book  Company,  Inc.,  New  York  1947 
Tvr  +  „•  Sam^in^  Z.nsPec^.on-  (Edited  by  H.  A.  Freeman,  M.  Friedman,  F. 
Mosteller,  and  W.  A.  Walhs.)  McGraw-Hill  Book  Company,  Inc.,  New  York, 

J.     4 


.  . 

26.    United  States  Government,    Department  of   Defense.  Sampling  Procedures 
and  Tables  for  Inspection  by   Variables  for  Percent  Defective    MIL-STD-414 
Washington,   D.C.,   1957.  *  ' 

27  '  ~~.  -  ~  •  Multi-Level  Continuous  Sampling  Procedures  and  Tables  for  Inspec- 
tzon  by  Attrzbutes.  Inspection  and  Quality  Control  Handbook  (Interim) 
H-106,  Washington,  D.C.,  1958. 

28-  -  1  -  •  Mathematical  and  Statistical  Principles  Underlying  Military  Stand 
ard  414.  Technical  Report,  Washington,  D.C.,  1958. 

29  '  ~^,  -  '  SamPlin9  Procedures  and  Tables  for  Inspection  by  Attributes.  MIL- 
STD-105C,  Washington,  D.C.,  1961. 

30.  -  .  Statistical  Procedures  for  Determining  Validity  of  Suppliers'  Attributes 
Inspection.    Quality    Control    and    Reliability    Handbook    (Interim)    H-109 
Washington,  D.C.,  1960. 

31.  Wald,  A.  Sequential  Analysis.  John  Wiley  and  Sons,  Inc.,  New  York,  1947. 

32.  ~-  -  ,  and  Wolfowitz,  J.  Sampling  inspection  plans  for  continuous  produc 
tion   which  insure  a  prescribed  limit  on  the  outgoing  quality.  Ann.   Math 
Stat.,  16:30,  1945. 

33.  Western     Electric     Company,     Inc.,     Statistical    Quality    Control    Handbook. 
Second  Ed.  Western  Ellectric  Company,  Inc.,  New  York,  1958. 

34.  Wiesen,   J.    M.    Extension   of  Existing    Tables  of  Attributes   Sampling  Plans. 
Sandia  Corporation  Technical  Memorandum  SCTM42-58(12).,  Sandia  Corp., 
Albuquerque,  1ST.  Mex.,  Feb.,  1958. 


CH  APTE  R    17 

SOME  OTHER  TECHNIQUES  AND 
APPLICATIONS 

IN  THE  PRECEDING  CHAPTERS  those  techniques  most  frequently  em 
ployed  by  users  of  statistical  methods  have  been  presented  as  integral 
parts  of  a  complete  discipline.  In  this  chapter  a  few  special  techniques 
will  be  discussed  and  one  or  two  interesting  applications  of  probability 
and/or  statistics  illustrated.  It  is  hoped  that  these  items  will  prove 
useful  to  many  of  the  readers  and  they  they,  the  items,  will  stimulate 
some  of  you  to  search  out  new  and  exciting  applications  in  your  own 
area  of  specialization. 

17-1      SOME  PSEUDO  t  STATISTICS 

Many  times,  the  researcher  may  feel  that  the  time  and  effort  in 
volved  in  the  calculation  of  s,  the  sample  standard  deviation,  is  too 
great  for  the  benefit  derived  therefrom.  Thus,  it  is  not  surprising  that 
techniques  have  been  devised  which  use  R,  the  sample  range,  in  place 
of  s.  One  of  these  techniques  involves  the  use  of  a  pseudo  t  statistic 
defined  by 

T,  -  &  -  ti/R.  (17.1) 

If  we  are  willing  to  assume  random  sampling  from  a  normal  popula 
tion,  tests  of  hypotheses  concerning  ^  or  confidence  interval  estima 
tion  of  ju  may  be  performed  by  considering  the  sampling  distribution  of 
T-L  and  utilizing  the  values  recorded  in  Table  1  of  Appendix  17. 

Example  17.1 

Consider  again  the  data  and  the  hypothesis  of  Example  7.8.  Using 
Equation  (17.1),  we  obtain  r1  =  (1267  — 1260) /8  =0.875.  Since  this  cal 
culated  value  of  rx  exceeds  the  critical  value  (for  n  =  4),  namely, 
ri(.975)  =  -717,  the  hypothesis  that  /x  =  1260  is  rejected.  (NOTE:  This 
is  in  agreement  with  the  conclusion  reached  in  Example  7.8) 

Example  17.2 

Utilizing  the  same  data  as  in  the  preceding  example,  a  99  per  cent 
confidence  interval  for  p.  may  be  obtained  as  follows: 

L  =  X  -  TI(  995yR  «  1267  —  1.316(8)  =  1256.47 
U  *  T  +  r1(995yR  *  1267  +  1.316(8)  =  1277.53. 

An  even  simpler  statistic  which  may  be  used  as  a  substitute  for 
"Student's"  t  is 

[5003 


17.3       EVOLUTIONARY    OPERATION  501 

(17.2) 

Critical  values  of  this  statistic  are  given  in  Table  3  of  Appendix  17. 
Because  of  the  similarity  (in  use)  of  this  statistic  to  the  one  discussed 
in  the  preceding  paragraph,  no  numerical  examples  will  be  presented. 
When  dealing  with  the  difference  between  two  sample  means,  a  third 
pseudo  t  statistic  may  be  utilized,  namely, 

rd  =  C^x  -  TsVCRx  +  £2).  (17.3) 

Critical  values  of  rd  are  tabulated,  for  samples  of  equal  size,  in  Table  2 
of  Appendix  17. 

Example  17.3 

Consider  the  data  and  hypothesis  of  Example  7.19.  Using  Equation 
(17.3),  we  obtain  rd=  (4  —  7)/(7+7)  =  —3/14=  —0.214.  Since  the  cal 
culated  value  to  rd  lies  between  —  Td(.96)  =  —.24:6  and  rd(^^^  =  .246,  we 
are  unable  to  reject  the  hypothesis  that  Mi  =M2.  (NOTE:  This  agrees  with 
the  conclusion  reached  in  Example  7.19,) 

17.2     A  PSEUDO  F  STATISTIC 

As  might  be  expected,  it  is  also  possible  to  utilize  the  range  in  place 
of  the  standard  deviation  to  provide  a  quick  substitute  for  the  familiar 
F  statistic.  The  suggested  statistic  is 

=  the  ratio  of  the  two  sample  ranges,  (17.4) 


and  critical  values  are  tabulated,  for  certain  selected  sample  sizes,  in 
Table  4  of  Appendix  17.  As  with  F,  the  critical  values  for  the  lower  end 
of  the  distribution  of  Ri/R*  may  be  found  by  interchanging  ni  and  n2 
(the  two  sample  sizes)  and  calculating  the  reciprocals  of  the  tabulated 
values.  Because  of  the  simplicity  of  this  test  and  its  similarity  to  those 
discussed  in  the  preceding  section,  no  numerical  examples  will  be  given. 

17.3      EVOLUTIONARY  OPERATION 

In  Section  13.6  the  reader  was  exposed  to,  but  not  indoctrinated  in, 
the  subject  known  as  "response  surface  techniques."  In  the  present 
section,  a  related  technique  wUl  be  introduced.  This  technique,  known 
as  evolutionary  operation  (or  EVOP),  is  an  application  of  the  concepts 
of  response  surface  methodology  to  the  problem  of  improving  the  per 
formance  of  industrial  processes.  Consequently,  this  technique  should 
be  of  great  interest  to  those  persons  concerned  with  production  proc- 


Q^CI 

Because  no  detailed  discussion  of  response  surface  methodology  was 
undertaken  in  Chapter  13,  a  full  description  of  EVOP  cannot  be  given 
here.  However,  in  capsule  form,  the  important  elements  of  the  tech 
nique  are: 

(1)   It  is  a  method  of  process  operation  which  includes  a  built-in 
procedure  for  improving  productivity. 


502  CHAPTER    17,    OTHER   TECHNIQUES    AND    APPLICATIONS 

(2)  It  uses  some  relatively  simple  statistical  concepts. 

(3)  It  is  run  during  normal  routine  production  by  plant  personnel. 

(4)  The  basic  philosophy  of  EVOP  is:  A  process  should  be  run 
not  only  to  produce  product  but  also  to  provide  information  on 
how  to  improve  the  process  and/or  product. 

(5)  Through  the  planned  introduction  of  minor  variants  into  the 
process,    the    customary    "static"    operating    conditions    are 
made  "dynamic"  in  nature. 

(6)  Utilizing  the  elementary  principles  of  response  surface  meth 
odology  and  making  small  changes  (in  a  prescribed  fashion) 
in  the  "controlled"  variables  of  the  process,  the  effects  of  the 
forced  changes  can  be  assessed. 

(7)  If  a  simple  pattern  of  operating  conditions  is  employed  (e.g., 
a  22  factorial  plus  a  center  point)1  and  if  the  operation  of  the 
process  under  each  of  these  conditions  is  termed  a  "cycle," 
the  running  of  several  cycles  will  yield  sufficient  information  to 
permit  a  judgment  to  be  made  as  to  what  is  a  better  nominal 
set  of  operating  conditions. 

(8)  Constant   repetition  of  this  program  will  lead  to   continual 
improvement  of  the  process. 

(9)  Two  important  items  in  an  EVOP  program  are : 

(a)  All  data  should  be  prominently  displayed  on  an  Infor 
mation  Board. 

(b)  An   Evolutionary   Operation   Committee    (composed   of   re 
search,   development,   and  production  personnel)    should 
make  periodic  reviews  of  the  EVOP  program  if  the  maxi 
mum  benefit  is  to  be  derived  from  this  approach. 

Inasmuch  as  the  preceding  remarks  are  only  a  skeletal  description 
of  the  EVOP  technique,  those  persons  desiring  more  details  must  con 
sult  other  sources.  In  particular,  the  publications  by  Box  (3)  and  Box 
and  Hunter  (4)  are  especially  recommended. 

17.4     TOLERANCES 

In  Section  5.14  some  remarks  were  made  with  regard  to  the  distri 
bution  of  a  linear  combination  of  random  variables.  At  this  time,  it  is 
appropriate  to  consider  a  specific  application  of  those  remarks,,  namely,, 
to  the  subject  of  tolerances. 

Before  embarking  on  this  discussion,  a  distinction  must  be  made 
between  specification  limits  (that  is,  a  nominal  value  plus  and  minus 
certain  engineering  tolerances)  and  natural,  or  statistical,  tolerances.  In 
general,  specification  limits  are  set  by  the  designer  as  a  statement  of 
his  requirements  with  respect  to  a  certain  dimension.  Thus,  in  many 
cases,  the  specification  limits  have  little  connection  with  production 
capabilities.  On  the  other  hand,  statistical  tolerance  limits  reflect  the 

1  This  center  point  will  usually  be  that  nominal  set  of  operating  conditions 
specified  in  the  engineering  drawings  or  production  manual. 


17.4      TOLERANCES  503 

capabilities  of  the  process  producing  the  dimensions  in  question.  (See 
Sections  6.11,  6.12,  and  6.13).  Therefore,  from  the  point  of  view  of 
managerial  decision  making,  the  subject  of  statistical  tolerances  is  of 
great  importance,  for  it  bears  directly  on  the  success  or  failure  of  the 
company's  product. 

In  what  follows,  we  are  interested  only  in  general  concepts  and  a 
method  of  approach  that  may  prove  helpful  in  the  design  of  complex 
equipment.  Accordingly,  our  attention  will  be  confined  to  those  cases 
in  which  the  several  variables  (dimensions)  are  normally  and  inde 
pendently  distributed  with  known  means  and  variances.  Restrictive 
as  these  assumptions  may  be,  they  will  not  seriously  limit  our  presen 
tation.  [NOTE  :  If  other  distributions  must  be  used,  the  same  concepts 
will  apply.  For  example,  see  Breipohl  (5).] 

Two  cases  will  be  examined:  (1)  linear  combinations  of  the  variables 
and  (2)  nonlinear  combinations  of  the  variables. 

Case  I:  Linear  Combinations  of  Independent  Random 
Variables 

This  case  was  covered  in  Section  5.14  where  it  was  noted  that  if 

U  =  Jb  a^-,  (17.5) 

1=1 

then 

n 

MCJ  =  ]C  #*Mi  (17.6) 

t=i 

and 

?<^  (17.7) 


where  M*  and  cr?  are  the  mean  and  variance,  respectively,  of 
Xi  (i=  17  -  •  •  ,  n).  If,  as  is  true  in  many  applications,  each  a»=  1,  then 
Equations  (17.6)  and  (17.7)  reduce  to 


i*  (17.8) 

«—  i 

and 

*.  (17.9) 


Because  the  above  results  are  used  in  a  variety  of  ways,  a  complete 
discussion  of  each  different  situation  is  not  planned.  However,  some 
typical  examples  will  be  presented  to  acquaint  the  reader  with  a  few  of 
the  possible  applications. 


5O4  CHAPTER    17,    OTHER   TECHNIQUES    AND    APPLICATIONS 

Example  17.4 

Consider  a  simple  addition  of  components  such  that  the  dimension  of 
the  assembly  (F)  is  the  sum  of  the  dimensions  of  the  individual  com 
ponents,  that  is,  F  =  53?.  i  Xi.  If  there  are  two  components,  Xi  and  X%, 
with  means  0.500  and  0.410  and  with  standard  deviations  0.008  and 
0.006,  respectively,  then  ^r  =  0.910  and  <rY  =  [(.008)2-K.006)2]1/2 
=  0.01.  If  Xi  and  X?  are  assumed  to  be  normally  distributed,  then  F 
is  normally  distributed  and  the  distribution  (of  F)  may  be  compared 
with  the  specification  limits  (for  F)  to  obtain  the  expected  percentage 
of  defective  assemblies. 

Example  17.5 

Consider  an  assembly  consisting  of  five  components  for  which 
F=  ]C*-i  Xi  represents  the  simple  addition  of  the  appropriate  dimen 
sions.  If  the  five  dimensions  are  normally  and  independently  dis 
tributed  with  means  0.500,  0.410,  0.200,  0.700,  and  0.210,  respectively, 
and  if  they  may  be  assumed  to  have  a  common  variance,  cr2,  how  large 
can  a  be  if  Mr  ±3cr:r  =  2-020±  0.030  units?  Using  Equation  (17.9),  it  may 
be  verified  that  cry  =  5cr2.  Therefore,  Scr^  =  3<r  V5  =  0.030,  which  means 
that  5a2  =  (0.01)2  =  0.0001,  and  thus  <r2  =  0.00002.  Consequently,  the 
maximum  allowable  value  of  cr  is  (0.00002)1/2  =  0.0045. 

Example  17.6 

Another  illustration  of  a  linear  combination  of  dimensions  is  the  clear 
ance  between  a  shaft  and  a  bearing.  Let  us  assume  that  shafts  are  mass 
produced  such  that  the  outside  diameters  are  normally  and  independ 
ently  distributed  with  jus  =  1.05  inches  and  with  standard  deviation  <rs. 
Let  us  also  assume  that  bearings  are  produced  such  that  the  inside 
diameters  are  normally  and  independently  distributed  with  jLtB  =  1.06 
inches  and  with  standard  deviation  crB  ==0.001  inch.  If  production  of  the 
shafts  is  to  be  controlled  so  that  no  more  than  5  per  cent  of  randomly 
mated  shafts  and  bearings  will  exhibit  interference,  what  is  the  maxi 
mum  allowable  value  of  crs? 

To  answer  this  question,  consider  F==JB  —  S.  Using  Equations  (17.6) 
and  (17.7),  it  may  be  verified  that  Atr  =  1.06  —  1.05=0.01  inch  and  that 
°>=  [(0.001)2  +  cr|]1/2  inch.  Now,  interference  will  occur  when  F<0. 
Thus,  it  is  appropriate  to  consider 

p{  y  <  0}  =  P{Z  <  (0  -  0.01)/<rF}  =  0.05 

where  Z==  (F— JUF)/O-F.  On  consulting  Appendix  3,  it  is  found  that 
—  0.01/oy=  —1.645  and,  consequently,  <r^  =  0.00000  l+<r|  =  (1/164.5)2 
=  0.000037.  As  a  result,  it  is  seen  that  the  maximum  <rs  is  0.006. 

It  is  hoped  that  the  preceding  examples  will  be  sufficient  to  indicate 
the  nature  and  scope  of  the  theory  of  statistical  tolerances  when  dealing 
with  linear  combinations  of  independent  random  variables.  For  those 
who  wish  to  investigate  the  subject  in  greater  depth,  more  details  and 
examples  are  available  in  Bowker  and  Lieberman  (2)  and  Breipohl  (5). 


17.4      TOLERANCES  5O5 


Case  II  :  Nonlinear  Combinations  of  Independent  Random 
Variables 

When  variables  are  combined  in  a  nonlinear  fashion,  the  tolerance 
problem  is  usually  much  more  difficult.  The  reason  for  this  difficulty  is 
that,  in  general,  it  is  not  easy  to  determine  the  distribution  of  a  non 
linear  combination  of  random  variables.  However,  if  a  linear  approxi 
mation  is  acceptable,  the  problem  may  be  handled  by  expanding  the 
function  in  a  Taylor  series  about  the  mean  values.  That  is,  if 


F  =  <K^i,  -  •  •  ,  -Yn),  (17.10) 

it  may  be  shown  that 


Y  = 

*—  i 


+  terms  of  higher  order.  (17.11) 

Then  if  the  "terms  of  higher  order"  are  neglected,  it  may  be  verified 
that 


•  •  •  ,M»)  (17.12) 

and 


r 


l,  '  '  -,Mn-I 


(17.13) 


where  /z*  and  cr\  are  the  mean  and  variance,  respectively,  of 
Xi  (i—  1,  •  •  •  ,  n).  As  a  consequence,  Equations  (17.12)  and  (17.  IS) 
may  be  used  to  obtain  approximate  tolerance  limits  for  Y, 

Example  17.7 

Suppose  that  Y  =  X^X^X^  Expanding  this  function  in  a  Taylor  series 
about  pi,  JJLZ,  and  ^3,  we  obtain 

Y  = 


If  the  X  i  are  normally  and  independently  distributed  with  mean  M 
variance  erf,  then  Y  is  approximately  normally  distributed  with  mean 
JAY  ss  /ziMuMs  and  variance  <r|r  =  (MSMS)  2cr?  +  (yuiMs)  2o|  +  (^1^2)  2oi-  These 
expressions  may  then  be  used  to  investigate  the  natural  tolerances  of  Y. 

Example  17.8 

Consider  an  electrical  circuit  consisting  of  two  resistors  connected  in 
parallel.  For  such  a  circuit,  Rc^RiRz/tRi+Rz)  where  Rt(i  =  I,  2) 
signify  the  resistances  of  the  two  resistors  connected  in  parallel  and  Re 
is  the  circuit  resistance.  Using  the  Taylor  series  approximation,  it  is 
easily  verified  that 


Ma)    and  <r*    «  | 

where  the  subscripts  conform  with  the  notation  used  above  for  the 


5O6  CHAPTER    17,    OTHER    TECHNIQUES    AND    APPLICATIONS 

R- values.  These  equations  may  then  be  used  to  study  the  relationships 
among  the  tolerances  of  the  resistors  and  the  tolerances  of  the  circuit, 

17.5     THE   ESTIMATION   OF  SYSTEM    RELIABILITY 

The  general  problem  of  estimating  the  reliability  of  systems  has 
received  considerable  attention  in  recent  years.  Stated  briefly,  the 
problem  is  as  follows:  Given  probabilities  for  the  successful  operation 
of  the  various  components  utilized  in  a  system,  estimate  the  probabil 
ity  that  the  system  will  operate  successfully.  Before  attempting  to 
present  the  solution  of  a  problem  such  as  posed  in  the  preceding  para 
graph,  it  will  be  wise  to  adopt  some  standard  notation.  Thus,  in  what 
follows,  we  shall  use: 

Pi  =  the  probability  that  the  ith  component  in  the  system  will  oper 
ate  successfully, 

gi  =  the  probability  that  the  ith  component  in  the  system  will  fail 
to  operate  successfully  =  1  —  p^ 

JP  =  the  probability  that  the  system  will  operate  successfully,  and 

Q  =  the  probability  that  the  system  will  fail  to  operate  successfully 
=  1-P. 

Given  this  notation  and  assuming  that  n  components  are  utilized  in  the 
system,  we  may  write 

P  =  f(pi,  •  •  •  ,  pj  (17.14) 

where  the  form  of  the  function  will  depend  on  the  nature  of  the  system- 
If  the  system  under  consideration  is  either  a  simple  series  system  (in 
which  all  components  must  work  for  the  system  to  succeed)  or  a  paral 
lel  sytem  (in  which  at  least  one  of  the  components  must  work  for  the 
system  to  succeed),  and  if  the  various  components  are  statistically 
independent  in  their  operation,  the  functional  form  is  easy  to  deter 
mine  and  the  basic  equations  are  as  follows: 
Series  System 


Q  =  1-  IIP*=  1-  11(1  -  5*)  (17.16) 

Parallel  System 


=  i  -  n  a  -  ^). 

i=i  i»=i 

If  the  system  involves  both  series  and  parallel  features,  an  equivalent 


17.5       ESTIMATION    OF   SYSTEM    RELIABILITY  5O7 

system  (or  circuit)  can  always  be  found  that  will  permit  utilization  of 
the  preceding  equations. 

Example  17.9 

Consider  a  system,  consisting  of  four  components  connected  in 
series,  in  which  all  components  must  operate  properly  if  the  system  is 
to  function  properly.  If  the  respective  failure  probabilities  of  the  four 
components  are  0.02,  0.03,  0.05,  and  0.02,  then  P  =  (0.98) (0.97) (0.95) 
(0.98)  and  Q  =  l— P. 

Example  17.10 

Consider  a  system,  consisting  of  four  components  connected  in 
parallel,  in  which  at  least  one  of  the  components  must  operate  properly 
if  the  system  is  to  function  properly.  If  the  respective  failure  proba 
bilities  of  the  four  components  are  0.05,  0.10,  0.02,  £nd  0.001,  then 
Q  =  (0.05)(0.10)(0.02)(0.001)  and  P  =  l— Q. 

Brief  as  the  preceding  treatment  has  been,  I  hope  that  it  has  been 
sufficient  to  acquaint  the  reader  with  the  nature  of  the  problem  and 
with  the  method  of  solution.  Some  of  the  problems  at  the  end  of  the 
chapter  will  require  the  extension  of  the  basic  principles  to  more  com 
plex  situations  while  others  will  introduce  certain  approximations  that 
are  often  used  by  reliability  analysts. 

Before  leaving  this  topic,  I  would  be  derelict  in  my  duty  if  I  did  not 
call  a  number  of  specific  points  to  your  attention.  These  are: 

(1)  The  assumption  of  statistical  independence  is  frequently  sub 
ject  to  question. 

(2)  Rarely  does  the  analyst  know  the  true  values  of  the  p*  and  q^ 
This  means  that  he  is  really  using  p*  and  fc=I  —  pi  and  thus 
p  =/(pi,  -  -  -  ,  pn)  and  Q  =  1  —  P  are  actually  point  estimates 
of  the  unknown  parameters,  P  and  Q. 

(3)  While  confidence  limits  for  the  true  pf  and  q*  are  easily  ob 
tained  (see  Chapter  6),  no  satisfactory  methods  have  yet  been 
developed    for    providing    confidence    limits    for    P    and    Q. 
(NOTE:  A  few  special  cases  have  been  solved  but  the  general 
case  is  still  under  investigation.) 

(4)  Frequently,  the  analyst  must  consider  more  than  just  success 
or  failure.  For  example,  in  nuclear  weaponry  the  possibility  of 
premature  operation  must  also  be  considered.  Thus,  for  each 
device,  and  for  the  system  as  a  whole,  we  now  have  three 
probabilities  to   contend  with.   Apart  from  the  added   com 
plexity  of  the  mathematics  involved,  this  can  lead  to  great 
difficulties  in  the  (logical)  analysis  of  the  system. 

(5)  In  many  applications,  the  probabilities  p*  and  g*  will  be  func 
tions  of  time,  that  is,  the  reliability  analyst  will  be  dealing 
with  pi(0  and  g^t)  where  t  represents  operating  time  or  age. 
When  such  is  the  case,  the  problems  of  analysis  are  compli 
cated  by  such  phenomena  as  early  failures  and  wearout. 


508  CHAPTER    17,    OTHER    TECHNIQUES    AND    APPLICATIONS 

The  preceding  are  just  a  few  of  the  points  that  plague  a  reliability  ana 
lyst  as  he  goes  about  his  daily  work.  I  am  certain  that  you  could,  with  a 
little  reflection,  add  many  other  items  to  the  above  list.  However,  such 
a  list  would  be  out  of  place  in  this  book.  The  items  mentioned,  though, 
are  distinctly  statistical  in  nature  and  thus  it  seems  appropriate  to 
call  them  to  your  attention  at  this  time. 

Problems 

17.1        Using  first  rt  and  then  T2  (as  defined  in  Section  17.1),   rework  the 
following  problems: 


(a)   6.1 

(/)    7.1 

(j)    7.6 

(6)    6.5 

(?)   7.2 

(fc)   7.29 

(c)    6.20 

/  7,  \      TO 
(A)      7.O 

(1)    7.32 

(<2)   6.22 

(t)    7.5 

(m)    7.33 

(e)    6.23 

17.2 

Using  rd 

(as  defined  in  Section 

17.1),  rework  the  following  problems: 

(a)   7.25 

(e)    7.36 

(6)   7.27 

C/)    7.37 

(c)    7.30 

(0)   7.39 

(d)  7.31 

(A)  7.40 

17.3 

TJsing  the   pseudo  F  statistic 

denned  in  Section   17.2,  rework  the 

following  problems: 

(a)   6.24  (d)  7.46 

(6)    6.26  (e)    7.47 

(<?)    7.45 

17.4  Rework  Example  17,5  assuming  that  /-tjr±  5crF  =  2.  020  ±0.030  units. 

17.5  Rework  Example  17.6  assuming  that  no  more  than  1  per  cent  of 
randomly  mated  bearings  and  shafts  must  exhibit  interference. 

17.6  Evaluate    the   results    of    Example    17.7   if  jui  =  40,    ^2  =  0.5,    ;u3  =  3, 
3cri=0.5,  3cr2  =  0.005,  and  3cr3  =  0.06. 

17.7  If  Y=*X  cos  6+UVW  sin  9,  where  X}   U,   V,  W,  and  9  are  random 
variables,  use  a  Taylor's  expansion  about  the  means  to  determine 
approximate  expressions  for  the  mean  and  variance  of  Y* 

17.8  Assuming  that  each  random  variable  is  normally  distributed  with  a 
mean  equal  to  the  nominal  value  and  with  a  standard  deviation  equal 
to  one-third  of  the   (one-sided)   engineering  tolerance,  evaluate  the 
results  determined  in  Problem  17.7,  for  the  following  specifications: 


?7=40±0.5 
F  =  0.5±l  per  cent 
W  =  3±2  per  cent 
0  =  60  ±0.25  degrees. 

17.9  Two  resistors  are  assembled  in  series.  Each  is  nominally  rated  at  20 
ohms.  The  resistors  are  known  to  be  normally  distributed  about  the 
nominal  value  with  a  standard  deviation  of  1.5  ohms.  What  is  the 
mean  and  standard  deviation  of  the  circuit  resistance? 

17.10  Rework  Problem  17.9  if  the  resistors  are  uniformly  distributed  over 
the  interval  18  to  22  ohms. 

17.11  If  two  mating  parts,  X  and  F,  are  each  normally  distributed  with 


PROBLEMS  5Q9 

*r==2.04  inches  and  standard  deviations 
)2  inch,  what  is  the  probability  of  inter- 

17.12  With  reference  to  Problem  17.11,  what  should^  be  (assuming  every- 
thing  else  is  unchanged)  if  the  probability  o^intekrenTe  fs  to  be 

17.13  For  each  of  the  following  cases,  use   a  Taylor's  series  to  find  the 
approximate  mean  and  variance  of  the  dependent  variable - 

(a)    R      "^ 
(6)    AC 
(<0    R" 


(e)    C  = 

17.14  Fora  system  involving  four  components  (A,  B,  C,  and  JD)  for  which 
PJL  -0.9,  PJ8-0.8,  pc  =  0.9,  and  ^=0.9,  determine  P  under  the  as 
sumption  of  mutual  independence  and  given  that  A  and  B  are  con 
nected  in  parallel  (branch  I)  and  that  this  branch  is  then  connected 

17.15  Consider  an  electrical  circuit  consisting  of  two  subcircuits,  the  first 
of  which  involves  the  components  Xl}  X2,  Xz,  and  X4  in  parallel,  and 
the  second  of  which  involves  the  components  X5  and  Xe  in  parallel. 
11   the   two   subscircuits   are   connected  in   series   and   mutual  inde 
pendence  can  be  assumed,   determine   Q  if  gi=0.10,   g2  =  0.05    and 
3s  =  £4  =  £5  =  #6  =  0.02.  *  *     ' 

17.16  An  equipment  consists  of  three  components  (£>,  E,  and  F)  connected 
in  series.  If  the  reliabilities  of  the  three  components  are  0,92,  0.95, 
and  0.96,  what  is  the  reliability  of  the  equipment?  State  all  assump 
tions  made  in  achieving  your  answer. 

17.17  An  equipment  consists  of  three  components  (D,  E,  and  F)  connected 
in  parallel.  If  the  reliabilities  of  the  three  components  are  0.92,  0.95, 
and  0,96,  what  is  the  reliability  of  the  equipment?  State  all  asssumpl 
tions  made  in  achieving  your  answer. 

17.18  An  equipment  consists  of  100  parts,  of  which  20  parts  are  tubes 
connected  functionally  in  series  (branch  A).  This  branch  is  in  turn 
connected   in   series   to  two  parallel   branches   of    60    and   20   parts 
(branches  B  and  C).  The  parts  which  comprise  each  of  these  branches 
are  connected  functionally  in  series.  The  reliability  of  each  tube  is 
0.95,   while  the  geometric  mean  reliability  of  branch  B  is  0.93  and 
the  geometric  mean  reliability  of  branch  C  is  0.96.  Draw  a  simplified 
equipment  diagram  and  determine  the  reliability  of  the  equipment. 
State  all  assumptions  made  in  achieving  your  answer. 

17.19  When  reliability  is   a  function  of  time,   it  is   common  practice  to 
assume  the  validity  of  the  exponential  probability  density  function 
as  a  failure   distribution.  That  is,   it  is  common  to  use  /(£)  ==\e~x* 
as  the  basic  failure  distribution.  In  such  a  case,  X  is  known  as  the 
hazard  rate   (or,  in  loose  language,  the  failure  rate)   and  m  — 1/X  is 
called  the  mean  time  between  failure  (MTBF).  Under  the  preceding 
conditions,  the  reliability  of  a  component  to  time  t  is 


Utilizing  the  above  information,  and  assuming  that  the  ith  compo- 


510  CHAPTER    17,    OTHER   TECHNIQUES   AND   APPLICATIONS 

nent  has  a  constant  hazard  rate\»(£  =  l,  •  •  •  ,  ri),  express  Equations 
(17.15)  through  (17.18)  in  terms  of  the  A's  and  *. 

17.20  Simplify  the  answer  to  the  preceding  problem  on  the  assumption  that 
all  the  \i  are  equal. 

17.21  If  /(<)=Xe-x«;  A>0,  Z>0;  what  effect  does  doubling  the  value  of  X 
have  on: 

(a)   the  MTBF,  and  (6)  R(f). 

17.22  Assume  an  equipment  consists  of  two  components  for  which  the 
failure  rates  are  Ai  =  .001  and  \2~.002,  respectively.  Calculate  the 
equipment  reliability  for  2  =  100  for  the  following  two  cases:  (a)  series 
connection,  (6)  parallel  connection. 

17.23  Show  that,  if  n  =  4,  Equation  (17.16)  can  be  approximated  by  either  4g 
or  4q  —  6g2.  (NOTE :  The  reliability  analyst  makes  frequent  use  of  such 
approximations,) 

17.24  The  failure  rate  for  a  television  receiver  is  0.02  failures/hour, 
(a)   Calculate  the  mean  time  between  failures. 

(6)   What  is  the  probability  of  such  a  receiver  failing  in  the  first  four 
hours? 

References  and  Further  Reading 

1.  Bazovsky,  I.  Reliability  Theory  and  Practice.  Prentice-Hall,  Inc.,  Englewood 
Cliffs,  N,J.,  1961. 

2.  Bowker,  A.  H.,  and  Lieberman,  G.  J.  Engineering  Statistics.  Prentice-Hall, 
Inc.,  Englewood  Cliffs,  N.J.,  1959. 

3.  Box,  G.  E.  P.  Evolutionary  operation:  A  method  for  increasing  industrial 
productivity.  Applied  Stat.,  6  (No.  2)  :3-23,  1957. 

4    ?  and  Hunter,  J.  S.   Condensed  calculations  for  evolutionary  opera 
tion  programs,  Technometrics,  l(No.  1):  77-95,  Feb.,  1959, 

5.  Breipohl,  A.  M.  A  Statistical  Approach  to  Tolerances.  Sandia  Corporation 
Technical    Memorandum   SCTM    65-60    (14),    Sandia    Corporation,    Albu 
querque,  N.  Mex.,  Mar.,  22,  1960. 

6.  Calabro,    S.    B,.    Reliability   Principles    and   Practices.    McGraw-Hill    Book 
Company,  Inc.,  New  York,  1962. 

7.  Chorafas,    D.    N.   Statistical  Processes  and  Reliability  Engineering.   D.   Van 
Nostrand  Company,  Inc.,  Princeton,  N.J.,  1960. 

8.  Dummer,  G.  W.  A.,  and  Griffin,  N.  Electronic  Equipment  Reliability.  John 
Wiley  and  Sons,  Inc.,  New  York,  1960. 

9.  Feigenbaum,    A.   V.    Total   Quality   Control:   Engineering  and  Management. 
McGraw-Hill  Book  Company,  Inc.,  New  York,  1961. 

10.  Gryna,  F.  M.,  Jr.,  McAfee,  N.  J.,  Ryerson,  C.  M.,  and  Zwerling,  S.  (editors). 
Reliability  Training  Text.  Second  Ed.  Institute  of  Radio  Engineers,  Inc., 
New  York,  1960. 

11.  Link,  R.  F.  The  sampling  distribution  of  the  ratio  of  two  ranges  from  inde 
pendent  samples.  Ann.  Math.  Stat.,  21:112,  1950. 

12.  Lloyd,  D.  K.,  and  Lipow,  M.  Reliability:  Management,  Methods,  and  Mathe 
matics.  Prentice-Hall,  Inc.,  Englewood  Cliffs,  N.J.,  1962. 

13.  Lord,  E.  The  use  of  the  range  in  place  of  the  standard  deviation  in  the  t  test. 
Biometrika,  34:41,  1947, 

14.  Walsh,  J.  E.  On  the  range-midrange  test  and  some  tests  with  bounded  signifi 
cance  levels.  Ann.  Math.  Stat.,  20:257,  1949. 


APPENDIX     1 


A  a  alpha 

B  /3  beta 

F  T  garanaa 

A  8  delta 

E  e  epsilon 

Z  f  zeta 

H  77  eta 

©  (9  theta 

I  L  iota 


K     K 

A      X 
M     fj, 


kappa 

lambda 

mu. 


N  v  nil 

S  4;  xi 

O  o  onxicron 

n  TT  pi 

P  p  rho 

S  cr  sigma 

T  T  tan 

T  u  upsilon 

<^>  <f>  ptii 

X  x.  chi 

^f  \f/  psi 

Q  co  omega 


C51  11 


APPENDIX    21 

CUMULATIVE  POISSON 
DISTRIBUTION23 


10 


0.01 
0.02 

0.990 
0.980 

0.03 

0.970 

0.04 

0.961 

0.999 

0.05 

0.951 

0.999 

0.06 

0.942 

0.998 

0.07 

0.932 

0.998 

0.08 

0.923 

0.997 

0.09 

0.914 

0.996 

0.10 

0.905 

0.995 

0.15 

0.861 

0.990 

0.999 

0.20 

0.819 

0.982 

0.999 

0.25 

0.779 

0.974 

0.998 

0.30 

0.741 

0.963 

0.996 

0.35 

0.705 

0.951 

0.994 

0.40 

0.670 

0.938 

0.992 

0.999 

0,45 

0.638 

0.925 

0.989 

0.999 

0.50 

0.607 

0.910 

0.986 

0.998 

0.55 

0.577 

0.894 

0.982 

0.998 

0.60 
0.65 

0.549 
0.522 

0.878 
0.861 

0.977 
0.972 

0.997 
0.996       0.999 

1  Reprinted  from  E.    C.   Molina,   Poisson's  Exponential  Binomial  Limit,    D. 
Van  Nostrand  Company,  Inc.,  New  York,   1947.  By  permission  of  the  author 
and  publishers. 

2  Entries  in  the  table  are  values  of  F(x)  where 

/?(#)  »  p(c  <  x)  —  £  e~^c/c\ 

c-O 

*  Blank  spaces  to  the  right  of  the  last  entry  in  any  row  of  the  table  may  be 
read  as  1;  blank  spaces  to  the  left  of  the  first  entry  in  any  row  of  the  table  may 
be  read  as  0. 


[5121 


10 


0.70 
0.75 

0.80 
0.85 
0.90 
0.95 
1.00 

1.1 
1.2 
1.3 
1.4 

1.5 

0.497 
0.472 

0.449 
0.427 
0.407 
0.387 
0.368 

0.333 
0.301 
0.273 
0.247 
0.223 

0.844 
0.827 

0.809 
0.791 
0.772 
0.754 
0.736 

0.699 
0.663 
0.627 
0.592 
0.558 

0.966 
0.959 

0.953 
0.945 
0.937 
0.929 
0.920 

0.900 
0.879 
0.857 
0.833 
0.809 

0.994 
0.993 

0.991 
0.989 
0.987 
0.9S4 
0.981 

0.974 
0.966 
0.957 
0.946 
0,934 

0.999 
0.999 

0.999 
0.998 
0.998 
0.997 
0.996 

0.995 
0.992 
0.989 
0.986 
0.981 

0.999 

0.999 
0.998 
0.998 
0.997 
0.996 

0.999 
0.999 

1.6 
1.7 
1.8 
1.9 
2.0 

0.202 
0.183 
0.165 
0.150 
0.135 

0.525 
0.493 
0.463 
0.434 
0.406 

0.783 
0.757 
0.731 
0.704 
0.677 

0.921 
0.907 
0.891 
0.875 
0.857 

0.976 
0.970 
0.964 
0.956 
0.947 

0.994 
0.992 
0.990 
0.987 
0.983 

0.999 
0.998 
0.997 
0.997 
0.995 

0.999 
0.999 
0.999 

2.1 
2.2 
2.3 
2.4 
2.5 
2.6 
2.7 
2.8 
2.9 
3.0 

0.122 
0.111 
0.100 
0.091 
0.082 
0.074 
0.067 
0,061 
0.055 
0.050 

0.380 
0.355 
0.331 
0.308 
0.287 
0.267 
0.249 
0.231 
0.215 
0.199 

0.650 
0.623 
0.596 
0,570 
0.544 
0.518 
0,494 
0.469 
0.446 
0.423 

0.839 
0.819 
0.799 
0.779 
0.758 
0.736 
0.714 
0.692 
0.670 
0.647 

0.938 
0.928 
0.916 
0.904 
0.891 
0.877 
0.863 
0.848 
0.832 
0.815 

0.980 
0.975 
0.970 
0.964 
0.958 
0.951 
0.943 
0.935 
0.926 
0.916 

0.994 
0.993 
0.991 
0.988 
0.986 
0.983 
0.979 
0.976 
0.971 
0.966 

0.999 
0.998 
0.997 
0.997 
0.996 
0.995 
0.993 
0.992 
0.990 
0.988 

0.999 
0.999 
0.999 
0.999 
0.998 
0.998 
0.997 
0.996 

0.999 
0.999 
0.999 
0.999 

0 

1 

2 

3 

4 

5 

6 

7 

8 

9 

10 

3.2 
3.4 
3.6 
3.8 
4.0 

0.041 
0.033 
0.027 
0.022 
0.018 

0.171 
0.147 
0.126 
0,107 
0.092 

0.380 
0.340 
0.303 
0.269 
0  238 

0.603 
0.558 
0.515 
0.473 
0.433 

0.781 
0.744 
0.706 
0.668 
0.629 

0.895 
0.871 
0.844 
0.816 
0.785 

0.955 
0.942 
0.927 
0.909 
0.889 

0.983 
0.977 
0.969 
0.960 
0.949 

0.994 
0.992 
0.988 
0.984 
0.979 

0.998 
0.997 
0.996 
0.994 
0.992 

0.999 
0.999 
0.998 
0.997 

4.2 

0.015 

0.078 

0.210 

0.395 

0.590 

0.753 

0,867 

0.936 

0.972 

0.989 

0.996 

4.4 

0.012 

0.066 

0.185 

0.359 

0.551 

0.720 

0.844 

0.921 

0.964 

0.985 

0.994 

4.6 

0.010 

0.056 

0.163 

0.326 

0.513 

0.686 

0.818 

0.905 

0.955 

0.980 

0.992 

4.8 

0.008 

0.048 

0.143 

0.294 

0.476 

0.651 

0.791 

0.887 

0.944 

0.975 

0.990 

5,0 

0.007 

0.040 

0.125 

0.-265 

0.440 

0.616 

0.762 

0.867 

0.932 

0.968 

0.986 

5.2 

0.006 

0.034 

0.109 

0.238 

0.406 

0.581 

0.732 

0.845 

0.918 

0.960 

0.982 

5.4 

0.005 

0.029 

0.095 

0.213 

0.373 

0.546 

0.702 

0.822 

0.903 

0.951 

0.977 

5.6 

0.004 

0.024 

0.082 

0.191 

0.342 

0.512 

0.670 

0.797 

0,886 

0.941 

0.972 

5.8 

0.003 

0.021 

0.072 

0.170 

0.313 

0.478 

0.638 

0.771 

0.867 

0.929 

0,965 

6.0 

0.002 

0.017 

0.062 

0,151 

0.285 

0.446 

0.606 

0.744 

0.847 

0.916 

0.957 

X 

11 

12 

13 

14 

15 

16 

17 

18 

19 

20 

21 

3.2 
3.4 

3.6 

3.8 

0.999 

4.0 

0.999 

4.2 

0.999 

4.4 

0.998 

0.999 

4.6 

0.997 

0.999 

4.8 

0.996 

0.999 

5.0 

0.995 

0.998 

0.999 

5.2 

0.993 

0.997 

0.999 

5.4 

0.990 

0.996 

0.999 

5.6 

0.988 

0.995 

0.998 

0.999 

5.8 

0.984 

0.993 

0.997 

0.999 

6.0 

0.98O 

0.991 

0.996 

0.999 

0.999 

[513] 


10 


6.2 
6.4 
6.6 
6.8 
7.0 

0.002 
0.002 
0.001 
0.001 
0.001 

0.015 
0.012 
0.010 
0.009 
0.007 

0.054 
0.046 
0.040 
O.034 
0.030 

0.134 
0.119 
0.105 
0.093 
0.082 

0.259 
0.235 
0.213 
0.192 
0.173 

0.414 
0.384 
0.355 
0.327 
0.301 

0.574 
0.542 
0.511 
0.480 
0.450 

0.716 
0.687 
0.658 
0.628 
0.599 

0,826 
0.803 
0.780 
0.755 
0.729 

0.902 
0.886 
0.869 
0.850 
0.830 

0.949 
0.939 
0.927 
0.915 
0.901 

7.2 
7.4 
7.6 
7.8 
8.0 

0.001 
0.001 
0.001 

0.006 
0.005 
0.004 
0.004 
0.003 

0.025 
0.022 
0.019 
0.016 
0.014 

0.072 
0.063 
0.055 
O.048 
0.042 

0.156 
0.140 
0.125 
0.112 
0.100 

0.276 
0.253 
0.231 
0.210 
0.191 

0.420 
0.392 
0.365 
0.338 
0.313 

0.569 
0.539 
0.510 
0.481 
0.453 

0.703 
0.676 
0.648 
0.620 
0.593 

0.810 
0.788 
0.765 
0.741 
0.717 

0.887 
0.871 
0.854 
0.835 
0.816 

8.2 
8.4 

0.003 
0.002 

O.012 
0.010 

0.037 
0.032 

0,089 
0.079 

0.174 
0.157 

0.290 
0.267 

0.425 
0.399 

0.565 
0.537 

0.692 
0.666 

0.796 
0.774 

11 

12 

13 

14 

15 

16 

17 

18 

19 

20 

21 

6.2 
6.4 
6.6 
6.8 
7.0 

0.975 
0.969 
0.963 
0.955 
0.947 

0.989 
0.986 
0.982 
0.978 
0.973 

0.995 
0.994 
0.992 
0.990 
0.987 

0.998 
0.997 
0.997 
0.996 
0.994 

0.999 
0.999 
0.999 
0.998 
0.998 

0,999 
0.999 
0.999 

7.2 
7.4 
7.6 
7,8 
8.0 

0.937 
0.926 
0.915 
0.902 
0.888 

0.967 
0.961 
0.954 
0.945 
0.936 

0.984 
0.980 
0.976 
0.971 
0.966 

0.993 
0.991 
0.989 
0.986 
0.983 

0.997 
0.996 
0.995 
0.993 
0.992 

0.999 
0.998 
0.998 
0.997 
0.996 

0.999 
0.999 
0.999 
0,999 
0.998 

0,999 

8.2 
8.4 

0.873 
0.857 

0.926 
0,915 

0,960 
0,952 

0.979 
0.975 

0.990 
0.987 

0.995 
0.994 

0.998 
0.997 

0.999 
0.999 

[514] 


ssssssg^gssgsg 

ONONONONOxOxOOOQoOr-«.|>-Co\OvX 


O  O  O 


"^OljOONCOt^CNIvd 
OQOOt--  NOvOiOUO^t1 


CMVOCOT^OOLOX^I 

>VO  O  O  oq  00  CN  < 

1  co  CO  C<J  es  TH  f-i  < 


O  O  O  O  OO 


VQ  _ 

OO  CN  vO  CN 
CO  CO  04  C<> 


rHr-tOOOO 


TH»--iOOOOOO 
OOOOOOOO 


>ON-HtO'»-HOOtr5^COCS--H~H 


O  O  O  O  O  O  O  c 


OO 

OO 


ON  *O  -rt^  co  <r<i  r—  i  •<  —  f  VH 

S88S88S8 

o  d  o"  d  o  o  o*  o* 


o 


o 

CO 


Ox  ON 

CN  OS 


o 


O^>  Ox  Ox 
0s*  Ox  Ox 

odd 


x  Ox  Ox 
ON  Ox  Ox 

odd 

ON  Ox  OO  10  O 
Ox  Ox  ON  Ox  Ox 
ON  Ox  ON  Ox  Ox 

d  d  dd  d 


Ox  ON  QO  vO  i— i  r^ 
ON  O\  Ox  Ox  O\  oo 
Ox  Ox  Ox  ON  Ox  Ox 


\  ON  OO  t—  co  m  CN 

x  Ox  Ox  Ox  ON  CO  t^- 
\  ON  ON  Ox  ON  Ox  Ox 


O  O  O  O  c 


OxdO\Or^^— i\OrHrot--Ox 
OxOxONOxOxCOcOOcOOx 
Ox  Ox  Ox  Ox  Ox  Ox  Ox  Ox  Ox  OO 


OO  O  O  O 


N  ON  ON  Ox  ON 
\  ON  ON  Ox  Ox 


•«-<   *O   O   -r 

ON  00  00  t 


i— i  Ox 
"O  Ox 


Ox  Ox  ON  Ox  Ox  Ox  Ox  CO  oo  J>» 


x  ON 
x  ON 


Ox  ON  Ox  OO  r 
ON  Ox  ON  ON  Ox  < 


ONOxOxOOOOlr-.vO-<i*coOOOuOT-i^hLovC5 
Ox  ON  Ox  ON  Ox  Ox  ON  Ox  ON  ON  OO  OO  OO  t~-  O  10 


[515] 


X—  ON  CM  VO  03  CO  vO 

CO  CM  CM  i—4  i— 4  CD  CD 

CD  CD*  CD*  CD*  CD  O  CD 


CM  i— I  CO  X>-  CM  VO  OO 
ON  03  \O  i— 1  CO  IO  CO 
03  O3  <r-H  i-H  CD  CD  CD 

CD  CD  CD  CD  CD  CD  CD 


xo  t—  i—  t 

„—  <  to  •»—  i 

CO  -*-(  i-< 


«^  03 
co  CM 
CD  CD 


CD  CD  o  CD  o  o  CD 


CD  xo  CM  OO  i— \  CD  03 
xo  CD  X—*  "^^  co  CM  i— i 
^H  ^H  CD  CD  CD  O  CD 

CD  CD  CD  CD  CD  CD  CD 


OO 

S 


S 


*— 

C5  O> 

o  O 


i  CD  CD 


1  01  T-M  T-H 

sss 


i— 

ss 


0  0  0»  0>  (0 


CN  <Z3   T-  1   CV| 
i—  «  04   04   CSI 


CO  OO  CO  CD  OO  OO  OO 
cot--  vo  ^=±1  CD  vo  i—  i 
ON  ON  ON  ON  ON  CO  OO 

« 

i 

CD  CD  CD  CD  CD  CD  CD 

CD  vQ  ^jH  CO  CO  rp  CQ 

ON  ON  ON  ON  OO  OO  X-  - 

CD 

CD  CD  CD  CD  CD  CD  CD 

NO  -^  T—  t  t—  CM  vo  CD 
ON  ON  ON  oO  oO  •£*•»  x>- 

C?N  ON 
ON  ON 

CD  CD 

CD  CD  CD  CD  CD  CD  CD 

i-H  03  CO  03  03  -^i  ON 
XO  CM  CO  CO  X—-  CD  03 
ON  ON  CO  GO  -I1-—  X"*"-  vO 

% 

O\  co  x--» 

ON  ON  ON 
ON  ON  ON 

CD*  CD  CD 

CD  CD  CD  CD  CD  CD  CD 

4T^-  CO  CO  t—  OO  CM  CO 
CM  CO  CO  X—-  CD  CO  xo 

ON  co  oo  X"*  JT—  vo  xo 

00 
co 

ON  ON  ON  ON 
ON  ON  ON  ON 

O  CD  CD  CD 

CD  CD  CD  CD  CD  CD  CD 

CO  CO  CM  CM  xo  Td-<  CO 
ON  •**!  cO  i—  *  co  xo  x*-. 
co  CO  -t~—  •  X"~-  vo  xo  TJ* 

CO 

ON  ON  x^-  xo  i—  i 
ON  ON  ON  ON  ON 
ON  ON  ON  ON  ON 

CD  CD  CD*  CD  CD 

CD  CD  CD  CD  CD  CD  CD 
ON  x**.  vO  X—  »  xo  co  "^ 

VO 
CO 

ON  co  vo  CM  xo 
ON  ON  ON  ON  co 
ON  ON  ON  ON  ON 

CD  CD  CD  CD  CD 

CD  CD  CD  CD  CD  CD  CD 

CO 

ON  ON  ON  ON  oo  X-— 

ON  ON  ON  ON  ON  ON 

CO  <•—  <  CD  vO  CM  CM  CO 

ON  CM  ^  XO  X—  •  ON  ^  —  1 

X—  J>-  ^O  XO  "Sf  CO  CO 

CD  CD  CD  CD  CD  CD 
ON  ON  !"••—•  ^^*  CO  ON  vo 

CD  CD  CD  CD  CD  CD  CD 

CO 

ON  ON  ON  ON  ON  ON  ON 

X-—  ^O  xo  ^  CO  CO  CM 

CO 

CD  CD  CD  CD  CD  CD  CD 

ON  x*-—  ^H  ON  T—  i  ON  CD 

ON  ON  ON  co  oo  vo  to 

£*^>  t  '-"  j  f  —  i  r~-t  r^^t  r*~-i  r-  -i 

co 

X^-.  ON  i-H  X"—  CD  CO  xo 
^  XO  X—  CO  •»—  f  -ctH.  CO 

VO  xo  -«t<  CO  CO  CM  i—  t 

O3 

CO 

CD  CD  CD  CD  CD  CD  CD 

CO  xo  ^H  co  I-H  co  ON 
ON  ON  ON  co  x  —  to  01 
ON  ON  ON  ON  ON  ON  ON 

CD  CD  CD  CD  CD  CD  CD 

*Tj  CD  ^  vo  00  CD  -=fr< 
vO  x^-  OO  CD  co  CO  co 

CO 

O  CD  CD  CD  CD  CD  CD 
VO  CM  xo  CO  VO  CM  CD 

ON  ON  co  X*—  xo  co  ^p 
ON  ON  ON  ON  ON  ON  ON 

CD  CD  CD  CD  CD  CD  CD 

ON  1—4  CM  CM  xo  CO  CM 
vo  QO  CD  co  x*-»  fM  ON 

S3 

CD  CD  CD  CD  CD  CD  CD 

ON  CO  X-—  xo  co  CD  vo 
ON  ON  ON  ON  ON  ON  oo 

•^h1  co  co  CM  i—  t  I-H  CD 

CD  CD  CD  CD  CD  CD  CD 

CD  CD  CD  CD  CD  CD  CD 

CD  CD  CD  CD  CD  CD  CD 

CD  CD  CD  CD  CD  CD  CD 

ON  CD  *-4  03  co  •***  xo 

ON  CD  T-H  CM  CO  ^^  xo 
i—4  CM  CM  CM  O3  CM  CM 

i-H  CM  CM  CM  CM  CM  O3 

£5163 


APPENDIX     31 

CUMULATIVE  STANDARD 
NORMAL  DISTRIBUTION2 


z 

G(& 

z 

= 

GOO 

= 
z 

G(z) 

—  4.  CO 
—  3.99 
—  3.98 
—  3.97 
—  3.96 

O.OOOO3 
O.OOOO3 
O.O0003 
O.  00004 
0.00004 

—  3.60 
—  3.59 
—3.58 
—  3.57 
—  3.56 

O.OO016 
O.OOO17 
O.  00017 
O.OOO18 
0.  00019 

—  3.  2O 
—  3.19 
—  3.18 
—  3.17 
—  3.16 

O.OOO69 
O.OOO71 
0.00074 
0.00076 
O.OOO79 

—  3.95 
—  3.94 
—  3.93 
—  3.92 
—  3.91 

O.OOOO4 
0.00004 
O.OOOO4 
O.OOOO4 
O.O0005 

—3.55 
—  3.54 
—  3.53 
—  3.52 
—3.51 

O.OO019 
O.OO02O 
O.OO021 
0.00022 
O.OOO22 

—  3.15 
—  3.14 
—  3.13 
—  3.12 
—3.11 

O.OOO82 
O.OOO84 
O.OOO87 
O.OOO9O 
0.00094 

—  3.90 
—  3.89 
—  3.88 
—  3.87 
—  3.86 

O.OOOO5 
0.00005 
O.OOOO5 
0.00005 
O.OO006 

—  3.5O 

—  3.49 
—  3.48 
—  3.47 
—  3.46 

O.OO023 
0.00024 
O.OOO25 
O.OO026 
0.00027 

—3.10 
—3.09 
—3.08 
—  3.07 
—  3.O6 

0.00097 
O.OO10O 
O.OO1O4 
O.OO107 
0.00111 

—  3.85 
—  3.84 
—  3.83 
—  3.82 
—  3.81 

0.00006 
O.OOOO6 
0.00006 
O.O0007 
0,00007 

—  3.45 
—  3.44 
—  3.43 
—  3.42 
—  3.41 

O.OO028 
O.OO029 
O.OOO3O 
0.00031 
O.OO032 

—  3.O5 
—3.04 
—  3.O3 
—  3.  02 
—  3.O1 

O.OO114 
0.00118 
O.  00122 
0.00126 
0.00131 

—  3.80 

O.OO007 

—  3.40 

O.OOO34 

—  3.OO 

O.  00135 

—  3.79 

O.OO008 

—  3.39 

0.00035 

—  2.99 

O.  00139 

—  3.78 

0.00008 

—  3.38 

O.OO036 

—  2.98 

O.OO144 

—  3.77 

O.OOOO8 

—  3.37 

O.OOO38 

—  2.97 

O.  00  149 

—  3.76 

O.OOOO8 

—  3.36 

0.00039 

—  2.96 

0.00154 

—  3.75 

O.OOOO9 

—  3.35 

O.OO04O 

—2.95 

O.OO159 

—  3.74 

O.OOOO9 

—  3.34 

O.OO042 

—  2.94 

O.OO164 

—  3.73 

O.OO010 

—  3.33 

O.OO043 

—  2.93 

O.OO169 

—  3.72 

O.OO01O 

—  3.32 

0.00045 

—  2.92 

O.OO175 

—  3.71 

0.00010 

—  3.31 

O.OOO47 

—  2.91 

0.00181 

—  3.7O 

O.OO011 

—  3.30 

0.00048 

—  2.9O 

O.OO187 

—  3.69 

0,00011 

—  3.29 

O.OO050 

—  2.89 

O.OO193 

—  3.68 

O.OO012 

—  3.28 

O.OO052 

—  2.88 

O.OO199 

—3,67 

O.OO012 

—  3.27 

O.OOO54 

—  2.87 

O.OO205 

—  3.66 

O.O0013 

—  3.26 

O.OO056 

—  2.86 

O.OO212 

—  3.65 

O.O0013 

—3.25 

O.OO058 

—  2.85 

O.OO219 

—  3.64 

O.OOO14 

—  3.24 

O.OOO6O 

—  2.84 

O.OO226 

—  3.63 

O.OO014 

—  3.23 

O.OO062 

—  2.83 

O.  00233 

—  3.62 

O.OO015 

—  3.22 

O.OOO64 

—  2.82 

O.OO24O 

—  3.61 

0,QP015 

—  3.21 

O.OO066 

—  2.81 

0.00248 

1  Abridged  from  Karl  Pearson,  Tables  for  Statisticians  and  Biometricians, 
Part  I,  Cambridge  University  Press,  London,  1924,  pp.  2-6.  By  permission  of 
the  author  and  publishers. 


GO) 


f  * 

J  —« 


z. 


(Gz) 


—  2.80 

0.00256 

—  2,30 

0.01072 

—  1.80 

O.O3593 

—  2.79 

0.00264 

—  2.29 

0.01101 

—  1.79 

0.03673 

—  2.78 

0.00272 

—  2.28 

0.01130 

—  1.78 

0.03754 

—  2.77 

0.00280 

—  2.27 

0.01160 

—  1.77 

0.03836 

—  2.76 

0  .  00289 

—  2.26 

0.01191 

—  1.76 

0  .  O3920 

—  2.75 

0.00298 

—  2.25 

0.01222 

—  1.75 

0.04006 

—  2.74 

0.00307 

—  2.24 

0.01255 

—  1.74 

O.O4093 

—  2.73 

0.00317 

—  2.23 

0.01287 

—  1.73 

O.O4182 

—  2.72 

0.00326 

—  2.22 

0.01321 

-1.72 

0.04272 

—  2.71 

0.00336 

—  2.21 

0.01355 

—  1.71 

0.04363 

—  2.70 

0.00347 

—  2.20 

0.01390 

—  1.70 

0.04457 

—  2.69 

0.00357 

—  2.19 

0.01426 

—  1.69 

0.04551 

—  2.68 

0.00368 

—  2.18 

0.01463 

—  1.68 

O.O4648 

—  2.67 

0.00379 

—  2.17 

0.01500 

—  1.67 

O.O4746 

—  2.66 

0.00391 

—  2.16 

0.01539 

—  1.66 

0  .  04846 

—  2.65 

0.00402 

—  2.15 

0.01578 

—  1.65 

0.04947 

—  2.64 

0.00415 

—  2.14 

0.01618 

—  1.64 

0.05050 

—  2.63 

0.00427 

—  2.13 

0.01659 

—  1.63 

0.05155 

—  2.62 

0.00440 

—  2.12 

0.01700 

—  1.62 

0.05262 

—  2.61 

0.00453 

—  2.11 

0.01743 

—  1.61 

0.05370 

—  2.60 

0.00466 

—  2.10 

0.01786 

—  1.60 

0.05480 

—  2.59 

0.00480 

—  2.09 

0.01831 

—  1.59 

0.05592 

—  2.58 

0.00494 

—  2.08 

0.01876 

—  1.58 

0.05705 

—  2.57 

0.00508 

—  2.07 

O.  01923 

—  1.57 

0.05821 

—  2.56 

0.00523 

—  2.06 

0.01970 

—  1.56 

0.05938 

—  2.55 

0.00539 

—  2.05 

0.02018 

—  1.55 

0.06057 

—  2.54 

0  .  00554 

—  2.04 

0.02068 

—  1.54 

0.06178 

—  2.53 

0.00570 

—  2.03 

0.02118 

—  1.53 

0.06301 

—  2.52 

0.00587 

—  2.02 

0.02169 

—  1.52 

0  .  06426 

—  2.51 

0.00604 

—  2.01 

0.02222 

—  1.51 

0.06552 

—  2.5O 

0.00621 

—  2.00 

0.02275 

—  1.50 

0.06681 

—  2.49 

0.00639 

—  1.99 

0.02330 

—  1.49 

0.06811 

—  2.48 

O.  0065  7 

—  1.98 

O.  02385 

—  1.48 

0.06944 

—  2.47 

O.  00676 

—  1.97 

O.  02442 

—  1.47 

0.07078 

—  2.46 

0.00695 

—  1.96 

0.02500 

—  1.46 

0.07215 

-2.45 

0.00714 

—  1.95 

0.02559 

—  1.45 

0.07353 

—  2.44 

0.00734 

—  1.94 

O.  02619 

—  1.44 

0.07493 

—  2.43 

0.00755 

-1.93 

0.02680 

—  1.43 

0.07636 

—  2.42 

0.00776 

—  1.92 

O.  02  743 

—  1.42 

0.07780 

—  2.41 

0.00798 

—  1.91 

O.  02807 

—  1.41 

0.07927 

—  2.40 

0  .  00820 

—  1.90 

0.02872 

—  1.40 

0.08076 

—  2.39 

0.00842 

—  1.89 

0.02938 

—  1.39 

0.08226 

—  2.38 

0,00866 

-1.88 

O  .  03005 

—  1.38 

0.08379 

—  2.37 

0,00889 

—  1.87 

0.03074 

—  1.37 

0.08534 

—  2.36 

0.00914 

—  1.86 

,0.03144 

"~1*3^ 

0.08691 

—  2.35 

0  .  00939 

-1.85 

0.03216 

—  1.35 

0.08851 

—  2.34 

0.00964 

—  1.84 

0,03288 

—  1.34 

0.09012 

—  2.33 

0.00990 

—  1.83 

O.  03362 

—  1.33 

0.09176 

—  2.32 

O.01O17 

—  1.82 

O  .  03438 

—  1.32 

0.09342 

—  2.31 

O.  01044 

—  1.81 

0.03515 

—  1.31 

0.09510 

[5181 


GOO 


GO) 


GGs) 


—  1.30 
—  1.29 
—  1.28 
-1.27 
-1.26 

0.09680 
0.09853 
0.10027 
0.10204 
0.10383 

-0.85 
-0.84 
—  0.83 
-0.82 
—  0.81 

0.19766 
0.20045 
0.2O327 
0.20611 
0.20897 

—0.40 
—  0.39 
-0.38 
-0.37 
-0.36 

0.34458 
0.34827 
0.35197 
0.35569 
0.35942 

-1.25 

0.10565 

—  0.80 

0.21186 

—  0.35 

0.36317 

—  1.24 

0.10749 

-0.79 

0.21476 

-0.34 

0.36693 

—  1.23 

0.10935 

—  0.78 

0.21770 

—  0.33 

0.37070 

—  1.22 

0.11123 

—0.77 

0.22065 

-0.32 

0.37448 

—  1.21 

0.11314 

-0.76 

0.22363 

—  0.31 

0.37828 

-1.20 

0.11507 

-0.75 

0.22663 

-0.30 

0.38209 

—  1.19 

0.11702 

-0.74 

0.22965 

-0.29 

0.38591 

—  1.18 

0.11900 

-0,73 

0.23270 

-0.28 

0.38974 

—  1.17 

0.12100 

-0.72 

0.23576 

—0.27 

0.39358 

—  1.16 

0.12302 

-0.71 

0.23885 

-0.26 

0.39743 

-1.15 

0.12507 

-0.70 

0.24196 

-0.25 

0.40129 

—  1.14 

0.12714 

-0.69 

0.24510 

-0.24 

0.40517 

—  1.13 

0.12924 

-0.68 

0.24825 

-0.23 

0.40905 

—  1.12 

0.13136 

-0.67 

0.25143 

—0.22 

0.41294 

-1.11 

0.13350 

-0.66 

0.25463 

-0.21 

0.41683 

—  1.10 

0.13567 

-0.65 

0.25785 

-0.20 

0.42074 

-1.09 

0.13786 

-0.64 

0.26109 

—  0.19 

0.42465 

-1.08 

0.14007 

-0.63 

0.26435 

—0.18 

0.42858 

-1.07 

0.14231 

-0.62 

0.26763 

-0.17 

0.43251 

—  1.06 

0.14457 

—  0.61 

0.27093 

—  0.16 

0.43644 

-1.05 

0  .  14686 

—  0.60 

0.27425 

-0.15 

0.44038 

—  1.04 

0.14917 

-0.59 

0.27760 

-0.14 

0.44433 

-1.03 

0.15150 

-0.58 

0.28096 

—0.13 

0.44828 

—  1.02 

0.15386 

-0.57 

0.28434 

—  0.12 

0.45224 

-1.01 

0.15625 

—  0.56 

0.28774 

-0.11 

0.45620 

-1.00 

0.15866 

-0.55 

0.29116 

-0.10 

0.46017 

—  0.99 

0.16109 

-0.54 

0.29460 

-0.09 

0.46414 

-0.98 

0.16354 

-0.53 

0.29806 

-0.08 

0.46812 

—  0.97 

0.16602 

-0.52 

0.30153 

-0.07 

0.47210 

-0.96 

0.16853 

-0.51 

0.30503 

-0.06 

0.47608 

-0.95 

0.17106 

-0.50 

0.30854 

-0.05 

0.48006 

-0.94 

0.17361 

-0.49 

0.31207 

-0.04 

0.48405 

-0.93 

0.17619 

-0.48 

0.31561 

-0.03 

0.48803 

—  0.92 

0.17879 

-0.47 

0.31918 

—  0.02 

0.49202 

—  0.91 

0.18141 

-0.46 

0.32276 

—0.01 

0.49601 

—  0.90 

0.18406 

-0.45 

0.32636 

0.00 

0.50000 

-0.89 

0.18673 

-0.44 

0.32997 

0.01 

0.50399 

-0.88 

0.18943 

-0.43 

0.33360 

0.02 

0.50798 

-0.87 

0.19215 

-0.42 

0.33724 

0.03 

0.51197 

-0.86 

0.19489 

—  0.41 

0.34090 

0.04 

0.51595 

E5193 


CO) 


0,05 

0.51994 

0.50 

0.69146 

0.95 

0.82894 

0,06 

0.52392 

0.51 

0.69497 

0.96 

0.83147 

0.07 

0.52790 

0.52 

0.69847 

0.97 

0.83398 

0.08 

0.53188 

0.53 

0.70194 

0.98 

0  .  83646 

0.09 

0.53586 

0.54 

0  .  70540 

0,99 

0.83891 

0.10 

0.53983 

0.55 

0.70884 

1.00 

0.84134 

0.11 

0.54380 

0.56 

0.71226 

1.01 

0.84375 

O.12 

0.54776 

0.57 

0.71566 

1.02 

0.84614 

0.13 

0.55172 

0.58 

0.71904 

1.03 

0.84850 

0.14 

0.55567 

0.59 

0.72240 

1.04 

0.85083 

O.15 

0.55962 

0.60 

0.72575 

1.05 

0.85314 

0,16 

0.56356 

0.61 

0.72907 

1.06 

0.85543 

0.17 

0.56749 

0.62 

0.73237 

1.07 

0.85769 

0.18 

O.  57142 

0.63 

0.73565 

1.08 

0.85993 

0.19 

0.57535 

0.64 

0.73891 

1.09 

0.86214 

0.20 

0.57926 

0.65 

0.74215 

1.10 

0.86433 

0.21 

0.58317 

0.66 

0.74537 

1.11 

0.8665O 

0.22 

0.58706 

0.67 

0.74857 

1.12 

0.86864 

0.23 

0.59095 

0.68 

0.75175 

1.13 

0.87076 

0.24 

0.59483 

0.69 

0.75490 

1.14 

0.87286 

0.25. 

0.59871 

0.70 

0.75804 

1.15 

0.87493 

0.26 

0.60257 

0.71 

0.76115 

1.16 

0.87698 

0.27 

0.60642 

0.72 

0.76424 

1.17 

O.  87900 

0.28 

0.61026 

0.73 

0.76730 

1.18 

0.88100 

0.29 

0.61409 

0.74 

0.77035 

1.19 

0.88298 

0.30 

0.61791 

0.75 

0.77337 

1.20 

0.88493 

0.31 

0.62172 

0.76 

0.77637 

1.21 

0.88686 

0.32 

0.62552 

0.77 

0.77935 

1.22 

0.88877 

0.33 

0.62930 

0.78 

0.78230 

1.23 

0.89065 

0.34 

0.63307 

0.79 

0.78524 

1.24 

0.89251 

0.35 

0.63683 

0.80 

0.78814 

1.25 

0.89435 

0.36 

0.64058 

0.81 

0.79103 

1,26 

0.89617 

O.37 

0.64431 

0.82 

0.79389 

1.27 

0.89796 

O.38 

0.64803 

0.83 

0.79673 

1.28 

0.89973 

0.39 

0.65173 

0,84 

0.79955 

1.29 

0,90147 

0.40 

0.65542 

0.85 

0.80234 

1.30 

0,90320 

O.41 

0.65910 

0.86 

O.  80511 

1.31 

0.90490 

0.42 

0.66276 

0.87 

0.80785 

1.32 

0.90658 

O.43 

O.  66640 

0.88 

0.81057 

1.33 

0.90824 

0.44 

0.67003 

0.89 

0,81327 

1.34 

0.90988 

0.45 

0.67364 

0.90 

0.81594 

1.35 

0.91149 

0.46 

0.67724 

0.91 

0,81859 

1.36 

0.91309 

0.47 

0.68082 

0.92 

0.82121 

1.37 

0.91466 

0.48 

0.68439 

0.93 

0.82381 

1.38 

0.91621 

0.49 

0.68793 

0.94 

0,82639 

1.39 

0,91774 

15201 


1.40 

0,91924 

1.85 

0.96784 

2.30 

0,98928 

1.41 

0.92073 

1.86 

0.96856 

2.31 

0.98956 

1.42 

0.92220 

1.87 

0.96926 

2.32 

0.98983 

1.43 

0.92364 

1.88 

0.96995 

2.33 

0.99010 

1.44 

0.92507 

1.89 

0.97062 

2.34 

0.99036 

1.45 

0.92647 

1.90 

0.97128 

2.35 

0.99061 

1.46 

0.92785 

1.91 

0.97193 

2.36 

0.99086 

1.47 

0.92922 

1.92 

0.97257 

2.37 

0.99111 

1.48 

0.93056 

1.93 

0.97320 

2.38 

O.  99134 

1.49 

0.93189 

1.94 

0.97381 

2.39 

0.99158 

1.50 

0.93319 

1.95 

0.97441 

2.40 

0.99180 

1.51 

0.93448 

1.96 

0.97500 

2.41 

0.99202 

1.52 

0.93574 

1.97 

0.97558 

2.42 

O.  99224 

1.53 

0.93699 

1.98 

0.97615 

2.43 

0.99245 

1.54 

0.93822 

1.99 

0.97670 

2.44 

0.99266 

1.55 

0.93943 

2.0O 

0.97725 

2.45 

O.  99286 

1.56 

0.94062 

2.01 

0.97778 

2.46 

0.99305 

1.57 

0.94179 

2.02 

0.97831 

2.47 

0.99324 

1.58 

0.94295 

2.03 

0.97882 

2.48 

0.99343 

1.59 

0.94408 

2.04 

0.97932 

2.49 

0.99361 

1.60 

0.94520 

2.05 

0.97982 

2.50 

0.99379 

1.61 

0.94630 

2.06 

0,98030 

2.51 

O.  99396 

1.62 

0.94738 

2.07 

0.98077 

2,52 

0.  99413 

1.63 

0.94845 

2.08 

0.98124 

2.53 

0.99430 

1.64 

0.94950 

2.09 

0.98169 

2.54 

0.99446 

1.65 

0.95053 

2.10 

0.98214 

2.55 

0.99461 

1.66 

0.95154 

2.11 

0.98257 

2.56 

0.99477 

1.67 

0.95254 

2.12 

0.98300 

2.57 

0.99492 

1.68 

0.95352 

2.13 

0.98341 

2.58 

0.99506 

1.69 

0.95449 

2.14 

0.98382 

2.59 

0.99520 

1.70 

0.95543 

2.15 

0.98422 

2.6O 

O.  99534 

1.71 

0.95637 

2.16 

0.98461 

2.61 

0.99547 

1.72 

0.95728 

2.17 

0.98500 

2,62 

0.99560 

1.73 

0.95818 

2.18 

0.98537 

2.63 

0.99573 

1.74 

0.95907 

2.19 

0.98574 

2.64 

0.99585 

1.75 

0.95994 

2,20 

0.98610 

2.65 

0.99598 

1.76 

0.96080 

2.21 

0.98645 

2.66 

O.  99609 

1.77 

0.96164 

2.22 

0.98679 

2.67 

0.99621 

1.78 

0.96246 

2.23 

0.98713 

2.68 

0.99632 

1.79 

0.96327 

2.24 

0.98745 

2.69 

0.  99  643 

1.80 

0.96407 

2.25 

0.98778 

2.70 

0.99653 

1.81 

0.96485 

2.26 

0.988O9 

2.71 

0.99664 

1.82 

0.96562 

2.27 

0.98840 

2.72 

0.  99674 

1.83 

0.96638 

2.28 

0.98870 

2.73 

0.99683 

1.84 

0.96712 

2.29 

0.98899 

2.74 

0.99693 

[sail 


z 

G(z) 

z 

G(z) 

z 

G(z) 

2.75 

0.99702 

3.20 

0.99931 

3.65 

0.99987 

2.76 

0.99711 

3.21 

0.99934 

3.66 

0.99987 

2.77 

0.99720 

3.22 

0.99936 

3.67 

0.99988 

2.78 

0.99728 

3.23 

0.99938 

3.68 

0.99988 

2.79 

0.99736 

3.24 

0.99940 

3.69 

0.99989 

2.80 

0.99744 

3.25 

0.99942 

3.70 

0.99989 

2.81 

0.99752 

3.26 

0  .  99944 

3.71 

0.99990 

2.82 

0.99760 

3.27 

0.99946 

3.72 

0.99990 

2.83 

0.99767 

3.28 

0.99948 

3.73 

0.99990 

2.84 

0.99774 

3.29 

0.99950 

3.74 

0.99991 

2.85 

0.99781 

3.30 

0.99952 

3.75 

0.99991 

2.86 

0.99788 

3.31 

0.99953 

3.76 

0.99992 

2.87 

0.99795 

3.32 

0.99955 

3.77 

0.99992 

2.88 

0.99801 

3.33 

0.99957 

3.78 

0,99992 

2.89 

0.99807 

3.34 

0.99958 

3.79 

0.99992 

2.90 

0.99813 

3.35 

0.99960 

3.80 

0.99993 

2.91 

0.99819 

3.36 

0.99961 

3.81 

0.99993 

2.92 

0.99825 

3.37 

0.99962 

3.82 

0.99993 

2.93 

0.99831 

3.38 

0.99964 

3.83 

0.99994 

2.94 

0.99836 

3.39 

0.99965 

3.84 

0.99994 

2.95 

0.99841 

3.40 

0.99966 

3.85 

0.99994 

2.96 

0.99846 

3.41 

0.99968 

3.86 

0  .  99994 

2.97 

0.99851 

3.42 

0.99969 

3.87 

0.99995 

2.98 

0.99856 

3.43 

0.99970 

3.88 

0.99995 

2.99 

0.99861 

3.44 

0.99971 

3.89 

0.99995 

3.00 

0.99865 

3.45 

0.99972 

3.90 

0  .  99995 

3.01 

0.99869 

3.46 

0.99973 

3.91 

0.99995 

3.02 

0.99874 

3.47 

0.99974 

3.92 

0.99996 

3.03 

0.99878 

3.48 

0.99975 

3.93 

0.99996 

3.04 

0.99882 

3.49 

0.99976 

3.94 

0.99996 

3.05 

0.99886 

3.50 

0.99977 

3.95 

0.99996 

3.06 

0.99889 

3.51 

0.99978 

3.96 

0.99996 

3.07 

0.99893 

3.52 

0.99978 

3.97 

0.99996 

3.08 

0.99897 

3.53 

0.99979 

3.98 

0.99997 

3.09 

0.99900 

3.54 

0.99980 

3.99 

0.99997 

3.10 

0.99903 

3.55 

0.99981 

4.00 

0.99997 

3.11 

O.  99906 

3.56 

0.99981 

3.12 

0.99910 

3.57 

0.99982 

3.13 

0.99913 

3.58 

0.99983 

3.14 

0.99916 

3.59 

0.99983 

3.15 

0.99918 

3.60 

0.99984 

3.16 

0.99921 

3.61 

0.99985 

3.17 

0.99924 

3.62 

0.99985 

3.18 

0.99926 

3.63 

0.99986 

3.19 

0.99929 

3.64 

0.99986 

15223 


APPENDIX    41 

CUMULATIVE  CHI-SQUARE 
DISTRIBUTION2 

1  Adapted  from  A.  Hald  and  S*  A.  Sinkbaek,  "A  table  of  percentage  points  of 
the   x2-distribution,J)   Skandinavisk  Aktuarietidskrift,  1950,   pp.   170—75.  By  per 
mission  of  the  authors  and  publishers. 

2  Entries  in  the  table  are  values  of  x|  where 


15231 


V 

P 

0.0005 

0.001 

0.005 

0.01 

0.025 

0.05 

0.10 

0.20 

0.30 

0.40 

1 

0.0*393 

0.0*157 

0.0*393 

0.0»157 

0.0»982 

0.02393 

0.0158 

0.0642 

0.148 

0.275 

2 

o.onoo 

0.0^200 

0.0100 

0.0201 

0.0506 

0.103 

0.211 

0.446 

0.713 

1.02 

3 

0.0153 

0.0243 

0.0717 

0.115 

0.216 

0.352 

0.584 

1.00 

1.42 

1.87 

4 

0.0639 

0.0908 

0.207 

0.297 

0.484 

0.711 

1.06 

1.65 

2.19 

2.75 

5 

0.158 

0.210 

0.412 

0.554 

0.831 

1.15 

1.61 

2.34 

3.00 

3.66 

6 

0.299 

0.381 

0.676 

0.872 

1.24 

1.64 

2.20 

3.07 

3.83 

4.57 

7 

0.485 

0.598 

0.989 

1.24 

1.69 

2.17 

2.83 

3.82 

4.67 

5.49 

8 

0.710 

0.857 

1.34 

1.65 

2.18 

2.73 

3.49 

4.59 

5.53 

6.42 

g 

0.972 

1.15 

1.73 

2.09 

2.70 

3.33 

4.17 

5.38 

6.39 

7.36 

10 

1.26 

1.48 

2.16 

2.56 

3.25 

3.94 

4.87 

6.18 

7.27 

8.30 

11 

1.59 

1.83 

2.60 

3.05 

3.82 

4.57 

5.58 

6.99 

8.15 

9.24 

12 

1.93 

2.21 

3.07 

3.57 

4.40 

5.23 

6.30 

7.81 

0.03 

10.2 

13 

2.31 

2.62 

3.57 

4.11 

5.01 

5.89 

7.04 

8,63 

9.93 

11.1 

14 

2.70 

3.04 

4.07 

4.66 

5.63 

6.57 

7.79 

9.47 

10.8 

12.1 

15 

3.11 

3.48 

4.60 

5.23 

6.26 

7.26 

8.55 

10.3 

11.7 

13.0 

16 

3.54 

3.94 

5.14 

5.81 

6.91 

7:96 

9.31 

11.2 

12.6 

14.0 

17 

3.98 

4.42 

5.70 

6.41 

7.56 

8.67 

10.1 

12.0 

13.5 

14.9 

18 

4.44 

4.90 

6.26 

7,01 

8.23 

9.39 

10.9 

12.9 

14.4 

15.9 

19 

4.91 

5.41 

6.84 

7.63 

8.91 

10.1 

11.7 

13.7 

15.4 

16.9 

20 

5.40 

5.92 

7.43 

8.26 

9.59 

10.9 

12.4 

14.6 

16.3 

17.8 

21 

5.90 

6.45 

8.03 

8.90 

10.3 

11.6 

13.2 

15.4 

17.2 

18.8 

22 

6.40 

6.98 

8.64 

9.54 

11.0 

12.3 

14.0 

16.3 

18.1 

19.7 

23 

6.92 

7.53 

9.26 

10.2 

11.7 

13.1 

14.8 

17.2 

19,0 

20.7 

24 

7.45 

8.08 

9.89 

10.9 

12.4 

13.8 

15.7 

18.1 

19.0 

21.7 

25 

7.99 

8.65 

10.5 

11.5 

13.1 

14.6 

16.5 

18.9 

20.0 

22.0 

26 

8.54 

9.22 

11.2 

12.2 

13.8 

15.4 

17.3 

19.8 

21.8 

23.6 

27 

9.09 

9.80 

11.8 

12.9 

14.6 

16.2 

18.1 

20.7 

22.7 

24.5 

28 

9.66 

10.4 

12.5 

13.6 

15.3 

16.9 

18.9 

21.6 

23.6 

25.5 

29 

10.2 

11.0 

13.1 

14.3 

16.0 

17.7 

19.8 

22.5 

24.6 

26.5 

30 

10,8 

11.6 

13.8 

15.0 

16,8 

18.5 

20.6 

23.4 

25.5 

27.4 

31 

11.4 

12.2 

14.5 

15.7 

17.5 

19.3 

21.4 

24.3 

26.4 

28.4 

32 

12.0 

12.8 

15.1 

16.4 

18.3 

20.1 

22.3 

25.1 

27.4 

29.4 

33 

12.6 

13.4 

15.8 

17.1 

19.0 

20.9 

23.1 

26.0 

28.3 

30.3 

34 

13.2 

14.1 

16.5 

17.8 

19-8 

21.7 

24.0 

26.9 

29.2 

31.3 

35 

13.8 

14.7 

17.2 

18.5 

20.6 

22.5 

24.8 

27.8 

30.2 

32.3 

36 

14.4 

15.3 

17.9 

19.2 

21.3 

23.3 

25.6 

28.7 

31.1 

33.3 

37 

15.0 

16.0 

18.6 

20.0 

22.1 

24.1 

26.5 

29.6 

32.1 

34.2 

38 

15.6 

16.6 

19.3 

20.7 

22.9 

24.9 

27.3 

30.5 

33.0 

35.2 

39 

16,3 

17.3 

20.0 

21.4 

23.7 

25.7 

28.2 

31.4 

33.9 

36.2 

40 

16.9 

17.9 

20.7 

22.2 

24.4 

26.5 

29.1 

32.3 

34.9 

37.1 

41 

17.5 

18.6 

21.4 

22.9 

25.2 

27.3 

29.9 

33.3 

35.8 

38.1 

42 

18.2 

19.2 

22.1 

23.7 

26.0 

28.1 

30.8 

34.2 

36.8 

39.1 

43 

18.8 

19.9 

22.9 

24.4 

26.8 

29.0 

31.6 

35.1 

37.7 

40.0 

44 

19.5 

20.6 

23.6 

25.1 

27.6 

29.8 

32.5 

36,0 

38.6 

41.0 

45 

20.1 

21,3 

24.3 

25.9 

28.4 

30.6 

33.4 

36.9 

39.6 

42.0 

46 

20.8 

21.9 

25.0 

26.7 

29.2 

31.4 

34.2 

37.8 

40.5 

43.0 

47 

21.5 

22.6 

25,8 

27.4 

30.0 

32.3 

35.1 

38.7 

41.5 

43.9 

48 

22.1 

23.3 

26.5 

28.2 

30.8 

33.1 

35.9 

39.6 

42.4 

44.0 

49 

22.8 

24.0 

27.2 

28.9 

31.6 

33.9 

36.8 

40.5 

43.4 

45.9 

50 

23.5 

24.7 

28.0 

29.7 

32.4 

34.8 

37.7 

41.4 

44.3 

46.9 

[524] 


V 

P 

0.50 

0.60 

0.70 

0.80 

0.90 

0.95 

0.975 

0.99 

0.995 

0.999 

0.9995 

1 

0.455 

0.708 

1.07 

1.64 

2.71 

3.84 

5.02 

6.63 

7.88 

10.8 

12.1 

2 

1.39 

1.83 

2.41 

3.22 

4.61 

5.99 

7.38 

9.21 

10.6 

13.8 

15.2 

3 

2.37 

2.95 

3.67 

4.64 

6.25 

7.81 

9.35 

11.3 

12.8 

16.3 

17.7 

4 

3.36 

4.04 

4.88 

5.99 

7.78 

9.49 

11.1 

13.3 

14.9 

18.5 

20.0 

5 

4.35 

5.13 

6.06 

7.29 

9.24 

11.1 

12.8 

15.1 

16,7 

20.5 

22.1 

6 

5.35 

6.21 

7.23 

8.56 

10.6 

12.6 

14.4 

16.8 

18.5 

22.5 

24.1 

7 

6.35 

7.28 

8.38 

9.80 

12.0 

14.1 

16.0 

18.5 

20.3 

24.3 

26.0 

8 

7.34 

8.35 

9.52 

11.0 

13.4 

15.5 

17.5 

20.1 

22,0 

26.1 

27.9 

9 

8.34 

9.41 

10.7 

12.2 

14.7 

16.9 

ia.o 

21.7 

23.6 

27.9 

29.7 

10 

9.34 

10.5 

11.8 

13.4 

16.0 

18.3 

20.5 

23.2 

25.2 

29.6 

31.4 

11 

10.3 

11.5 

12.9 

14.6 

17.3 

19.7 

21.9 

24.7 

26.8 

31.3 

33.1 

12 

11.3 

12.6 

14.0 

15.8 

18.5 

21.0 

23,3 

26.2 

28.3 

32.9 

34.8 

13 

12.3 

13.6 

15.1 

17.0 

19.8 

22.4 

24.7 

27.7 

29.8 

34.5 

36.5 

14 

13.3 

14.7 

16.2 

18.2 

21.1 

23.7 

26.1 

29.1 

31.3 

36.1 

38.1 

15 

14.3 

15.7 

17.3 

19.3 

22.3 

25.0 

27.5 

30.6 

32.8 

37.7 

39.7 

16 

15.3 

16.8 

18.4 

20.5 

23.5 

26.3 

28.8 

32.0 

34.3 

39.3 

41.3 

17 

16.3 

17.8 

19.5 

21.6 

24.8 

27.6 

30.2 

33.4 

35.7 

40.8 

42.9 

18 

17.3 

18.9 

20.6 

22.8 

26.0 

28.9 

31.5 

34.8 

37.2 

42.3 

44.4 

19 

18.3 

19.9 

21.7 

23.9 

27.2 

30.1 

32.9 

36.2 

38.6 

43.8 

46.0 

20 

19.3 

21.0 

22.8 

25.0 

28.4 

31.4 

34.2 

37.6 

40.0 

45.3 

47.5 

21 

20.3 

22.0 

23.9 

26.2 

29.6 

32.7 

35.5 

38.9 

41.4 

46.8 

49.0 

22 

21.3 

23.0 

24.9 

27.3 

30.8 

33.9 

36.8 

40.3 

42.8 

48.3 

50.5 

23 

22.3 

24.1 

26.0 

28.4 

32.0 

35.2 

38.1 

41.6 

44.2 

49,7 

52.0 

24 

23.3 

25.1 

27.1 

29.6 

33.2 

36.4 

39.4 

43.0 

45.6 

51.2 

53.5 

25 

24.3 

26.1 

28.2 

30.7 

34.4 

37.7 

40.6 

44.3 

46.9 

52.6 

54.9 

26 

25.3 

27.2 

29.2 

31.8 

35.6 

38.9 

41.9 

45.6 

48.3 

54,1 

56.4 

27 

26.3 

28.2 

30.3 

32.9 

36.7 

40.1 

43.2 

47.0 

49.6 

55.5 

57.9 

28 

27.3 

29.2 

31.4 

34,0 

37.9 

41.3 

44.5 

48.3 

51.0 

56.9 

59.3 

29 

28.3 

30.3 

32.5 

35.1 

39.1 

42.6 

45.7 

49.6 

52.3 

58.3 

60.7 

30 

29.3 

31.3 

33.5 

36.3 

40.3 

43.8 

47.0 

50.9 

53.7 

59.7 

62.2 

31 

30.3 

32.3 

34.6 

37.4 

41.4 

45.0 

48.2 

52.2 

55.0 

61.1 

63.6 

32 

31.3 

33.4 

35.7 

38.5 

42.6 

46.2 

49.5 

53.5 

56.3 

62.5 

65.0 

33 

32.3 

34.4 

36.7 

39.6 

43.7 

47.4 

50.7 

54.8 

57.6 

63.9 

66.4 

34 

33.3 

35.4 

37.8 

40.7 

44.9 

48.6 

52.0 

56.1 

59.0 

65.2 

67.8 

35 

34.3 

36.5 

38.9 

41.8 

46.1 

49.8 

53.2 

57.3 

60.3 

66.6 

69.2 

36 
37 

35.3 
36  3 

37.5 
38.5 

39,9 
41.0 

42.9 
44.0 

47.2 
48.4 

51.0 
52.2 

54.4 
55.7 

58.6 
59.9 

61.6 
62,9 

68.0 
69.3 

70.6 
72.0 

38 
39 
40 

37.3 
38.3 
39.3 

39.6 
40.6 
41.6 

42.0 
43.1 
44.2 

45.1 
46.2 
47.3 

49.5 
50.7 
51.8 

53.4 
54.6 
55.8 

56.9 
58.1 
59.3 

61.2 
62.4 
63.7 

64.2 
65.5 
66.8 

70.7 
72.1 
73.4 

73.4 
74.7 
76.1 

41 
42 
43 
44 
45 

40.3 
41.3 
42.3 
43.3 
44.3 

42.7 
43.7 
44.7 
45.7 
46.8 

45.2 
46.3 
47.3 
48.4 
49.5 

48.4 
49.5 
50.5 
51.6 
52.7 

52.9 
54.1 
55.2 
56.4 
57.5 

56.9 
58.1 
59.3 
60.5 
61.7 

60.6 
61.8 
63.0 
64.2 
65.4 

65.0 
66.2 
67.5 
68.7 
70.0 

68.1 
69.3 
70.6 
71.9 
73.2 

74.7 
76.1 
77.4 
78.7 
80.1 

77.5 
78.8 
80.2 
81.5 
82.9 

46 
47 
48 
49 
50 

45.3 
46.3 
47.3 
48.3 
49.3 

47.8 
48.8 
49.8 
50.9 
51.9 

50.5 
51.6 
52.6 
53.7 
54.7 

53.8 
54.9 
56.0 
57.1 
58.2 

58.6 
59.8 
60.9 
62.0 
63.2 

62.8 
64.0 
65.2 
66.3 
67.5 

66.6 
67.8 
69.0 
70.2 
71.4 

71.2 
72.4 
73,7 
74.9 
76.2 

74.4 
75.7 
77.0 

78.2 
79.5 

81.4 
82.7 
84.0 
85.4 
86.7 

84.2 
85.6 
86.9 
88.2 
89.6 

[525] 


If 

V 

0.0005 

0.001 

0.005 

0.01 

0.025 

0.05 

0.10 

0.20 

0.30 

0.40 

0.50 

51 

24.1 

25.4 

28.7 

30.5 

33.2 

35.6 

38.6 

42.4 

45.3 

47.8 

50.3 

52 

24.8 

26.1 

29.5 

31.2 

34.0 

36.4 

39.4 

43.3 

46.2 

48.8 

51.3 

53 

25.5 

26.8 

30.2 

32.0 

34.8 

37.3 

40.3 

44.2 

47.2 

49.8 

52.3 

54 

26.2 

27.5 

31.0 

32.8 

35.6 

38.1 

41.2 

45.1 

48.1 

50.8 

53.3 

55 

26.9 

28.2 

31.7 

33.6 

36.4 

39.0 

42.1 

46.0 

49.1 

51.7 

54.3 

56 

27.6 

28.9 

32.5 

34,3 

37.2 

39.8 

42.9 

47.0 

50.0 

52.7 

55.3 

57 

28.2 

29.6 

33.2 

35.1 

38.0 

40.6 

43.8 

47.9 

51.0 

53.7 

56.3 

58 

28.9 

30.3 

34.0 

35.9 

38.8 

41.5 

44,7 

48.8 

51.9 

54.7 

57.3 

59 

29.6 

31.0 

34.8 

36.7 

39.7 

42.3 

45.6 

49.7 

52.9 

55.6 

58.3 

60 

30.3 

31.7 

35.5 

37.5 

40.5 

43.2 

46.5 

50.6 

53.8 

56,6 

59.3 

61 

31.0 

32.5 

36.3 

38.3 

41.3 

44.0 

47.3 

51.6 

54.8 

57.6 

60.3 

62 

31,7 

33.2 

37.1 

39.1 

42.1 

44.9 

48,2 

52.5 

55.7 

58.6 

61.3 

63 

32.5 

33.9 

37.8 

39.9 

43.0 

45.7 

49.1 

53.4 

56.7 

59.6 

62.3 

64 

33.2 

34.6 

38.6 

40.6 

43.8 

46.6 

50.0 

54.3 

57.6 

60.5 

63.3 

65 

33.9 

35.4 

39.4 

41.4 

44.6 

47.4 

50.9 

55.3 

58.6 

61.5 

64.3 

66 

34.6 

36.1 

40.2 

42.2 

45.4 

48.3 

51.8 

56.2 

59.5 

62.5 

65.3 

67 

35.3 

36.8 

40.9 

43.0 

46.3 

49.2 

52.7 

57.1 

60.5 

63.5 

66.3 

68 

36.0 

37.6 

41.7 

43.8 

47.1 

50.0 

53.5 

58.0 

61.4 

64.4 

67.3 

69 

36.7 

38.3 

42.5 

44.6 

47.9 

50.9 

54.4 

59.0 

62.4 

65.4 

68.3 

70 

37.5 

39.0 

43.3 

45.4 

48.8 

51.7 

55.3 

59.9 

63.3 

66.4 

69.3 

71 

38.2 

39.8 

44.1 

46.2 

49.6 

52.6 

56.2 

60.8 

64.3 

67.4 

70.3 

72 

38.9 

40.5 

44.8 

47.1 

50.4 

53.5 

57.1 

61.8 

65.3 

68.4 

71.3 

73 

39.6 

41.3 

45.6 

47.9 

51.3 

54.3 

58.0 

62.7 

66.2 

69.3 

72.3 

74 

40.4 

42.0 

46.4 

48.7 

52.1 

55.2 

58.9 

63.6 

67.2 

70.3 

73.3 

75 

41.1 

42.8 

47.2 

49.5 

52.9 

56.1 

59.8 

64.5 

68.1 

71.3 

74.3 

76 

41.8 

43.5 

48.0 

50.3 

53.8 

56.9 

60.7 

65.5 

69.1 

72.3 

75.3 

77 

42.6 

44.3 

48.8 

51.1 

54.6 

57.8 

61.6 

66.4 

70.0 

73.2 

76.3 

78 

43.3 

45.0 

49.6 

51.9 

55.5 

58.7 

62.5 

67.3 

71.0 

74.2 

77.3 

79 

44.1 

45.8 

50.4 

52.7 

56.3 

59.5 

63.4 

68.3 

72.0 

75.2 

78.3 

80 

44.8 

46.5 

51.2 

53.5 

57.2 

60.4 

64.3 

69.2 

72.9 

76.2 

79.3 

81 

45.5 

47.3 

52.0 

54.4 

58.0 

61.3 

65.2 

70.1 

73.9 

77.2 

80.3 

82 

46.3 

48.0 

52.8 

55.2 

58.8 

62.1 

66.1 

71.1 

74.8 

78.1 

81.3 

83 

47.0 

48.8 

53.6 

56.0 

59.7 

63.0 

67.0 

72.0 

75.8 

79.1 

82.3 

84 

47.8 

49.6 

54.4 

56.8 

60.5 

63.9 

67.9 

72.9 

76.8 

80.1 

83.3 

85 

48.5 

50.3 

55.2 

57.6 

61.4 

64.7 

68.8 

73.9 

77.7 

81.1 

84.3 

86 

49.3 

51.1 

56.0 

58.5 

62.2 

65.6 

69.7 

74.8 

78.7 

82.1 

85.3 

87 

50.0 

51.9 

56.8 

59.3 

63.1 

66.5 

70.6 

75.7 

79.6 

83.0 

86.3 

88 

50.8 

52.6 

57.6 

60.1 

63.9 

67.4 

71.5 

76.7 

80.6 

84.0 

87.3 

89 

51.5 

53.4 

58.4 

60.9 

64.8 

68.2 

72.4 

77.6 

81.6 

85.0 

88.3 

90 

52.3 

54,2 

59.2 

61.8 

65.6 

69.1 

73.3 

78.6 

82.5 

86.0 

89.3 

91 

53.0 

54.9 

60.0 

62.6 

66.5 

70.0 

74.2 

79.5 

83.5 

87.0 

90.3 

92 

53.8 

55.7 

60.8 

63.4 

67.4 

70-9 

75.1 

80.4 

84.4 

88.0 

91.3 

93 

54.5 

56.5 

61.6 

64.2 

68.2 

71.8 

76.0 

81.4 

85.4 

88.9 

92.3 

94 

55.3 

57.2 

62.4 

65.1 

69.1 

72,6 

76.9 

82.3 

86,4 

89.9 

93.3 

95 

56.1 

58.0 

63.2 

65.9 

69.9 

73.5 

77.8 

83.2 

87.3 

90.9 

94.3 

96 

56.8 

58.8 

64.1 

66.7 

70.8 

74.4 

78.7 

84.2 

88.3 

91.9 

95.3 

97 

57.6 

59.6 

64.9 

67.6 

71.6 

75.3 

79.6 

85.1 

89.2 

92.9 

96.3 

98 

58.4 

60.4 

65.7 

68.4 

72.5 

76,2 

80.5 

86.1 

90.2 

93.8 

97.3 

99 

59.1 

61.1 

66.5 

69.2 

73.4 

77.0 

81.4 

87.0 

91.2 

94.8 

98.3 

100 

59.9 

61.9 

67.3 

70.1 

74.2 

77.9 

82.4 

87.9 

92.1 

95.8 

99.3 

[5261 


V 

P 

0.60 

0.70 

0.80 

0.90 

0.95 

0.975 

0.99 

0.995 

0.999 

0.9995 

51 
52 
53 
54 
55 

52.9 
53.9 
55.0 
56.0 
57.0 

55.8 
56.8 
57.9 
58.9 
60.0 

59.2 
60.3 
61.4 
62,5 
63.6 

64.3 
65.4 
66.5 
67.7 
68.8 

68.7 
69.8 
71.0 
72.2 
73.3 

72.6 
73.8 
75.0 
76.2 
77.4 

77.4 
78.6 
79.8 
81.1 
82.3 

80.7 
82.0 
83.3 
84.5 
85.7 

88.0 
89,3 
90.6 
91.9 
93.2 

90.9 
92.2 
93.5 
94.8 
96.2 

56 
57 
58 

58.0 
59,1 
60.1 

61.0 
62.1 
63.1 

64.7 
65.7 
66.8 

69.9 
71.0 
72.2 

74.5 
75.6 
76.8 

78.6 
79.8 
80,9 

83.5 

84.7 
86.0 

87.0 

88.2 
89.5 

94.5 
95.8 
97.0 

97.5 
98.8 
100.1 

59 
60 

61.1 
62.1 

64.2 
65.2 

67.9 
69.0 

73.3 

74.4 

77.9 
79.1 

82.1 
83.3 

87.2 
88.4 

90.7 
92.0 

98.3 

99.6 

101.4 
102.7 

61 

63.2 

66.3 

70.0 

75.5 

80.2 

84.5 

89.6 

93.2 

100.9 

104.0 

62 

64.2 

67.3 

71.1 

76.6 

81.4 

85.7 

90,8 

94.4 

102.2 

105.3 

63 

65.2 

68,4 

72.2 

77.7 

82.5 

86.8 

92.0 

95.6 

103.4 

106  6 

64 

66.2 

69.4 

73.3 

78.9 

83.7 

88.0 

93.2 

96.9 

104.7 

107.9 

65 

67.2 

70.5 

74.4 

80.0 

84.8 

89.2 

94.4 

98.1 

106.0 

109.2 

66 

68.3 

71.5 

75.4 

81.1 

86.0 

90.3 

95.6 

99.3 

107.3 

110.5 

67 

69.3 

72.6 

76.5 

82.2 

87.1 

91.5 

96.8 

100.6 

108.5 

111.7 

68 

70.3 

73.6 

77.6 

83.3 

88.3 

92.7 

98.0 

101.8 

109.8 

113.0 

69 

71.3 

74.6 

78.6 

84.4 

89.4 

93.9 

99.2 

103.0 

111.1 

114.3 

70 

72.4 

75.7 

79.7 

85.5 

90.5 

95.0 

100.4 

104.2 

112.3 

115.6 

71 

73.4 

76.7 

80.8 

86.6 

91.7 

96.2 

101.6 

105.4 

113.6 

116.9 

72 

74.4 

77.8 

81.9 

87.7 

92.8 

97.4 

102.8 

106.6 

114.8 

118.1 

73 

75.4 

78.8 

82.9 

88.8 

93.9 

98.5 

104.0 

107.9 

116.1 

119.4 

74 

76.4 

79.9 

84.0 

90.0 

95.1 

99.7 

105.2 

109.1 

117.3 

120.7 

75 

77.5 

80.9 

85.1 

91.1 

96.2 

100.8 

106.4 

110.3 

118.6 

121,9 

76 

78.5 

82.0 

86.1 

92.2 

97.4 

102.0 

107.6 

111.5 

119.9 

123.2 

77 

79.5 

83.0 

87.2 

93.3 

98.5 

103.2 

108.8 

112.7 

121.1 

124.5 

78 

80.5 

84.0 

88.3 

94.4 

99.6 

104.3 

110.0 

113.9 

122.3 

125.7 

7Q 

81.5 

85.1 

89.3 

95.5 

100.7 

105.5 

111.1 

115.1 

123.6 

127.0 

80 

82.0 

86.1 

90.4 

96.6 

101.9 

106.6 

112.3 

116.3 

124.8 

128.3 

81 

83.6 

87.2 

91.5 

97.7 

103.0 

107.8 

113.5 

117.5 

126.1 

129.5 

82 

84.6 

88.2 

92.5 

98.8 

104.1 

108.9 

114.7 

118.7 

127.3 

130.8 

83 

85.6 

89.2 

93.6 

99.9 

105.3 

110.1 

115.9 

119.9 

128.6 

132.0 

84 

86.6 

90.3 

94.7 

101.0 

106.4 

111.2 

117.1 

121.1 

129.8 

133,3 

85 

87.7 

91.3 

95.7 

102.1 

107.5 

112.4 

118.2 

122.3 

131.0 

134.5 

86 

88.7 

92.4 

96.8 

103.2 

108.6 

113.5 

119.4 

123.5 

132.3 

135.8 

87 

89.7 

93.4 

97.9 

104.3 

109.8 

114.7 

120.6 

124.7 

133.5 

137.0 

88 

90.7 

94.4 

98.9 

105.4 

110.9 

115.8 

121.8 

125.9 

134.7 

138,3 

89 

91.7 

95.5 

100.0 

106.5 

112.0 

117.0 

122.9 

127.1 

136.0 

139.5 

90 

92.8 

96.5 

101.1 

107.6 

113.1 

118.1 

124.1 

128.3 

137.2 

140.8 

91 

93.8 

97.6 

102.1 

108.7 

114.3 

119.3 

125.3 

129.5 

138.4 

142.0 

92 

94.8 

98.6 

103.2 

109.8 

115.4 

120.4 

126.5 

130.7 

139.7 

143.3 

93 

95.8 

99.6 

104.2 

110.9 

116.5 

121.6 

127.6 

131.9 

140.9 

144.5 

94 

96,8 

100.7 

105.3 

111.9 

117.6 

122.7 

128.8 

133.1 

142.1 

145.8 

95 

97.9 

101.7 

106.4 

113.0 

118.8 

123.9 

130.0 

134.2 

143.3 

147,0 

96 

98.9 

102.8 

107.4 

114.1 

119.9 

125.0 

131.1 

135.4 

144.6 

148.2 

97 

99.9 

103.8 

108.5 

115.2 

121.0 

126.1 

132.3 

136.6 

145.8 

149.5 

98 

100.9 

104.8 

109.5 

116.3 

122,1 

127.3 

133.5 

137.8 

147.0 

150.7 

99 

101.9 

105.9 

110.6 

117.4 

123.2 

128.4 

134.6 

139.0 

148.2 

151.0 

100 

102.9 

106.9 

111.7 

118.5 

124.3 

129.6 

135.8 

140.2 

149.4 

153.2 

C5273 


APPENDIX     51 


CUMULATIVE  /-DISTRIBUTION 


V 

p 

0.75 

0.8O 

O.85 

O.90 

O.95 

0,975 

0.995 

O.9995 

1 

1.OOO5 

1.376 

1.963 

3.078 

6.314 

12.706 

63.657 

636.619 

2 

0.816 

1.061 

1.386 

1.886 

2.920 

4.3O3 

9.925 

31.598 

3 

0.765 

0.978 

1.25O 

1.638 

2.353 

3.182 

5.841 

12.941 

4 

0.741 

0.941 

1.190 

1.533 

2.132 

2.776 

4.604 

8.61O 

5 

0.727 

0.92O 

1.156 

1.476 

2.O15 

2.571 

4.032 

6.859 

6 

0.718 

0.906 

1.134 

1.440 

1  .943 

2.447 

3.707 

5.959 

7 

0.711 

O.896 

1.119 

1.415 

1.895 

2.365 

3.499 

5.405 

8 

0.7O6 

0.889 

1.108 

1.397 

1.86O 

2.306 

3.355 

5.041 

9 

O.7O3 

O.883 

1.10O 

1.383 

1.833 

2.262 

3.250 

4.781 

10 

0.7OO 

O.879 

1.093 

1.372 

1.812 

2.228 

3.169 

4.587 

11 

0.697 

O.876 

1.088 

1.363 

1.796 

2.201 

3.106 

4.437 

12 

O.695 

O.873 

1.O83 

1.356 

1.782 

2.179 

3.055 

4.318 

13 

O.694 

O.87O 

1.079 

1.35O 

1.771 

2.160 

3,012 

4.221 

14 

0.692 

O.868 

1.076 

1.345 

1.761 

2.145 

2.977 

4.140 

15 

0.691 

0.866 

1.074 

1.341 

1.753 

2.131 

2.947 

4.073 

16 

0.690 

O.866 

1.071 

1.337 

1.746 

2.120 

2.921 

4.015 

17 

O.689 

O.863 

1.O69 

1.333 

1.740 

2.110 

2.898 

3.965 

18 

O.688 

O.862 

1.O67 

1.33O 

1.734 

2.101 

2.878 

3.922 

19 

0.688 

0.861 

1.066 

1.328 

1.729 

2.O93 

2.861 

3.883 

20 

0.687 

0.86O 

1.064 

1.325 

1.725 

2.086 

2.845 

3.850 

21 

O.686 

0.859 

1.063 

1.323 

1.721 

2.080 

2.831 

3.819 

22 

O.686 

0.858 

1.061 

1.321 

1.717 

2.074 

2.819 

3.792 

23 

0.685 

0.858 

1.060 

1.319 

1.714 

2.069 

2.8O7 

3.767 

24 

O.685 

0.857 

1.059 

1.318 

1.711 

2.064 

2.797 

3.745 

25 

O.684 

O.856 

1.058 

1.316 

1.7O8 

2.06O 

2.787 

3.725 

26 

0.684 

0.856 

1.058 

1.315 

1.706 

2.056 

2.779 

3.7O7 

27 

0.684 

0.855 

1.057 

1.314 

1.7O3 

2.O52 

2.771 

3,690 

28 

0.683 

0.855 

1.056 

1.313 

1.7O1 

2.048 

2,763 

3.674 

29 

O.683 

0.854 

1.O55 

1.311 

1.699 

2.045 

2.756 

3.659 

30 

O.683 

0.854 

1.O55 

1.310 

1.697 

2.042 

2.750 

3.646 

35 

0.682 

O.852 

1.052 

1.306 

1.690 

2.03O 

2.724 

3.591 

40 

O.681 

0.851 

1.050 

1.303 

1.684 

2.021 

2.7O4 

3.551 

45 

O.680 

0.85O 

1.O48 

1.3O1 

1.680 

2.014 

2.69O 

3.52O 

50 

O.68O 

0.849 

1.047 

1.299 

1.676 

2.0O8 

2.678 

3.496 

55 

O.679 

0.849 

1.047 

1.297 

1.673 

2.004 

2.669 

3.476 

60 

O.679 

0.848 

1.046 

1.296 

1.671 

2.0OO 

2.660 

3.460 

70 

0.678 

0.847 

1.045 

1.294 

1.667 

1.994 

2.648 

3.435 

80 

0.678 

0.847 

1.044 

1.293 

1.665 

1,990 

2.638 

3.416 

90 

0.678 

0.846 

1.043 

1.291 

1.662 

1.987 

2.632 

3.402 

100 

0.677 

0.846 

1.042 

1.290 

1.661 

1.984 

2.626 

3.39O 

20O 

0.676 

0.844 

1.O39 

1.286 

1.653 

1.972 

2.601 

3.34O 

300 

O.676 

0.843 

1.O38 

1.285 

1.650 

1.968 

2.592 

3.323 

400 

0.676 

0.843 

1.038 

1.284 

1.649 

1.966 

2.588 

3.315 

500 

0.676 

0.843 

1.O37 

1.284 

1.648 

1.965 

2.586 

3.31O 

1OOO 

O.675 

0.842 

1,037 

1.283 

1.647 

1.962 

2.581 

3.301 

CO 

0  .  67449 

0.84162 

1.O3643 

1.28155 

1.64485 

1  .  95996 

2.57582 

3.29053 

1  Partly  from  Table  III  of  R.  A.  Fisher  and  Frank  Yates,  Statistical  Tables  for 
Biological,  Agricultural  and  Medical  Research,  third  ed.,  Oliver  and  Boyd,  Edin 
burgh,  1948.  By  permission  of  the  authors  and  publishers. 

2  Entries  in  the  table  are  values  of  tp  "where 


APPENDIX    61 

CUMULATIVE  F-DISTRIBUTION 

i  Reproduced  from  Table  A-7c  of  W.  J.  Dixon  and  F.  J.  Massey,  Introduction 
to  Stat^st^cal  Analysis,  second  ed0  McGraw-Hill  Book  Company,  Inc.,  New 
York,  1957.  By  permission  of  the  authors  and  publishers.  However,  since  most 
of  the  values  in  Dixon  and  Massey  were  extracted  from  other  publications, 
permission  was  also  requested  of  the  primary  sources  noted  below.  In  each  case' 
permission  was  granted  to  reproduce  the  needed  material. 

(a)    All  values  for  v19  vz  equal  to   50,   100,   200,   and  500  are  from  A.  Hald, 
Statistical  Tables  and  Formulas,  John  Wiley  and  Sons,  Inc.,  New  York 


(b)  For  cumulative  proportions  .5,   ,75,   .9,   .95,  .975,  .99,  and  .995,  most  of 
the  values  are  from  M.   Merrington  and   C.   M.  Thompson,   "Tables  of 
percentage   points   of  the   inverted    beta    <JF}    distribution/'    Biometrika 
Vol.  33,  Part  I,  April,  1943,  pp.  74-87. 

(c)  For  cumulative  proportions  .999,  the  values  are  from  C.  C.  Colcord  and 
L.  S.  Deming,  "The  one-tenth  percent  level  of  Z,"  Sankhya,  Vol.  2,  Part  4, 
Dec.,  1936,  pp.  423-24. 

(d)  As  noted  in   Dixon   and   Massey,   the   remaining  values  were  found  by 
forming  reciprocals  or  by  interpolation. 

2  Entries  in  the  table  are  values  of  Fp  where 


[5293 


1 

2 

3 

4 

5 

6 

7 

8 

9 

10 

11 

12 

^2 

1 

.0005 

.0662 

.0350 

.0238 

,0294 

.016 

.022 

.027 

.032 

.036 

.039 

.042 

,045 

.001 

.0525 

.ono 

.0260 

.013 

.021 

.028 

.034 

.039 

.044 

.048 

.051 

.054 

.005 

.0462 

-0251 

.018 

.032 

.044 

.054 

.062 

.068 

.073 

.078 

.082 

.085 

.010 

.0325 

.010 

.029 

.047 

.062 

.073 

.082 

.089 

.095 

,100 

.104 

.107 

.025 

.0215 

.026 

.057 

.082 

.100 

.113 

.124 

.132 

.139 

.144 

.149 

.153 

.05 

.0262 

.054 

.099 

.130 

.151 

.167 

.179 

.188 

.195 

.201 

.207 

.211 

.10 

.025 

.117 

.181 

.220 

.246 

.265 

.279 

.289 

.298 

.304 

.310 

.315 

.25 

.172 

.389 

.494 

.553 

.591 

.617 

.637 

.650 

.661 

.670 

.680 

.684 

.50 

1.00 

1.50 

1.71 

1.82 

1.89 

1.94 

1.98 

2.00 

2.03 

2.04 

2.05 

2.07 

.75 

5.83 

7.50 

8.20 

8.58 

8.82 

8.98 

9.10 

9.19 

9.26 

9.32 

9.36 

9.41 

.90 

39.9 

49.5 

53.6 

55.8 

57.2 

58.2 

58.9 

59.4 

59.9 

60.2 

60.5 

60.7 

.95 

161 

200 

216 

225 

230 

234 

237 

239 

241 

242 

243 

244 

.975 

648 

800 

864 

900 

922 

937 

948 

957 

963 

969 

973 

977 

.99 

4051 

500i 

5401 

562i 

5761 

5861 

5931 

5981 

6021 

6061 

60S1 

611i 

.995 

1622 

2002 

2162 

2252 

2312 

2342 

2372 

2392 

2412 

2422 

2432 

2442 

.999 

4063 

5003 

5403 

5623 

5763 

5863 

5933 

5983 

6023 

6063 

6093 

611 

.9995 

162* 

200* 

216* 

225* 

2314 

234* 

2374 

2394 

2414 

242* 

243* 

244* 

2 

.0005 

.0650 

.0350 

.0=42 

.011 

.020 

.029 

.037 

.044 

.050 

.056 

.061 

.065 

.001 

.0520 

.0210 

.0268 

.016 

.027 

.037 

.046 

.054 

.061 

.067 

.072 

.077 

.005 

.0*50 

.0250 

.020 

.038 

.055 

,069 

.081 

.091 

.099 

.106 

.112 

.118 

.01 

.0320 

.010 

.032 

.056 

.075 

.092 

.105 

.116 

.125 

.132 

.139 

.144 

.025 

.0213 

.026 

.062 

.094 

.119 

.138 

.153 

.165 

.175 

.183 

.190 

.196 

.05 

.0^50 

.053 

.105 

.144 

.173 

.194 

.211 

.224 

.235 

.244 

.251 

.257 

,10 

.020 

.111 

.183 

.231 

.265 

.289 

.307 

.321 

.333 

,342 

.350 

.356 

,25 

.133 

.333 

.439 

.500 

.540 

.568 

.588 

.604 

.616 

.626 

.633 

.641 

.50 

.667 

1.00 

1.13 

1.21 

1.25 

1.28 

1.30 

1.32 

1.33 

1,34 

1.35 

1.36 

.75 

2.57 

3.00 

3.15 

3.23 

3.28 

3.31 

3.34 

3.35 

3.37 

3.38 

3.39 

3.39 

.90 

8.53 

9.00 

9.16 

9.24 

9.29 

9.33 

9.35 

9.37 

9.38 

9.39 

9.40 

9.41 

.95 

18.5 

19.0 

19.2 

19.2 

19.3 

19.3 

19.4 

19.4 

19.4 

19.4 

19.4 

19.4 

.975 

38.5 

39.0 

39.2 

39.2 

39.3 

39.3 

39.4 

39.4 

39.4 

39.4 

39.4 

39.4 

.99 

98.5 

99.0 

99.2 

99.2 

99.3 

99.3 

99.4 

99.4 

99.4 

99.4 

99.4 

99.4 

.995 

198 

199 

199 

199 

199 

199 

199 

199 

199 

199 

199 

199 

.999 

998 

999 

999 

999 

999 

999 

999 

999 

999 

999 

999 

999 

.9995 

2001 

200i 

2001 

2001 

2001 

2001 

2001 

200i 

2001 

20Qi 

2001 

2001 

3 

.0005 

.0«46 

.0350 

.0244 

.012 

.023 

.033 

.043 

.052 

.060 

.067 

.074 

.079 

.001 

.0*19 

.0210 

-0271 

.018 

.030 

.042 

.053 

.063 

.032 

.079 

.086 

.093 

.005 

.0*46 

.0250 

.021 

.041 

.060 

.077 

.092 

.104 

.115 

.124 

.132 

-138 

.01 

.On9 

.010 

.034 

.060 

.083 

.102 

.118 

.132 

.143 

.153 

.161 

.168 

.025 

.0212 

.026 

.065 

.100 

.129 

.152 

.170 

.185 

.197 

.207 

.216 

.224 

.05 

.0246 

.052 

.108 

.152 

.185 

.210 

.230 

.246 

.259 

.270 

.279 

.287 

.10 

.019 

.109 

,185 

.239 

.276 

.304 

.325 

.342 

.356 

.367 

.376 

.384 

.25 

.122 

.317 

.424 

.489 

.531 

.561 

.582 

.600 

.613 

.624 

.633 

.641 

.50 

.585 

.881 

1.00 

1.06 

1.10 

1.13 

1.15 

1.16 

1.17 

1.18 

1.19 

1.20 

.75 

2.02 

2.28 

2.36 

2.39 

2.41 

2.42 

2.43 

2.44 

2.44 

2.44 

2.45 

2.45 

.90 

5.54 

5.46 

5.39 

5.34 

5.31 

5.28 

5.27 

5.25 

5.24 

5.23 

5.22 

5.22 

.95 

10.1 

9.55 

9.28 

9.12 

9.01 

8.94 

8.89 

8,85 

8.81 

8.79 

8.76 

8.74 

.975 

17.4 

16.0 

15.4 

15.1 

14.9 

14.7 

14.6 

14.5 

14.5 

14.4 

14.4 

14.3 

.99 

34.1 

30.8 

29.5 

28.7 

28.2 

27.9 

27.7 

27.5 

27.3 

27,2 

27.1 

27.1 

.995 

55.6 

49.8 

47.5 

46.2 

45.4 

44,3 

44.4 

44.1 

43.9 

43.7 

43.5 

43.4 

.999 

167 

149 

141 

137 

135 

133 

132 

131 

130 

129 

129 

128 

.9995 

266 

237 

225 

218 

214 

211 

209 

208 

207 

206 

204 

204 

Read  .0^56  0    .00056,  200*  as  2000,  1624  as  1620000,  etc. 


,t 

P    ^\ 

15 

20 

24 

30 

40 

50 

60 

100 

120 

200 

500 

CO 

1 

.0005 

.051 

.058 

.062 

.066 

.069 

.072 

.074 

.077 

,078 

.080 

.081 

.083 

.001 

.060 

.067 

.071 

.075 

.079 

.082 

.084 

.087 

.088 

.089 

.091 

.092 

.005 

.093 

.101 

.105 

.109 

.113 

.116 

.118 

.121 

.122 

.124 

.126 

.127 

.01 

.115 

.124 

.128 

.132 

.  137 

.139 

.141 

.145 

.146 

.148 

.150 

.151 

.025 

.  161 

.170 

.175 

.180 

.184 

.187 

.189 

.193 

.194 

.196 

.198 

.199 

.05 

.220 

.230 

.235 

.240 

.245 

.248 

.250 

.254 

.255 

.257 

.259 

.261 

.10 

.325 

.336 

.342 

.347 

.353 

.356 

.358 

.362 

.364 

.366 

.368 

.370 

.25 

.698 

.712 

.719 

.727 

.734 

.738 

.741 

.747 

.749 

.752 

.754 

,756 

.50 

2.09 

2.12 

2.13 

2.15 

2.16 

2.17 

2,17 

2.18 

2.  IS 

2.19 

2.19 

2.20 

.75 

9.49 

9.58 

9.63 

9.67 

9.71 

9.74 

9.76 

9.78 

9.80 

9.82 

9.84 

9.85 

.90 

61.2 

61.7 

62.0 

62.3 

62.5 

62.7 

62.8 

63.0 

63.1 

63.2 

63.3 

63.3 

.95 

246 

248 

249 

250 

251 

252 

252 

253 

253 

254 

254 

254 

.975 

985 

993 

997 

lOOi 

1011 

1011 

1011 

1011 

1011 

102L 

1021 

102* 

.99 

6161 

62li 

623i 

626i 

6291 

63  0L 

631i 

6331 

6341 

635i 

6361 

6371 

.995 

2462 

2482 

249* 

2502 

2512 

2522 

253= 

2532 

2542 

254* 

2542 

255* 

.999 

6163 

621» 

6233 

6263 

6293 

63  O3 

6313 

6333 

6343 

6353 

637* 

.9995 

246* 

248* 

249* 

250* 

251* 

252* 

252* 

253* 

253* 

253* 

254* 

2544 

2 

.0005 

.076 

.088 

.094 

,101 

.108 

.113 

.116 

.122 

.124 

.127 

.130 

,132 

.001 

.088 

.100 

.107 

.114 

.121 

.126 

.129 

.135 

.137 

.140 

.143 

.145 

,005 

.130 

.143 

.150 

.157 

.165 

.169 

.173 

.179 

.181 

.184 

.187 

.189 

.01 

.157 

.171 

.178 

.186 

.193 

.198 

.201 

.207 

.209 

.212 

.215 

.217 

.025 

.210 

.224 

.232 

.239 

.247 

.251 

.255 

.261 

.263 

.266 

.269 

.271 

.05 

.272 

.286 

.294 

.302 

.309 

.314 

.317 

.324 

.326 

.329 

.332 

.334 

.10 

.371 

.386 

.394 

.402 

.410 

.415 

.418 

.424 

.426 

.429 

.433 

.434 

.25 

.657 

.672 

.680 

.689 

.697 

.702 

.705 

.711 

.713 

.716 

.719 

.721 

.50 

1.38 

1.39 

1.40 

1.41 

1.42 

1.42 

1.43 

1.43 

1.43 

1.44 

1.44 

1.44 

.75 

3.41 

3.43 

3.43 

3.44 

3.45 

3.45 

3.46 

3.47 

3.47 

3.48 

3.48 

3.48 

.90 

9.42 

9.44 

9.45 

9.46 

9.47 

9.47 

9.47 

9.48 

9.48 

9.49 

9.49 

9.49 

.95 

19.4 

19.4 

19.5 

19.5 

19.5 

19.5 

19.5 

19.5 

19.5 

19.5 

19.5 

19.5 

.975 

39.4 

39.4 

39.5 

39.5 

39.5 

39.5 

39.5 

39.5 

39.5 

39.5 

39.5 

39.5 

.99 

99.4 

99.4 

99.5 

99.5 

99.5 

99.5 

99.5 

99.5 

99.5 

99.5 

99.5 

99.5 

.995 

199 

199 

199 

199 

199 

199 

199 

199 

199 

199 

199 

200 

.999 

999 

999 

999 

999 

999 

999 

999 

999 

999 

999 

999 

999 

.9995 

2001 

2001 

2001 

2001 

20Q1 

2001 

2001 

2001 

2001 

2001 

2001 

2001 

3 

.0005 

.093 

.109 

.117 

.127 

.136 

.143 

.147 

.156 

.158 

.162 

.166 

.169 

.001 

.107 

.123 

.132 

.142 

.152 

.158 

.162 

.171 

.173 

.177 

.181 

.184 

.005 

.154 

.172 

.181 

.191 

.201 

.207 

.211 

.220 

.222 

.227 

.231 

.234 

.01 

.185 

.203 

.212 

.222 

.232 

.238 

.242 

.251 

.253 

.258 

.262 

.264 

.025 

.241 

.259 

.269 

.279 

.289 

.295 

.299 

.308 

.310 

.314 

.318 

.321 

.05 

.304 

,323 

.332 

,342 

.352 

.358 

.363 

.370 

.373 

.377 

.382 

.384 

.10 

.402 

.420 

.430 

.439 

.449 

.455 

.459 

.467 

.469 

.474 

.476 

.480 

.25 

.658 

,675 

.684 

.693 

.702 

.708 

.711 

.719 

.721 

.724 

.728 

.730 

.50 

1.21 

1.23 

1.23 

1.24 

1.25 

1.25 

1.25 

1.26 

1.26 

1.26 

1.27 

1,27 

.75 

2.46 

2.46 

2,46 

2.47 

2.47 

2.47 

2.47 

2.47 

2.47 

2.47 

2.47 

2.47 

.90 

5.20 

5.18 

5.18 

5.17 

5.16 

5.15 

5.15 

5.14 

5.14 

5.14 

5.14 

5.13 

.95 

8.70 

8.66 

8.63 

8.62 

8.59 

8.58 

8.57 

8.55 

8.55 

8.54 

8.53 

8.53 

.975 

14.3 

14.2 

14.1 

14.1 

14.0 

14.0 

14.0 

14.0 

13.9 

13.9 

13.9 

13.9 

.99 

26.9 

26.7 

26.6 

26.5 

26.4 

26.4 

26.3 

26.2 

26.2 

26.2 

26.1 

26.1 

.995 

43.1 

42.8 

42.6 

42.5 

42.3 

42.2 

42.1 

42.0 

42,0 

41.9 

41.9 

41.8 

.999 

127 

126 

126 

125 

125 

125 

124 

124 

124 

124 

124 

123 

.9995 

203 

201 

200 

199 

199 

198 

198 

197 

197 

197 

196 

196 

[5311 


"2 

1 

2 

3 

4 

5 

6 

7 

8 

9 

10 

11 

12 

4 

.0005 

.0«44 

.0*50 

.0246 

.013 

.024 

.036 

.047 

.057 

.066 

.075 

.082 

.089 

.001 

.0*18 

.0210 

.0273 

.019 

.032 

.046 

.058 

.069 

.079 

.089 

.097 

.104 

.005 

.CH44 

.0250 

.022 

.043 

.064 

.083 

.100 

.114 

.126 

.137 

.145 

.153 

.01 

.0318 

.010 

.035 

.063 

.088 

.109 

.127 

.143 

.156 

.167 

.176 

.185 

.025 

.0211 

.026 

.066 

.104 

.135 

.161 

.181 

.198 

.212 

.224 

.234 

.243 

05 

.0244 

.052 

.110 

.157 

,193 

.221 

.243 

.261 

.275 

.288 

.298 

.307 

.10 

.018 

.108 

.187 

.243 

,284 

.314 

.338 

.356 

.371 

.384 

.394 

.403 

.25 

.117 

.309 

.418 

.484 

.528 

.560 

.583 

.601 

.615 

.627 

.637 

.645 

.50 

.549 

.828 

.941 

1.00 

1.04 

1.06 

1.08 

1.09 

1.10 

1.11 

1.12 

1.13 

.75 

1.81 

2.00 

2.05 

2.06 

2.07 

2.08 

2.08 

2.08 

2.08 

2.08 

2.08 

2.08 

.90 

4.54 

4.32 

4.19 

4.11 

4.05 

4.01 

3.98 

3.95 

3.94 

3.92 

3.91 

3.90 

.95 

7.71 

6.94 

6.59 

6.39 

6.26 

6.16 

6.09 

6.04 

6.00 

5.96 

5.94 

5.91 

.975 

12.2 

10.6 

9.98 

9.60 

9.36 

9.20 

9.07 

8.98 

8.90 

8.84 

8.79 

8.75 

.99 

21.2 

18.0 

16.7 

16.0 

15.5 

15.2 

15.0 

14.8 

14.7 

14.5 

14.4 

14.4 

.995 

31.3 

26.3 

24.3 

23.2 

22.5 

22.0 

21.6 

21.4 

21.1 

21.0 

20.8 

20.7 

.999 

74.1 

61.2 

56.2 

53.4 

51.7 

50.5 

49.7 

49.0 

48.5 

48.0 

47.7 

47.4 

.9995 

106 

87.4 

80.1 

76.1 

73.6 

71.9 

70.6 

69.7 

68.9 

68.3 

67.8 

67.4 

5 

.0005 

.0«43 

.0350 

.0247 

.014 

.025 

.038 

.050 

.061 

.070 

.081 

.089 

.096 

.001 

.0*17 

.0210 

.0275 

,019 

.034 

.048 

.062 

.074 

.085 

.095 

.104 

.112 

.005 

.CH43 

.0250 

.022 

.045 

,067 

.087 

.105 

.120 

.134 

.146 

.156 

.165 

.01 

.0317 

.010 

.035 

.064 

.091 

.114 

.134 

.151 

.165 

.177 

.188 

.197 

.025 

.0211 

.025 

.067 

.107 

.140 

.167 

.189 

.208 

.223 

.236 

.248 

.257 

.05 

.0243 

.052 

.111 

.160 

.198 

.228 

.252 

.271 

.287 

.301 

.313 

.322 

,10 

.017 

.  108 

.188 

.247 

.290 

.322 

.347 

.367 

.383 

.397 

.408 

.418 

.25 

.113 

.305 

.415 

.483 

.528 

.560 

.584 

.604 

.618 

.631 

.641 

.650 

.50 

.528 

.799 

.907 

.965 

1.00 

1.02 

1.04 

1.05 

1.06 

1.07 

1.08 

1.09 

.75 

1.69 

1.85 

1.88 

1.89 

1.89 

1.89 

1.89 

1.89 

1.89 

1.89 

1.89 

1.89 

.90 

4.06 

3.78 

3.62 

3.52 

3.45 

3.40 

3.37 

3.34 

3.32 

3.30 

3.28 

3.27 

.95 

6.61 

5.79 

5.41 

5.19 

5.05 

4,95 

4.88 

4.82 

4.77 

4.74 

4.71 

4.68 

.975 

10.0 

8.43 

7.76 

7.39 

7.15 

6.98 

6.85 

6.76 

6.68 

6.62 

6.57 

6.52 

.99 

16.3 

13.3 

12.1 

11.4 

11.0 

10.7 

10.5 

10.3 

10.2 

10.1 

9.96 

9.89 

.995 

22.8 

18.3 

16.5 

15.6 

14.9 

14.5 

14.2 

14.0 

13.8 

13.6 

13.5 

13.4 

.999 

47.2 

37.1 

33.2 

31.1 

29.7 

28.8 

28.2 

27.6 

27,2 

26.9 

26.6 

26.4 

.9995 

63,6 

49.8 

44.4 

41.5 

39.7 

38.5 

37.6 

36.9 

36.4 

35.9 

35.6 

35.2 

6 

.0005 

.0^43 

.0350 

.0247 

.014 

.026 

.039 

.052 

.064 

.075 

.085 

.094 

.103 

.001 

.0617 

.0210 

.0275 

.020 

.035 

.050 

.064 

.078 

.090 

.101 

.111 

.119 

.005 

.0*43 

.0250 

.022 

.045 

.069 

.090 

.109 

.126 

.140 

.153 

.164 

.174 

.01 

.0317 

.010 

.036 

.066 

.094 

.118 

.139 

.157 

.172 

.186 

.197 

.207 

.025 

.0211 

.025 

.068 

.109 

.143 

.172 

,195 

.215 

.231 

.246 

.258 

.268 

.05 

.0243 

.052 

.112 

.162 

.202 

.233 

.259 

.279 

.296 

.311 

.324 

.334 

.10 

.017 

.107 

.189 

.249 

.294 

.327 

.354 

.375 

.392 

.406 

.418 

.429 

.25 

.111 

.302 

.413 

.481 

.524 

.561 

.586 

.606 

.622 

.635 

.645 

.654 

.50 

.515 

.780 

.886 

.942 

.977 

1.00 

1.02 

1.03 

1.04 

1.05 

1.05 

1.06 

.75 

1.62 

1.76 

1.78 

1.79 

1.79 

1.78 

1.78 

1.78 

1.77 

1.77 

1.77 

1.77 

.90 

3.78 

3.46 

3.29 

3.18 

3.11 

3.05 

3.01 

2.98 

2.96 

2.94 

2.92 

2.90 

.95 

5.99 

5.14 

4,76 

4.53 

4.39 

4.28 

4.21 

4.15 

4.10 

4.06 

4.03 

4.00 

.975 

8.81 

7.26 

6.60 

6.23 

5.99 

5.82 

5.70 

5.60 

5.52 

5.46 

5.41 

5.37 

.99 

13.7 

10.9 

9.78 

9.15 

8.75 

8.47 

8.26 

8.10 

7.98 

7.87 

7.79 

7.72 

.995 

18.6 

14.5 

12.9 

12.0 

11.5 

11.1 

10.8 

10.6 

10.4 

10.2 

10.1 

10.0 

.999 

35.5 

27.0 

23.7 

21.9 

20.8 

20.0 

19.5 

19.0 

18.7 

18.4 

18.2 

18.0 

.9995 

46.1 

34,8 

30.4 

28.1 

26.6 

25.6 

24.9 

24.3 

23.9 

23.5 

23.2 

23.0 

[532] 


>>Z 

15 

20 

24 

30 

40 

50 

60 

100 

120 

200 

5OO 

00 

4 

.0005 

.105 

.125 

.135 

.147 

.159 

.166 

.172 

.183 

.186 

.191 

.196 

.200 

.001 

.121 

.141 

,152 

.163 

.176 

.183 

.  188 

.200 

.202 

.208 

.213 

.217 

.005 

.172 

.193 

.204 

.216 

.229 

.237 

.242 

.253 

.255 

.260 

.266 

.269 

.01 

.204 

.226 

.237 

.249 

.261 

.269 

.274 

.285 

.287 

.293 

.298 

.301 

.025 

.263 

.284 

.296 

.308 

.320 

.327 

.332 

.342 

.346 

.351 

.356 

.359 

.05 

.327 

.349 

.360 

.372 

.384 

.391 

.396 

.407 

.409 

.413 

.418 

.422 

.10 

.424 

.445 

.456 

.467 

,478 

.485 

.490 

.500 

.502 

.508 

.510 

.514 

.25 

.664 

.683 

.692 

.702 

.712 

.718 

.722 

.731 

.733 

.737 

.740 

,743 

.50 

1.14 

1.15 

1.16 

1.16 

1.17 

1.18 

1.18 

1.18 

1.18 

1.19 

1.19 

1.19 

.75 

2.08 

2.08 

2.08 

2.08 

2.08 

2.08 

2.08 

2.08 

2.08 

2.08 

2.08 

2.08 

.90 

3.87 

3.84 

3.83 

3.82 

3.80 

3.80 

3.79 

3.78 

3.78 

3.77 

3.76 

3.76 

.95 

5.8<5 

5.80 

5.77 

5.75 

5.72 

5.70 

5.69 

5.66 

5.66 

5,65 

5.64 

5.63 

.975 

8.66 

8.56 

8.51 

8.46 

8.41 

8.38 

8.36 

8.32 

8.31 

8.29 

8.27 

8.26 

.99 

14.2 

14.0 

13.9 

13.8 

13.7 

13.7 

13.7 

13.6 

13.6 

13.5 

13.5 

13.5 

.995 

20.4 

20.2 

20.0 

19.9 

19.8 

19.7 

19.6 

19.5 

19.5 

19.4 

19.4 

19.3 

.999 

46.8 

46.1 

45.8 

45.4 

45.1 

44.9 

44.7 

44.5 

44.4 

44,3 

44.1 

44,0 

.9995 

66.5 

65.5 

65.1 

64.6 

64.1 

63.8 

63.6 

63.2 

63.1 

62.9 

62.7 

62.6 

5 

.0005 

.115 

.137 

.150 

.163 

.177 

.186 

.192 

.205 

.209 

.216 

.222 

.226 

.001 

.132 

.155 

.167 

.181 

.195 

.204 

.210 

.223 

.227 

.233 

.239 

.244 

,005 

.186 

.210 

,223 

.237 

.251 

.260 

.266 

.279 

.282 

.288 

.294 

.299 

.01 

.219 

.244 

.257 

.270 

.285 

,293 

.299 

,312 

,315 

.322 

.328 

.331 

.025 

.280 

.304 

.317 

.330 

.344 

.353 

.359 

.370 

.374 

.380 

.386 

.390 

.05 

.345 

.369 

.382 

.395 

.408 

.417 

.422 

.432 

.437 

.442 

.448 

.452 

.10 

.440 

,463 

.476 

.488 

.501 

.508 

.514 

.524 

.527 

.532 

.538 

.541 

.25 

.669 

.690 

.700 

.711 

.722 

.728 

.732 

.741 

.743 

.748 

.752 

.755 

,50 

1.10 

1.11 

1.12 

1.12 

1.  13 

1,13 

1.14 

1.14 

1.14 

1.15 

1.15 

1.15 

.75 

1.89 

1.88 

1.88 

1.88 

1.88 

1.88 

1.87 

1.87 

1.87 

1.87 

1.87 

1.87 

.90 

3.24 

3.21 

3.19 

3.17 

3.16 

3.15 

3,14 

3.13 

3.12 

3.12 

3.11 

3.10 

.95 

4.62 

4,56 

4.53 

4.50 

4.46 

4.44 

4.43 

4.41 

4.40 

4.39 

4.37 

4.36 

.975 

6.43 

6.33 

6.28 

6.23 

6.18 

6.14 

6.12 

6.08 

6.07 

6.05 

6,03 

6.02 

.99 

9.72 

9.55 

9.47 

9.38 

9.29 

9.24 

9.20 

9-13 

9.11 

9.08 

9.04 

9.02 

.995 

13.1 

12,9 

12.8 

12.7 

12.5 

12.5 

12.4 

12.3 

12.3 

12.2 

12.2 

12.1 

.999 

25.9 

25.4 

25.1 

24.9 

24.6 

24.4 

24.3 

24.1 

24.1 

23.9 

23.8 

23.8 

.9995 

34.6 

33.9 

33.5 

33.1 

32.7 

32.5 

32.3 

32.1 

32.0 

31.8 

31.7 

31.6 

6 

.0005 

.123 

.148 

.162 

.177 

.193 

.203 

,210 

.225 

.229 

.236 

.244 

.249 

.001 

.141 

.166 

.180 

.195 

.211 

.222 

.229 

.243 

.247 

.255 

.262 

.267 

.005 

.197 

.224 

.238 

.253 

.269 

.279 

.286 

.301 

.304 

.312 

.318 

.324 

.01 

.232 

.258 

.273 

.288 

.304 

.313 

.321 

.334 

.338 

.346 

.352 

.357 

.025 

.293 

.320 

.334 

.349 

.364 

.375 

.381 

.394 

.398 

.405 

.412 

.415 

.05 

.358 

.385 

.399 

.413 

.428 

.437 

.444 

.457 

.460 

.467 

.472 

.476 

.10 

.453 

.478 

.491 

.505 

.519 

.526 

,533 

.546 

.548 

.556 

.559 

.564 

.25 

.675 

.696 

.707 

.718 

.729 

.736 

.741 

.751 

.753 

,758 

.762 

.765 

.50 

1.07 

1.08 

1.09 

1.10 

1,10 

1.11 

1.11 

1.11 

1.12 

1.12 

1.12 

1.12 

.75 

1.76 

1.76 

1.75 

1.75 

1.75 

1.75 

1,74 

1.74 

1.74 

1.74 

1,74 

1.74 

.90 

2.87 

2.84 

2.82 

2.80 

2.78 

2.77 

2,76 

2.75 

2.74 

2.73 

2.73 

2.72 

.95 

3.94 

3.87 

3.84 

3.81 

3.77 

3.75 

3.74 

3.71 

3.70 

3.69 

3.68 

3.67 

.,975 

5.27 

5.17 

5.12 

5.07 

5.01 

4.98 

4.96 

4.92 

4.90 

4.88 

4.86 

4.85 

.99 

7.56 

7.40 

7.31 

7.23 

7.14 

7.09 

7.06 

6,99 

6.97 

6.93 

6.90 

6.88 

.995 

9.81 

9.59 

9.47 

9.36 

9.24 

9.17 

9.12 

9.03 

9.00 

8.95 

8.91 

8.88 

.999 

17.6 

17.1 

16.9 

16,7 

16.4 

16.3 

16.2 

16.0 

16.0 

15.9 

15.8 

15.7 

.9995 

22.4 

21.9 

21.7 

21.4 

21.1 

20.9 

20.7 

20.5 

20.4 

20.3 

20.2 

20.1 

[533] 


"2 

1 

2 

3 

4 

5 

6 

7 

8 

9 

10 

11 

12 

7 

.0005 

.0642 

.0350 

.0248 

.014 

.027 

.040 

.053 

.066 

.078 

.088 

.099 

.  108 

.001 

.0517 

.0210 

.0276 

.020 

.035 

.051 

.067 

.081 

.093 

.105 

.115 

.125 

.005 

.0*42 

.0250 

.023 

.046 

.070 

.093 

.113 

.130 

.145 

.159 

.171 

.181 

.01 

.0317 

.010 

.036 

.067 

.096 

.121 

.143 

.162 

.178 

.192 

.205 

.216 

.025 

.0210 

.025 

.068 

.110 

.146 

.176 

.200 

.221 

.238 

.253 

.266 

.277 

.05 

.0242 

.052 

.113 

.164 

.205 

.238 

.264 

.286 

.304 

.319 

.332 

.343 

.10 

.017 

.107 

.190 

.251 

.297 

.332 

.359 

.381 

.399 

.414 

.427 

.438 

.25 

.110 

.300 

.412 

.481 

.528 

.562 

.588 

.608 

.624 

.637 

.649 

.658 

.50 

.506 

.767 

.871 

.926 

.960 

.983 

1.00 

1.01 

1.02 

1.03 

1.04 

1.04 

.75 

1.57 

1,70 

1.72 

1.72 

1.71 

1.71 

1.70 

1.70 

1.69 

1.69 

1.69 

1.68 

.90 

3.59 

3.26 

3.07 

2.96 

2.88 

2.83 

2.78 

2.75 

2.72 

2.70 

2.68 

2.67 

.95 

5.59 

4.74 

4.35 

4.12 

3.97 

3.87 

3.79 

3.73 

3.68 

3.64 

3.60 

3,57 

.975 

8.07 

6.54 

5.89 

5.52 

5.29 

5.12 

4.99 

4.90 

4.82 

4.76 

4.71 

4.67 

.99 

12.2 

9.55 

8.45 

7.85 

7.46 

7.19 

6.99 

6.84 

6.72 

6.62 

6.54 

6.47 

.995 

16.2 

12.4 

10.9 

10.0 

9.52 

9.16 

8.89 

8.68 

8.51 

8.38 

8.27 

8,18 

.999 

29.2 

21.7 

18.8 

17.2 

16.2 

15.5 

15.0 

14.6 

14.3 

14.1 

13.9 

13.7 

.9995 

37.0 

27.2 

23.5 

21.4 

20.2 

19.3 

18.7 

18.2 

17.8 

17.5 

17.2 

17.0 

8 

.0005 

.0642 

.0350 

.0248 

.014 

.027 

.041 

.055 

.068 

.081 

.092 

.102 

.112 

.001 

.0517 

.0210 

.0276 

.020 

.036 

.053 

.068 

.083 

.096 

.109 

.120 

.130 

.005 

.0442 

.0250 

,027 

.047 

.072 

.095 

.115 

.133 

.149 

.164 

.176 

.187 

.01 

.0317 

.010 

,036 

.068 

.097 

.123 

.146 

.166 

.183 

.198 

.211 

.222 

.025 

.0210 

.025 

.069 

.111 

.148 

.179 

.204 

.226 

.244 

.259 

.273 

.285 

.05 

.0242 

.052 

.113 

.166 

.208 

.241 

.268 

.291 

.310 

.326 

.339 

.351 

.10 

.017 

.107 

.190 

.253 

.299 

.335 

.363 

.386 

.405 

.421 

.435 

.445 

.25 

.109 

.298 

.411 

.481 

.529 

.563 

.589 

.610 

.627 

.640 

.654 

.661 

.50 

.499 

.757 

.860 

.915 

.948 

.971 

.988 

1.00 

1.01 

1.02 

1.02 

1.03 

,75 

1.54 

1.66 

1.67 

1.66 

1.66 

1.65 

1.64 

1.64 

1.64 

1.63 

1.63 

1.62 

.90 

3.46 

3.  11 

2.92 

2.81 

2.73 

2.67 

2.62 

2.59 

2.56 

2.54 

2.52 

2.50 

.95 

5.32 

4.46 

4.07 

3.84 

3.69 

3,58 

3.50 

3.44 

3.39 

3.35 

3.31 

3.28 

,975 

7.57 

6.06 

5.42 

5.05 

4.82 

4.65 

4.53 

4.43 

4.36 

4.30 

4,24 

4.20 

,99 

11.3 

8.65 

7.59 

7.01 

6.63 

6.37 

6.18 

6.03 

5.91 

5.81 

5.73 

5.67 

,995 

14.7 

11.0 

9.60 

8.81 

8.30 

7.95 

7.69 

7.50 

7.34 

7.21 

7.10 

7.01 

.999 

25.4 

18.5 

15.8 

14.4 

13.5 

12.9 

12.4 

12.0 

11.8 

11.5 

11.4 

11.2 

.9995 

31.6 

22.8 

19.4 

17.6 

16.4 

15.7 

15.1 

14.6 

14.3 

14.0 

13.8 

13.6 

9 

.0005 

.0641 

.0350 

.0248 

.015 

.027 

.042 

.056 

.070 

.083 

.094 

.105 

.115 

.001 

.0517 

.0210 

.0277 

.021 

.037 

.054 

.070 

.085 

.099 

.112 

.123 

.134 

.005 

.0^42 

.0250 

.023 

.047 

.073 

.096 

.117 

.136 

.153 

.168 

.181 

.192 

.01 

.0317 

.010 

.037 

.068 

.098 

.125 

.149 

.169 

.187 

.202 

.216 

.228 

.025 

.0210 

.025 

.069 

.112 

.150 

.181 

.207 

.230 

.248 

.265 

.279 

.291 

.05 

.0240 

.052 

.113 

.167 

.210 

.244 

.272 

.296 

.315 

.331 

.345 

.358 

.10 

.017 

.107 

.191 

.254 

.302 

.338 

.367 

.390 

.410 

.426 

.441 

.452 

.25 

.108 

.297 

.410 

.480 

.529 

.564 

.591 

.612 

.629 

.643 

.654 

.664 

.50 

.494 

.749 

.852 

.906 

.939 

.962 

.978 

.990 

1.00 

1.01 

1.01 

1.02 

.75 

1.51 

1.62 

1.63 

1.63 

1.62 

1.61 

1.60 

1.60 

1.59 

1.59 

1.58 

1.58 

.90 

3.36 

3.01 

2.81 

2.69 

2.61 

2.55 

2.51 

2.47 

2.44 

2.42 

2.40 

2.38 

.95 

5.12 

4.26 

3.86 

3.63 

3.48 

3.37 

3.29 

3.23 

3.18 

3.14 

3.10 

3.07 

.975 

7.21 

5.71 

5.08 

4.72 

4.48 

4.32 

4.20 

4.10 

4.03 

3.96 

3.91 

3.87 

.99 

10.6 

8.02 

6.99 

6.42 

6.06 

5.80 

5.61 

5.47 

5.35 

5.26 

5.18 

5.11 

.995 

13.6 

10.1 

8.72 

7.96 

7.47 

7.13 

6.88 

6.69 

6.54 

6.42 

6.31 

6.23 

.999 

22.9 

16.4 

13.9 

12.6 

11.7 

11.1 

10.7 

10.4 

10.1 

9.89 

9.71 

9.57 

.9995 

28.  0 

19.9 

16.8 

15.1 

14.1 

13.3 

12.8 

12.4 

12.1 

11.8 

11.6 

11.4 

[534] 


*2 

15 

20 

24 

30 

40 

50 

60 

100 

120 

200 

500 

w 

7 

.0005 
.001 

.130 
.148 

.157 
.176 

.172 
.191 

.188 
.208 

.206 
.225 

.217 
.237 

.225 
.245 

.242 
.261 

.246 
.266 

.255 
.274 

.263 
.282 

.268 
.288 

.005 

-206 

.235 

.251 

.267 

.285 

.296 

.304 

.319 

.324 

.332 

.340 

.345 

.01 

.241 

.270 

.286 

.303 

.320 

.331 

.339 

.355 

.358 

.366 

.373 

.379 

.025 

.304 

.333 

.348 

.364 

.381 

.392 

.399 

.413 

.418 

.426 

.433 

.437 

.05 

.369 

.398 

.413 

.428 

.445 

.455 

.461 

.476 

.479 

.485 

.493 

.498 

.10 

.463 

.491 

.504 

.519 

.534 

.543 

.550 

.562 

.566 

.571 

.578 

.582 

.25 

.679 

.702 

.713 

.725 

.737 

.745 

.749 

.760 

.762 

.767 

.772 

.775 

.50 

1.05 

1.07 

1.07 

1.08 

1.08 

1.09 

1.09 

1.10 

1.10 

1.10 

1.10 

1.10 

.75 

1.68 

1.67 

1.67 

1.66 

1.66 

1.66 

1.65 

1.65 

1.65 

1.65 

1.65 

1.65 

.90 

2.63 

2.59 

2.58 

2.56 

2.54 

2.52 

2.51 

2.50 

2.49 

2.48 

2.48 

2.47 

.95 

3.51 

3.44 

3.41 

3.38 

3.34 

3.32 

3.30 

3.27 

3.27 

3.25 

3.24 

3.23 

.975 

4.57 

4.47 

4.42 

4.36 

4.31 

4.28 

4.25 

4.21 

4.20 

4.18 

4.16 

4.14 

.99 

6.31 

6.16 

6.O7 

5.99 

5.91 

5.86 

5.82 

5.75 

5.74 

5.70 

5.67 

5.65 

.995 

7.97 

7.75 

7.65 

7.53 

7.42 

7.35 

7.31 

7.22 

7.19 

7.15 

7.10 

7.08 

.999 

13.3 

12.9 

12.7 

12.5 

12.3 

12.2 

12.1 

11.9 

11.9 

11.8 

11.7 

11.7 

.9995 

16.5 

16.0 

15.7 

15.5 

15.2 

15.1 

15.0 

14.7 

14.7 

14.6 

14.5 

14.4 

8 

.0005 

.136 

.164 

.181 

.198 

.218 

.230 

.239 

.257 

.262 

.271 

.281 

.287 

.001 

.155 

.184 

.200 

.218 

.238 

.250 

.259 

.277 

.282 

.292 

.300 

.306 

.005 

.214 

.244 

.261 

.279 

.299 

.311 

.319 

.337 

.341 

.351 

.358 

.364 

,01 

.250 

.281 

.297 

.315 

.334 

.346 

.354 

.372 

.376 

.385 

.392 

.398 

.025 

.313 

.343 

.360 

.377 

.395 

.407 

.415 

.431 

.435 

.442 

.450 

.456 

.05 

.379 

.409 

.425 

.441 

.459 

.469 

.477 

.493 

.496 

.505 

.510 

.516 

.10 

.472 

.500 

.515 

.531 

.547 

.556 

.563 

.578 

.581 

,588 

.595 

.599 

.25 

.684 

.707 

.718 

.730 

.743 

.751 

.756 

.767 

.769 

.775 

.780 

.783 

.50 

1.04 

1.05 

1.06 

1.07 

1.07 

1.07 

1.08 

1.08 

1.08 

1.09 

1.O9 

1.09 

.75 

1.62 

1.61 

1.60 

1.60 

1.59 

1.59 

1.59 

1.58 

1.58 

1.58 

1.58 

1.58 

.90 

2.46 

2.42 

2.40 

2.38 

2.36 

2.35 

2.34 

2.32 

2.32 

2.31 

2.30 

2.29 

.95 

3.22 

3.15 

3.12 

3.08 

3.04 

3.02 

3.01 

2.97 

2.97 

2.95 

2.94 

2.93 

.975 

4.10 

4.00 

3.95 

3.89 

3.84 

3.81 

3.78 

3.74 

3.73 

3.70 

3.68 

3.67 

.99 

5.52 

5.36 

5.28 

5.20 

5.12 

5.07 

5.03 

4.96 

4.95 

4.91 

4.88 

4.86 

.995 

6.81 

6.61 

6.50 

6.40 

6.29 

6.22 

6.18 

6.09 

6.06 

6.02 

5.98 

5.95 

.999 

10.8 

10.5 

10.3 

10.1 

9.92 

9.80 

9.73 

9.57 

9.54 

9.46 

9.39 

9.34 

.9995 

13.1 

12.7 

12.5 

12.2 

12.0 

11.8 

11.8 

11.6 

11.5 

11.4 

11.4 

11.3 

9 

.0005 

.141 

.171 

.188 

.207 

.228 

.242 

.251 

.270 

.276 

.287 

.297 

.303 

.001 

.160 

.191 

.208 

.228 

.249 

.262 

.271 

.291 

.296 

.307 

.316 

.323 

.005 

.220 

.253 

.271 

.290 

.310 

.324 

.332 

.351 

.356 

.366 

.376 

.382 

.01 

.257 

.289 

.307 

.326 

.346 

.358 

.368 

.386 

.391 

.400 

.410 

.415 

.025 

.320 

.352 

.370 

.388 

.408 

.420 

.428 

.446 

.450 

.459 

.467 

.473 

.05 

.386 

.418 

.435 

.452 

.471 

.483 

.490 

.508 

.510 

.518 

.526 

.532 

,10 

.479 

.509 

.525 

.541 

.558 

.568 

.575 

.588 

.594 

.602 

.610 

.613 

.25 

.687 

.711 

.723 

.736 

.749 

.757 

.762 

.773 

.776 

.782 

.787 

.791 

.50 

1.03 

1.04 

1.05 

1.05 

1.06 

1.06 

1.07 

1.07 

1.07 

1.08 

1.08 

1.08 

.75 

1.57 

1.56 

1.56 

1.55 

1.55 

1.54 

1.54 

1.53 

1.53 

1.53 

1.53 

1.53 

.90 

2.34 

2.30 

2.28 

2.25 

2.23 

2.22 

2.21 

2.19 

2.18 

2.17 

2.17 

2.16 

.95 

3.01 

2.94 

2.90 

2.86 

2.83 

2.80 

2.79 

2.76 

2.75 

2.73 

2.72 

2.71 

.975 

3.77 

3.67 

3.61 

3.56 

3.51 

3.47 

3.45 

3.40 

3.39 

3.37 

3.35 

3.33 

.99 

4.96 

4.81 

4.73 

4.65 

4.57 

4.52 

4.48 

4.42 

4.40 

4.36 

4.33 

4.31 

.995 

6.03 

5.83 

5.73 

5.62 

5.52 

5.45 

5.41 

5.32 

5.30 

5.26 

5.21 

5.19 

.999 

9.24 

8.90 

8.72 

8.55 

8.37 

8,26 

8.19 

8.04 

8.00 

7.93 

7.86 

7.81 

.9995 

11.0 

10.6 

10.4 

10.2 

9.94 

9.80 

9.71 

9.53 

9.49 

9.40 

9.32 

9.26 

[535] 


vz 

1 

2 

3 

4 

5 

6 

7 

8 

9 

10 

11 

12 

10 

,0005 

.0*41 

.0350 

.0249 

.015 

.028 

.043 

.057 

.071 

.085 

.097 

.108 

.119 

.001 

.0617 

.0*10 

.0277 

.021 

.037 

.054 

.071 

.087 

.101 

.114 

.126 

.137 

.005 

.0*41 

.0250 

.023 

.048 

.073 

.098 

.119 

.139 

.156 

.171 

.185 

.197 

.01 

.On7 

.010 

.037 

.069 

.  100 

.127 

.151 

.172 

.190 

.206 

.220 

.233 

.025 

.0210 

.025 

.069 

.113 

.151 

.183 

.210 

.233 

.252 

.269 

.283 

.296 

.05 

.0241 

.052 

.  114 

.168 

.211 

.246 

.275 

.299 

.319 

.336 

.351 

.363 

.10 

.017 

.  106 

.191 

.255 

.303 

.340 

.370 

.394 

.414 

.430 

.444 

.457 

.25 

.107 

.296 

.409 

.480 

.529 

.565 

.592 

.613 

.631 

.645 

.657 

.667 

.50 

.490 

.743 

.845 

.899 

.932 

.954 

.971 

.983 

.992 

1.00 

1.01 

1.01 

.75 

1.49 

1.60 

1.60 

1.59 

1.59 

1.58 

1.57 

1.56 

1.56 

1.55 

1.55 

1.54 

.90 

3.28 

2.92 

2.73 

2.61 

2.52 

2,46 

2.41 

2.38 

2.35 

2.32 

2.30 

2.28 

.95 

4.96 

4.10 

3.71 

3.48 

3.33 

3.22 

3.14 

3.07 

3.02 

2.98 

2,94 

2.91 

.975 

6,94 

5.46 

4.83 

4.47 

4.24 

4.07 

3.95 

3.85 

3.78 

3.72 

3.66 

3.62 

.99 

10.0 

7.56 

6.55 

5.99 

5.64 

5.39 

5.20 

5.06 

4.94 

4.85 

4.77 

4.71 

.995 

12.8 

9.43 

8.08 

7.34 

6.87 

6.54 

6.30 

6.12 

5.97 

5.85 

5.75 

5.66 

.999 

21.0 

14.9 

12.6 

11.3 

10.5 

9.92 

9.52 

9.20 

8.96 

8.75 

8.58 

8.44 

.9995 

25.5 

17.9 

15.0 

13.4 

12.4 

11.8 

11.3 

10.9 

10.6 

10.3 

10.1 

9.93 

11 

,0005 

.0641 

.0350 

.0249 

.015 

.028 

.043 

.058 

.072 

.086 

.099 

.111 

.121 

.001 

.0<516 

.0210 

.0278 

.021 

.038 

.055 

.072 

.088 

.103 

.116 

.129 

.140 

.005 

,0440 

.0250 

.023 

.048 

.074 

.099 

.121 

.141 

.158 

.174 

.188 

.200 

.01 

.0316 

.010 

.037 

.069 

.100 

.128 

.153 

.175 

.  193 

.210 

.224 

.237 

.025 

.0210 

.025 

.069 

.114 

.152 

.185 

.212 

.236 

.256 

.273 

.288 

.301 

.05 

.0241 

.052 

.114 

.168 

.212 

.248 

.278 

.302 

.323 

.340 

.355 

.368 

.10 

.017 

.106 

.192 

.256 

.305 

.342 

.373 

.397 

.417 

.435 

.448 

.461 

.25 

.107 

.295 

.408 

.481 

.529 

.565 

.592 

.614 

.633 

.645 

.658 

.667 

.50 

,486 

.739 

.840 

.893 

.926 

.948 

.964 

.977 

.986 

.994 

1.00 

1.01 

.75 

1.47 

1,58 

1.58 

1.57 

1.56 

1,55 

1.54 

1.53 

1.53 

1.52 

1.52 

1.51 

.90 

3.23 

2.86 

2.66 

2.54 

2.45 

2.39 

2.34 

2.30 

2.27 

2.25 

2.23 

2.21 

.95 

4.84 

3.98 

3.59 

3.36 

3.20 

3.09 

3.01 

2.95 

2.90 

2.85 

2,82 

2.79 

.975 

6.72 

5.26 

4,63 

4.28 

4.04 

3.88 

3.76 

3.66 

3.59 

3.53 

3.47 

3,43 

.99 

9.65 

7.21 

6.22 

5.67 

5.32 

5.07 

4.89 

4.74 

4.63 

4.54 

4.46 

4.40 

.995 

12.2 

8.91 

7.60 

6.88 

6.42 

6.10 

5.86 

5.68 

5.54 

5.42 

5.32 

5.24 

.999 

19.7 

13.8 

11.6 

10.3 

9.58 

9.05 

8.66 

8.35 

8.12 

7.92 

7.76 

7.62 

.9995 

23.6 

16.4 

13.6 

12.2 

11.2 

10.6 

10.1 

9.76 

9.48 

9.24 

9.04 

8.88 

12 

.0005 

.0^41 

.0*50 

.0249 

.015 

.028 

.044 

.058 

.073 

.087 

.101 

.113 

.124 

.001 

.0516 

.0*10 

.0278 

.021 

.038 

.056 

.073 

.089 

.104 

,118 

.131 

.143 

.005 

.0*39 

.0250 

.023 

.048 

.075 

.100 

.122 

.143 

.161 

.177 

.191 

.204 

.01 

.0316 

.010 

.037 

.070 

.101 

.130 

.  155 

.176 

.196 

.212 

.227 

.241 

.025 

.0^10 

.025 

.070 

.114 

,153 

.186 

.214 

.238 

.259 

.276 

.292 

.305 

.05 

.0241 

.052 

.114 

.169 

.214 

.250 

.280 

.305 

.325 

.343 

.358 

.372 

.10 

.016 

.106 

.192 

.257 

.306 

.344 

.375 

.400 

.420 

.438 

.452 

.466 

.25 

.106 

.295 

.408 

.480 

.530 

.566 

.594 

.616 

.633 

.649 

.662 

.671 

.50 

.484 

.735 

.835 

.888 

.921 

.943 

.959 

.972 

.981 

.989 

.995 

1.00 

.75 

1.46 

1.56 

1.56 

1.55 

1.54 

1.53 

1.52 

1.51 

1.51 

1.50 

1.50 

1.49 

.90 

3.18 

2.81 

2.61 

2.48 

2.39 

2.33 

2.28 

2.24 

2.21 

2.19 

2.17 

2.15 

.95 

4.75 

3.89 

3.49 

3.26 

3.11 

3.00 

2.91 

2.85 

2.80 

2.75 

2.72 

2.69 

.975 

6.55 

5.10 

4.47 

4.12 

3.89 

3.73 

3.61 

3.51 

3.44 

3.37 

3.32 

3.28 

.99 

9.33 

6.93 

5.95 

5.41 

5.06 

4.82 

4.64 

4.50 

4.39 

4.30 

4.22 

4.16 

.995 

11.8 

8.51 

7.23 

6.52 

6.07 

5.76 

5.52 

5.35 

5.20 

5.09 

4.99 

4.91 

.999 

18.6 

13-0 

10.8 

9.63 

8.89 

8.38 

8.00 

7.71 

7.48 

7.29 

7.14 

7.01 

.9995 

22.2 

15.3 

12.7 

11.2 

10.4 

9.74 

9,28 

8.94 

8.66 

8.43 

8.24 

8.08 

[5361 


, 

P    ^\ 

15 

20 

24 

30 

40 

50 

60 

100 

120 

20O 

500 

oo 

10 

.0005 
.001 
.005 

.145 
.164 
.226 

.  177 
.197 
.260 

.195 
.216 
.279 

.215 
.236 
.299 

.238 
.258 

.321 

.251 

.272 
.334 

.262 
.282 

.344 

.282 
.303 
.365 

.288 
.309 
.370 

.299 

.321 
.380 

.311 
.331 

.391 

.319 
.338 
.397 

.01 

.263 

.297 

.316 

.336 

.357 

.370 

.380 

.400 

.405 

.415 

.424 

.431 

.025 

.327 

.360 

.379 

.398 

.419 

.431 

.441 

.459 

.464 

.474 

.483 

.488 

.05 

.393 

.426 

.444 

.462 

.481 

.493 

.502 

.518 

.523 

.532 

,541 

,546 

.10 

.486 

.516 

.532 

.549 

.567 

.578 

.586 

.602 

.605 

.614 

.621 

,625 

.25 

.691 

.714 

.727 

.740 

.754 

.762 

.767 

.779 

.782 

.788 

.793 

.797 

.50 

1.02 

1.03 

1.04 

1.05 

1.05 

1.06 

1.06 

1.06 

1.06 

1.07 

1.07 

1.07 

.75 

1.53 

1.52 

1.52 

1.51 

1.51 

1.50 

1.50 

1.49 

1.49 

1.49 

1.48 

1.48 

.90 

2.24 

2.20 

2.18 

2.16 

2.13 

2.12 

2.11 

2.09 

2.08 

2.07 

2.06 

2.06 

.95 

2.85 

2.77 

2.74 

2.70 

2.66 

2.64 

2.62 

2.59 

2.58 

2.56 

2.55 

2.54 

.975 

3.52 

3.42 

3.37 

3.31 

3.26 

3.22 

3.20 

3,15 

3.14 

3.12 

3.09 

3.08 

.99 

4.56 

4.41 

4.33 

4.25 

4.17 

4.12 

4.08 

4.01 

4.00 

3.96 

3,93 

3.91 

.995 

5.47 

5.27 

5.17 

5.07 

4.97 

4.90 

4.86 

4.77 

4.75 

4.71 

4.67 

4.64 

.999 

8.13 

7.80 

7.64 

7.47 

7.30 

7.19 

7.12 

6.98 

6.94 

6.87 

6.81 

6.76 

.9995 

9.56 

9.16 

8.96 

8.75 

8.54 

8,42 

8.33 

8.16 

8.12 

8.04 

7.96 

7.90 

11 

.0005 

.148 

.182 

.201 

.222 

.246 

.261 

.271 

.293 

.299 

.312 

.324 

.331 

-OO1 

.168 

.202 

.222 

.243 

.266 

.282 

.292 

.313 

.320 

.332 

.343 

.353 

.005 

.231 

.266 

.286 

.308 

.330 

.345 

.355 

.376 

.382 

.394 

.403 

.412 

.01 

.268 

.304 

.324 

.344 

366 

.380 

.391 

.412 

,417 

.427 

.439 

.444 

.025 

.332 

.368 

.386 

.407 

.429 

.442 

.450 

.472 

.476 

,485 

.495 

,503 

.05 

.398 

.433 

.452 

.469 

.490 

.503 

.513 

.529 

.535 

.543 

.552 

.559 

.10 

.490 

.524 

.541 

.559 

.578 

.588 

.595 

.614 

.617 

.625 

.633 

.637 

.25 

.694 

.719 

.730 

.744 

.758 

.767 

.773 

.780 

.788 

.794 

.799 

.803 

.50 

1.02 

1.03 

1.03 

1.04 

1.05 

1.05 

1.05 

1.06 

1.06 

1.06 

1.06 

1.06 

.75 

1.50 

1.49 

1.49 

1.48 

1.47 

1.47 

1.47 

1.46 

1.46 

1.46 

1.45 

1.45 

.90 

2.17 

2.12 

2.10 

2.08 

2.05 

2.04 

2,03 

2.00 

2.00 

1.99 

1.98 

1.97 

.95 

2.72 

2.65 

2.61 

2.57 

2.53 

2.51 

2.49 

2.46 

2.45 

2.43 

2.42 

2.40 

.975 

3.33 

3.23 

3.17 

3.12 

3.06 

3.03 

3,00 

2.96 

2.94 

2,92 

2.90 

2.88 

.99 

4.25 

4.10 

4.02 

3.94 

3.86 

3.81 

3,78 

3.71 

3.69 

3.66 

3.62 

3.60 

.995 

5.05 

4.86 

4.76 

4.65 

4.55 

4.49 

4.45 

4.36 

4.34 

4.29 

4.25 

4.23 

.999 

7.32 

7.01 

6,85 

6.68 

6.52 

6.41 

6.35 

6.21 

6.17 

6.10 

6.04 

6.00 

.9995 

8.52 

8.14 

7.94 

7.75 

7.55 

7.43 

7.35 

7.18 

7.14 

7.06 

6.98 

6.93 

12 

.0005 

.152 

.186 

.206 

.228 

.253 

.269 

.280 

.305 

,311 

.323 

.337 

.345 

.001 

.172 

.207 

.228 

.250 

,275 

.291 

.302 

.326 

.332 

.344 

.357 

.365 

.005 

.235 

.272 

.292 

.315 

.339 

.355 

.365 

.388 

.393 

.405 

.417 

.424 

.01 

.273 

,310 

.330 

.352 

.375 

.391 

.401 

.422 

.428 

.441 

.450 

.458 

.025 

.337 

.374 

.394 

.416 

.437 

.450 

.461 

.481 

.487 

.498 

.508 

.514 

.05 

.404 

.439 

.458 

.478 

.499 

.513 

.522 

.541 

.545 

.556 

.565 

.571 

.10 

.496 

.528 

.546 

.564 

.583 

.595 

.604 

.621 

.625 

.633 

.641 

.647 

.25 

.695 

.721 

.734 

.748 

.762 

.771 

.777 

.789 

,792 

.799 

.804 

.808 

.50 

1.01 

1.02 

1.03 

1.03 

1.04 

1.04 

1.05 

1.05 

1.05 

1.05 

1.06 

1.06 

.75 

1.48 

1.47 

1.46 

1.45 

1.45 

1.44 

1.44 

1.43 

1.43 

1.43 

1.42 

1.42 

.90 

2.11 

2.06 

2.04 

2.01 

1.99 

1.97 

1,96 

1.94 

1.93 

1.92 

1.91 

1.90 

.95 

2.62 

2.54 

2.51 

2.47 

2.43 

2.40 

2.38 

2.35 

2.34 

2.32 

2.31 

2.30 

.975 

3.18 

3.07 

3,02 

2.96 

2.91 

2.87 

2.85 

2.80 

2.79 

2.76 

2.74 

2.72 

.99 

4.01 

3.86 

3.78 

3.70 

3.62 

3.57 

3.54 

3.47 

3.45 

3.41 

3.38 

3.36 

.995 

4.72 

4.53 

4.43 

4.33 

4.23 

4.17 

4.12 

4.04 

4.01 

3.97 

3.93 

3.90 

.999 

6.71 

6.40 

6.25 

6.09 

5.93 

5.83 

5.76 

5.63 

5.59 

5.52 

5.46 

5,42 

.9995 

7.74 

7.37 

7.18 

7.00 

6,80 

6.68 

6.61 

6.45 

6,41 

6.33 

6.25 

6.20 

£5371 


1 

Vl 

1 

2 

3 

4 

5 

6 

7 

8 

9 

10 

11 

12 

15 

.0005 

.0341 

.0350 

.0249 

.015 

.029 

.045 

.061 

.076 

.091 

.105 

.117 

.129 

.001 

.0516 

.0310 

.0279 

.021 

.039 

.057 

.075 

.092 

.108 

.123 

.137 

.149 

,005 

.0*39 

.0250 

.023 

.049 

.076 

.102 

.125 

.147 

.166 

.  183 

.198 

.212 

,01 

,0316 

.010 

.037 

.070 

.103 

.132 

.158 

.181 

.202 

.219 

.235 

.249 

.025 

.0210 

.025 

.070 

.116 

.156 

.190 

.219 

.244 

.265 

.284 

.300 

.315 

.05 

.0241 

.051 

.115 

.170 

.216 

.254 

.285 

.311 

.333 

.351 

.368 

.382 

.10 

.016 

.106 

.192 

.258 

.309 

.348 

.380 

.406 

.427 

.446 

.461 

.475 

,25 

.105 

.293 

.407 

.480 

.531 

.568 

.596 

.618 

.637 

.652 

.667 

.676 

.50 

,478 

.726 

.826 

.878 

.911 

.933 

.948 

.960 

.970 

.977 

.984 

.989 

.75 

1.43 

1.52 

1.52 

1.51 

1.49 

1.48 

1.47 

1.46 

1.46 

1.45 

1.44 

1.44 

.90 

3.07 

2.70 

2.49 

2.36 

2.27 

2.21 

2.16 

2.12 

2.09 

2.06 

2.04 

2.02 

.95 

4.54 

3.68 

3.29 

3.06 

2.90 

2.79 

2.71 

2.64 

2.59 

2.54 

2.51 

2.48 

.975 

6.20 

4.76 

4.15 

3.80 

3.58 

3.41 

3.29 

3.20 

3.12 

3.06 

3.01 

2.96 

.99 

8.68 

6.36 

5.42 

4.89 

4.56 

4.32 

4.  14 

4.00 

3.89 

3.80 

3.73 

3.67 

.995 

10.8 

7.70 

6.48 

5.80 

5.37 

5.07 

4.85 

4.67 

4.54 

4.42 

4.33 

4.25 

.999 

16.6 

11  .3 

9.34 

8.25 

7.57 

7.09 

6.74 

6.47 

6.26 

6.08 

5.93 

5.81 

.9995 

19.5 

13.2 

10.8 

9.48 

8.66 

8.10 

7.68 

7.36 

7.11 

6.91 

6.75 

6.60 

20 

.0005 

.0°40 

.0350 

,0250 

,015 

.029 

.046 

.063 

.079 

.094 

.109 

.123 

.136 

.001 

.0516 

.0210 

.0279 

.022 

.039 

.058 

.077 

.095 

.112 

.128 

.143 

.156 

.005 

.0139 

.0250 

.023 

.050 

.077 

.104 

.129 

.151 

.171 

.190 

.206 

.221 

.01 

.0316 

.010 

.037 

.071 

.105 

.135 

.  162 

.187 

.208 

.227 

.244 

.259 

.025 

.0210 

.025 

.071 

.117 

.  158 

.193 

.224 

.250 

.273 

.292 

.310 

.325 

.05 

.0240 

.051 

.115 

.172 

.219 

.258 

.290 

.318 

.340 

.360 

.377 

.393 

.  10 

.016 

.  106 

,193 

.260 

.312 

.353 

.385 

.412 

.435 

.454 

.472 

.485 

.25 

.104 

.292 

.407 

.480 

.531 

.569 

.598 

.622 

.641 

.656 

.671 

.681 

,50 

.472 

.718 

.816 

.868 

,900 

,922 

.938 

.950 

.959 

.966 

.972 

.977 

.75 

1.40 

1.49 

1.48 

1.47 

1.45 

1.44 

1.43 

1.42 

1.41 

1.40 

1.39 

1,39 

.90 

2,97 

2.59 

2.38 

2.25 

2.16 

2.09 

2.04 

2.00 

1.96 

1.94 

1.91 

1,89 

.95 

4.35 

3.49 

3.10 

2.87 

2.71 

2.60 

2.51 

2.45 

2.39 

2.35 

2.31 

2,28 

.975 

5.87 

4.46 

3.86 

3.51 

3.29 

3.13 

3.01 

2.91 

2.84 

2.77 

2.72 

2.68 

.99 

8.10 

5.85 

4.94 

4.43 

4.10 

3.87 

3.70 

3.56 

3.46 

3.37 

3.29 

3.23 

.995 

9.94 

6.99 

5.82 

5.17 

4.76 

4.47 

4.26 

4.09 

3.96 

3.85 

3.76 

3.68 

.999 

14.8 

9,95 

8.10 

7.10 

6.46 

6.02 

5.69 

5.44 

5.24 

5.08 

4.94 

4.82 

.9995 

17.2 

11.4 

9.20 

8.02 

7.28 

6.76 

6.38 

6.08 

5.85 

5.66 

5.51 

5.38 

24 

.0005 

.0°40 

.0^50 

.0250 

.015 

.030 

.046 

.064 

.080 

.096 

.112 

.126 

,139 

,001 

.OM6 

.0210 

.0279 

.022 

.040 

.059 

.079 

.097 

.115 

.131 

.146 

.160 

.005 

.0<40 

.0250 

,023 

.050 

.078 

.106 

.  131 

.154 

.175 

,193 

.210 

.226 

.01 

.0^16 

.010 

.038 

.072 

.106 

.137 

,165 

.189 

.211 

.231 

.249 

.264 

.025 

.0210 

.025 

.071 

,117 

.159 

.195 

.227 

.253 

.277 

,297 

.315 

.331 

.05 

.0240 

.051 

,116 

.173 

.221 

.260 

.293 

.321 

.345 

.365 

.383 

.399 

.10 

.016 

.106 

.193 

.261 

.313 

.355 

.388 

.416 

.439 

.459 

.476 

.491 

.25 

.104 

.291 

.406 

.480 

.532 

.570 

.600 

.623 

.643 

,659 

.671 

.684 

,50 

.469 

.714 

.812 

.863 

.895 

.917 

.932 

.944 

.953 

.961 

.967 

.972 

.75 

1.39 

1.47 

1.46 

1.44 

1.43 

1.41 

1.40 

1.39 

1,38 

1.38 

1.37 

1.36 

.90 

2.93 

2.54 

2.33 

2.19 

2.10 

2,04 

1,98 

1.94 

1.91 

1.88 

1.85 

1.83 

.95 

4.26 

3,40 

3-01 

2.78 

2.62 

2.51 

2.42 

2.36 

2.30 

2.25 

2.21 

2.18 

.975 

5.72 

4.32 

3.72 

3.38 

3.15 

2.99 

2.87 

2.78 

2.70 

2.64 

2.59 

2.54 

.99 

7.82 

5.61 

4.72 

4.22 

3.90 

3.67 

3.50 

3.36 

3.26 

3,17 

3.09 

3.03 

.995 

9.55 

6.66 

5.52 

4.89 

4.49 

4.20 

3.99 

3.83 

3,69 

3.59 

3.50 

3.42 

.999 

14.0 

9.34 

7.55 

6.59 

5.98 

5.55 

5.23 

4.99 

4,80 

4.64 

4.50 

4.39 

.  9995 

16.2 

10.6 

8.52 

7.39 

6.68 

6.18 

5.82 

5.54 

5.31 

5.13 

4.98 

4.85 

[5383 


»* 

x 

15 

20 

24 

30 

40 

50 

60 

100 

120 

200 

500 

OO 

15 

.0005 

.159 

.197 

.220 

.244 

.272 

.290 

.303 

.330 

.339 

.353 

.368 

.377 

.001 

.181 

.219 

.242 

.266 

.294 

.313 

.325 

.352 

.360 

.375 

.388 

.398 

.005 

.246 

.286 

.308 

-333 

.360 

.377 

.389 

.415 

.422 

.435 

.448 

.457 

.01 

.284 

.324 

.346 

.370 

.397 

.413 

.425 

.450 

.456 

.469 

.483 

.490 

.025 

.349 

.389 

.410 

.433 

.458 

.474 

.485 

.508 

.514 

.526 

.538 

-546 

.05 

,416 

.454 

.474 

.496 

.519 

.535 

.545 

.565 

.571 

.581 

.592 

.600 

.10 

.507 

.542 

.561 

.581 

.602 

.614 

.624 

.641 

.647 

.658 

.667 

.672 

.25 

.701 

.728 

.742 

.757 

.772 

.782 

.788 

.802 

.805 

-812 

.818 

-822 

.50 

1.00 

1.01 

1.02 

1.02 

1,03 

1.03 

1.03 

1.04 

1.04 

1-04 

1.04 

1.05 

.75 

1.43 

1.41 

1.41 

1.40 

1.39 

1.39 

1.38 

1.38 

1.37 

1.37 

1.36 

1.36 

.90 

1.97 

1.92 

1.90 

1.87 

1.85 

1.83 

1.82 

1.79 

1.79 

1.77 

1.76 

1.76 

.95 

2,40 

2.33 

2.39 

2.25 

2.20 

2.18 

2.16 

2.12 

2.11 

2.10 

2,08 

2.07 

,975 

2.86 

2.76 

2,70 

2.64 

2.59 

2.55 

2.52 

2.47 

2.46 

2.44 

2.41 

2.40 

.99 

3.52 

3.37 

3.29 

3.21 

3.13 

3.08 

3.05 

2.98 

2.96 

2,92 

2.89 

2.  87 

.995 

4.07 

3.88 

3.79 

3.69 

3.59 

3.52 

3.48 

3.39 

3.37 

3.33 

3.29 

3.26 

,999 

5.54 

5.25 

5.10 

4.95 

4.80 

4.70 

4.64 

4.51 

4.47 

4.41 

4.35 

4.31 

.9995 

6.27 

5.93 

5.75 

5.58 

5.40 

5.29 

5.21 

5.06 

5.02 

4.94 

4.87 

4.83 

20 

.0005 

.169 

.211 

.235 

.263 

.295 

.316 

.331 

.364 

.375 

.391 

.408 

.422 

.001 

.191 

.233 

.258 

.286 

.318 

.339 

.354 

.386 

.395 

-413 

.429 

.441 

.005 

.258 

.301 

.327 

-354 

.385 

.405 

.419 

.448 

.457 

.474 

.490 

.500 

,01 

.297 

.340 

.365 

.392 

.422 

.441 

.455 

.483 

.491 

.508 

.521 

.532 

.025 

.363 

.406 

.430 

.456 

.484 

,503 

.514 

.541 

.548 

.562 

.575 

.585 

.05 

.430 

.471 

.493 

.518 

.544 

.562 

.572 

.595 

.603 

.617 

.629 

.637 

.10 

.520 

.557 

.578 

.600 

.623 

.637 

.648 

.671 

.675 

.685 

.694 

.704 

.25 

.708 

.736 

.751 

.767 

.784 

.794 

.801 

,816 

.820 

-827 

.835 

.840 

.50 

.989 

1.00 

1.01 

1.01 

1.02 

1.02 

1.02 

1.03 

1.03 

1.03 

1.03 

1.03 

.75 

1.37 

1.36 

1.35 

1.34 

1.33 

1.33 

1,32 

1.31 

1.31 

1.30 

1.30 

1.29 

,90 

1-84 

1.79 

1.77 

1.74 

1.71 

1.69 

1.68 

1.65 

1.64 

1.63 

1.62 

1.61 

.95 

2.20 

2.12 

2.08 

2.04 

1.99 

1.97 

1.95 

1.91 

1.90 

1.88 

1.86 

1.84 

.975 

2.57 

2.46 

2.41 

2.35 

2.29 

2.25 

2.22 

2.17 

2.16 

2.13 

2.10 

2.09 

.99 

3.09 

2.94 

2.86 

2.78 

2.69 

2.64 

2.61 

2.54 

2,52 

2.48 

2.44 

2.42 

.995 

3.50 

3.32 

3.22 

3.12 

3.02 

2.96 

2.92 

2.83 

2.81 

2.76 

2.72 

2.69 

.999 

4.56 

4.29 

4.  15 

4.01 

3.86 

3.77 

3.70 

3.58 

3.54 

3.48 

3.42 

3.38 

.9995 

5.07 

4.75 

4.58 

4.42 

4.24 

4.15 

4.07 

3.93 

3-90 

3.82 

3.75 

3,70 

24 

.0005 

.174 

.218 

.244 

.274 

.309 

.331 

.349 

.384 

.395 

.416 

.434 

.449 

.001 

.196 

.241 

.268 

.298 

.332 

.354 

.371 

.405 

.417 

.437 

.455 

.469 

.005 

.264 

.310 

.337 

.367 

.400 

.422 

.437 

.469 

.479 

.498 

.515 

.527 

.01 

.304 

.350 

,376 

.405 

.437 

.459 

.473 

.505 

.513 

.529 

.546 

.558 

.025 

.370 

.415 

.441 

.468 

.498 

.518 

.531 

.562 

.568 

.585 

.599 

.610 

.05 

.437 

.480 

.504 

.530 

.558 

.575 

.588 

.613 

.622 

.637 

.649 

-659 

.10 

.527 

.566 

.588 

.611 

.635 

.651 

.662 

.685 

.691 

.704 

.715 

.723 

.25 

.712 

.741 

.757 

,773 

.791 

.802 

.809 

.825 

.829 

,837 

.844 

.850 

.50 

.983 

.994 

1.00 

1.01 

1.01 

1.02 

1.02 

1.02 

1.02 

1,02 

1.03 

1.03 

.75 

1.35 

1.33 

1.32 

1.31 

1.30 

1.29 

1.29 

1.28 

1.28 

1.27 

1.27 

1.26 

.90 

1.78 

1.73 

1.70 

1.67 

1.64 

1.62 

1.61 

1.58 

1.57 

1.56 

1.54 

1.53 

.95 

2.11 

2.03 

1.98 

1.94 

1.89 

1.86 

1.84 

1.80 

1.79 

1.77 

1,75 

1.73 

.975 

2,44 

2.33 

2.27 

2.21 

2.15 

2.11 

2.08 

2.02 

2.01 

1.98 

1.95 

1.94 

.99 

2.89 

2.74 

2.66 

2.58 

2.49 

2.44 

2.40 

2.33 

2.31 

2.27 

2.24 

2.21 

.995 

3.25 

3.06 

2.97 

2.87 

2.77 

2.70 

2.66 

2.57 

2,55 

2,50 

2.46 

2.43 

,999 

4.14 

3.87 

3.74 

3.59 

3.45 

3.35 

3.29 

3.16 

3.14 

3.07 

3.01 

2.97 

.9995 

4.55 

4.25 

4.09 

3.93 

3.76 

3.66 

3.59 

3.44 

3.41 

3.33 

3.27 

3.22 

[539] 


*2 

1 

2 

3 

4 

5 

6 

7 

8 

9 

10 

11 

12 

30 

.0005 

,0840 

.0250 

.0250 

.015 

,030 

.047 

.065 

.082 

.098 

.114 

.129 

.143 

.001 

,0516 

.0210 

.0280 

.022 

.040 

.060 

.080 

.099 

.117 

.134 

.150 

.164 

.005 

.OHO 

.0250 

.024 

.050 

.079 

.107 

.133 

.156 

.178 

.197 

.215 

.231 

.01 

.0316 

.010 

.038 

.072 

.107 

.138 

.167 

.192 

.215 

.235 

.254 

.270 

.025 

.0210 

.025 

.071 

.118 

.161 

.197 

.229 

.257 

.281 

.302 

.321 

.337 

.05 

.0240 

.051 

.116 

.174 

.222 

.263 

.296 

.325 

.349 

.370 

.389 

.406 

.10 

.016 

.106 

.193 

.262 

.315 

.357 

.391 

.420 

.443 

,464 

.481 

.497 

.25 

.103 

.290 

.406 

.480 

.532 

.571 

.601 

.625 

.645 

.661 

.676 

,688 

.50 

.466 

.709 

.807 

.858 

.890 

.912 

.927 

.939 

.948 

.955 

.961 

.966 

.75 

1,38 

1.45 

1.44 

1.42 

1.41 

1.39 

1.38 

1.37 

1.36 

1.35 

1.35 

1.34 

.90 

2,88 

2.49 

2.28 

2.14 

2.05 

1.98 

1.93 

1.88 

1.85 

1.82 

1.79 

1.77 

.95 

4.17 

3.32 

2.92 

2.69 

2.53 

2.42 

2.33 

2.27 

2.21 

2.16 

2.13 

2.09 

.975 

5.57 

4.18 

3.59 

3.25 

3.03 

2.87 

2.75 

2.65 

2.57 

2.51 

2.46 

2.41 

.99 

7.56 

5.39 

4.51 

4.02 

3.70 

3.47 

3.30 

3.17 

3.07 

2.98 

2.91 

2.84 

.995 

9.18 

6.35 

5.24 

4.62 

4.23 

3.95 

3.74 

3.58 

3.45 

3.34 

3.25 

3.18 

.999 

13.3 

8.77 

7.05 

6.12 

5.53 

5.12 

4.82 

4.58 

4.39 

4.24 

4.11 

4.00 

.9995 

15.2 

9.90 

7.90 

6.82 

6.14 

5.66 

5.31 

5,04 

4.82 

4.65 

4.51 

4.38 

40 

.0005 

,0^40 

.0350 

.0250 

,016 

.030 

.048 

.066 

.084 

.100 

.117 

.132 

.147 

.001 

.OM6 

.0210 

.0280 

.022 

.042 

.061 

.081 

.101 

.119 

.137 

.153 

.169 

.005 

.0^40 

.0250 

.024 

.051 

.080 

.108 

.135 

.159 

.181 

.201 

.220 

.237 

.01 

.On6 

.010 

.038 

.073 

.108 

.140 

.169 

.195 

.219 

.240 

.259 

.276 

.025 

.0399 

.025 

.071 

.119 

.162 

.199 

.232 

.260 

.285 

.307 

.327 

.344 

,05 

.0^40 

.051 

.116 

,175 

.224 

.265 

.299 

.329 

.354 

.376 

.395 

.412 

.10 

.016 

.106 

.194 

,263 

.317 

.360 

.394 

,424 

.448 

.469 

.488 

.504 

.25 

.103 

-290 

.405 

.480 

.533 

.572 

,603 

.627 

.647 

.664 

.680 

.691 

.50 

.463 

.705 

.802 

.854 

,885 

,907 

.922 

.934 

.943 

.950 

.956 

.961 

.75 

1.36 

1.44 

1.42 

1.40 

1.39 

1.37 

1.36 

1.35 

1.34 

1.33 

1.32 

1.31 

.90 

2.84 

2.44 

2.23 

2.09 

2.00 

1.93 

1.87 

1.83 

1.79 

1.76 

1.73 

1.71 

.95 

4.08 

3.23 

2,84 

2.61 

2.45 

2.34 

2.25 

2.18 

2.12 

2.08 

2.04 

2.00 

.975 

5.42 

4.05 

3.46 

3.13 

2.90 

2.74 

2.62 

2.53 

2.45 

2.39 

2.33 

2.29 

.99 

7.31 

5.18 

4.31 

3.83 

3.51 

3.29 

3.12 

2.99 

2.89 

2.80 

2.73 

2.66 

.995 

8.83 

6,07 

4.98 

4,37 

3.99 

3.71 

3.51 

3.35 

3.22 

3.12 

3.03 

2.95 

.999 

12.6 

8,25 

6.60 

5.70 

5.13 

4.73 

4.44 

4.21 

4.02 

3.87 

3.75 

3.64 

.9995 

14.4 

9.25 

7.33 

6.30 

5.64 

5.19 

4.85 

4.59 

4.38 

4.21 

4.07 

3.95 

60 

.0005 

.0640 

.0350 

,0251 

.016 

.031 

.048 

.067 

.085 

.103 

.120 

.136 

.152 

.001 

.0816 

.0^10 

.0280 

.022 

.041 

.062 

.083 

.103 

.122 

.140 

.157 

.174 

.005 

.OHO 

.0250 

.024 

.051 

.081 

.110 

.137 

.162 

.185 

.206 

.225 

.243 

.01 

.0^16 

.010 

.038 

.073 

.109 

.142 

.172 

.199 

.223 

.245 

.265 

.283 

.025 

,0399 

.025 

,071 

.120 

.163 

.202 

.235 

.264 

.290 

.313 

.333 

.351 

.05 

-0240 

.051 

.116 

.176 

.226 

,267 

.303 

.333 

.359 

.382 

.402 

.419 

.10 

.016 

.106 

.194 

.264 

.318 

.362 

.398 

.428 

.453 

.475 

.493 

.510 

.25 

.102 

.289 

.405 

.480 

.534 

.573 

.604 

.629 

.650 

.667 

.680 

.695 

.50 

.461 

.701 

.798 

.849 

.880 

.901 

.917 

.928 

.937 

.945 

.951 

.956 

.75 

1.35 

1.42 

1.41 

1.38 

1.37 

1.35 

1,33 

1.32 

1.31 

1.30 

1.29 

1.29 

.90 

2.79 

2.39 

2.18 

2.04 

1.95 

1.87 

1.82 

1.77 

1.74 

1.71 

1.68 

1.66 

.95 

4.00 

3.15 

2.76 

2.53 

2.37 

2.25 

2.  17 

2.10 

2.04 

1.99 

1.95 

1.92 

.975 

5.29 

3.93 

3.34 

3.01 

2.79 

2.63 

2.51 

2.41 

2.33 

2.27 

2.22 

2.17 

.99 

7,08 

4.98 

4.13 

3.65 

3.34 

3.12 

2.95 

2.82 

2.72 

2,63 

2.56 

2.50 

.995 

8,49 

5.80 

4.73 

4.  14 

3.76 

3.49 

3.29 

3.13 

3.01 

2.90 

2.82 

2.74 

.999 

12.0 

7.76 

6.17 

5.31 

4.76 

4.37 

4.09 

3.87 

3.69 

3.54 

3.43 

3.31 

.9995 

13.6 

8.65 

6.81 

5.82 

5.20 

4.76 

4.44 

4.18 

3.98 

3.82 

3.69 

3.57 

[5401 


"2 

>< 

15 

20 

24 

30 

40 

50 

60 

100 

120 

2OO 

500 

CO 

30 

.0005 

,179 

,226 

,254 

.287 

.325 

.350 

.369 

.410 

.420 

.444 

.467 

.483 

.001 

.202 

.250 

.278 

.311 

.348 

.373 

.391 

.451 

.442 

.465 

.488 

.503 

.005 

.271 

.320 

,349 

.381 

.416 

,441 

.457 

.495 

.504 

.524 

.543 

.559 

.01 

.311 

.360 

.388 

.419 

.454 

.476 

.493 

.529 

.538 

.559 

.575 

.590 

.025 

.378 

.426 

.453 

.482 

.515 

.535 

,551 

.585 

.592 

.610 

.625 

.639 

.05 

.445 

.490 

.516 

.543 

.573 

.592 

.606 

.637 

.644 

.658 

.676 

.685 

.10 

.534 

,575 

.598 

.623 

.649 

.667 

.678 

.704 

.710 

,725 

.735 

.746 

.25 

.716 

.746 

.763 

.780 

.798 

.810 

.818 

.835 

.839 

.848 

.856 

.862 

.50 

.978 

.989 

.994 

1.00 

1.01 

1.01 

1.01 

1.02 

1.02 

1.02 

1.02 

1.02 

.75 

1,32 

1.30 

1.29 

1.28 

1.27 

1.26 

1.26 

1.25 

1.24 

1.24 

1.23 

1.23 

.90 

1.72 

1.67 

1.64 

1.61 

1.57 

1.55 

1.54 

1.51 

1.50 

1.48 

1.47 

1.46 

.95 

2.01 

1.93 

1.89 

1.84 

1.79 

1.76 

1.74 

1.70 

1.68 

1.66 

1,64 

1.62 

.975 

2.31 

2.20 

2.14 

2.O7 

2.01 

1.97 

1.94 

1.88 

1,87 

1.84 

1,81 

1.79 

.99 

2.70 

2.55 

2.47 

2.39 

2.30 

2.25 

2.21 

2.13 

2.11 

2.07 

2.03 

2.01 

.995 

3.01 

2.82 

2.73 

2.63 

2.52 

2.46 

2.42 

2.32 

2.30 

2.25 

2.21 

2.18 

.999 

3.75 

3.49 

3.36 

3.22 

3.07 

2.98 

2.92 

2.79 

2.76 

2.69 

2.63 

2.59 

.9995 

4.10 

3.80 

3.65 

3.48 

3.32 

3.22 

3.15 

3.00 

2.97 

2.89 

2,82 

2.78 

40 

.0005 

.185 

.236 

.266 

.301 

.343 

.373 

.393 

.441 

.453 

.480 

.504 

,525 

,001 

.209 

.259 

.290 

.326 

.367 

.396 

.415 

.461 

.473 

.500 

.524 

.545 

.005 

.279 

.331 

,362 

.396 

.436 

.463 

.481 

.524 

.534 

.559 

.581 

,599 

.01 

.319 

.371 

.401 

.435 

.473 

.498 

.516 

.556 

.567 

.592 

.613 

.628 

.025 

.387 

.437 

.466 

.498 

.533 

.556 

.573 

.610 

.620 

,641 

.662 

.674 

,05 

.454 

.502 

.529 

.558 

.591 

.613 

.627 

.658 

.669 

.685 

.704 

,717 

.10 

.542 

.585 

.609 

.636 

.664 

.683 

.696 

.724 

.731 

,747 

.762 

.772 

.25 

.720 

.752 

.769 

.787 

,806 

.819 

.828 

.846 

,851 

.861 

.870 

.877 

.50 

.972 

.983 

.989 

.994 

1.00 

1.00 

1,01 

1.01 

1.01 

1.01 

1.02 

1.02 

.75 

1.30 

1.28 

1.26 

1.25 

1.24 

1.23 

1.22 

1.21 

1.21 

1.20 

1.19 

1.19 

.90 

1.66 

1.61 

1.57 

1.54 

1,51 

1.48 

1.47 

1.43 

1.42 

1.41 

1,39 

1.38 

.95 

1.92 

1.84 

1.79 

1.74 

1.69 

1.66 

1.64 

1.59 

1.58 

1.55 

1.53 

1.51 

.975 

2,18 

2.07 

2.01 

1.94 

1.88 

1.83 

1.80 

1.74 

1.72 

1,69 

1,66 

1.64 

.99 

2.52 

2.37 

2.29 

2.20 

2.11 

2.06 

2.02 

1.94 

1.92 

1,87 

1.83 

1.80 

.995 

2.78 

2.60 

2.50 

2,40 

2.30 

2.23 

2.18 

2.09 

2.06 

2,01 

1.96 

1.93 

.999 

3.40 

3,15 

3.01 

2.87 

2.73 

2.64 

2.57 

2.44 

2.41 

2.34 

2,28 

2.23 

.9995 

3.68 

3.39 

3.24 

3,08 

2.92 

2.82 

2.74 

2.60 

2.57 

2.49 

2.41 

2.37 

60 

.0005 

.192 

.246 

.278 

.318 

.365 

.398 

.421 

.478 

.493 

.527 

.561 

.585 

.001 

.216 

.270 

.304 

.343 

.389 

.421 

.444 

.497 

.512 

.545 

.579 

.602 

.005 

.287 

.343 

.376 

.414 

.458 

.488 

.510 

.559 

.572 

.602 

.633 

.652 

.01 

.328 

.383 

.416 

.453 

.495 

,524 

.545 

.592 

.604 

,633 

,658 

.679 

.025 

.396 

.450 

.481 

.515 

.555 

.581 

.600 

.641 

.654 

.680 

.704 

.720 

.05 

.463 

.514 

.543 

.575 

.611 

.633 

.652 

.690 

.700 

.719 

.746 

.759 

.10 

.550 

.596 

.622 

.650 

.682 

.703 

.717 

,750 

.758 

.776 

.793 

.806 

.25 

.725 

.758 

.776 

.796 

.816 

.830 

.840 

.860 

.865 

.877 

.888 

.896 

.50 

.967 

.978 

.983 

.989 

.994 

.998 

1.00 

1.00 

1.01 

1.01 

1.01 

1,01 

.75 

1.27 

1.25 

1.24 

1.22 

1.21 

1.20 

1.19 

1.17 

1.17 

1.16 

1.15 

1.15 

.90 

1,60 

1.54 

1.51 

1.48 

1.44 

1,41 

1.40 

1.36 

1.35 

1.33 

1.31 

1.29 

.95 

1.84 

1.75 

1.70 

1.65 

1.59 

1.56 

1.53 

1.48 

1.47 

1.44 

1.41 

1.39 

.975 

2.06 

1.94 

1.88 

1.82 

1.74 

1.70 

1.67 

1.60 

1.58 

1.54 

1.51 

1.48 

.99 

2.35 

2.20 

2.12 

2.03 

1.94 

1.88 

1.84 

1.75 

1.73 

1.68 

1.63 

1.60 

.995 

2.57 

2.39 

2.29 

2.19 

2.08 

2.01 

1.96 

1,86 

1,83 

1.78 

1.73 

1.69 

.999 

3,08 

2.83 

2.69 

2.56 

2.41 

2.31 

2.25 

2.11 

2.09 

2.01 

1.93 

1  .89 

.9995 

3.30 

3.02 

2.87 

2.71 

2.55 

2.45 

2.38 

2.23 

2.19 

2.11 

2.03 

1.98 

C5413 


*1 

1 

2 

3 

4 

5 

6 

7 

8 

9 

10 

11 

12 

120 

.0005 

.0640 

.0350 

.0^51 

.016 

.031 

.049 

.067 

.087 

.105 

.  123 

.140 

.156 

.001 

.OH6 

.0210 

.0^81 

.023 

.042 

.063 

.084 

.105 

.125 

.144 

.162 

.179 

.005 

.0*39 

.0250 

.024 

.051 

.081 

.111 

.139 

.165 

.189 

.211 

.230 

.249 

.01 

.0*16 

.010 

.038 

.074 

.110 

,143 

.174 

.202 

.227 

.250 

.271 

.290 

.025 

.0399 

.025 

.072 

.120 

.165 

.204 

.238 

.268 

.295 

.318 

.340 

.359 

.05 

.0239 

.051 

.117 

.177 

.227 

.270 

.306 

.337 

.364 

.388 

.408 

.427 

.10 

.016 

.105 

.194 

.265 

.320 

.365 

.401 

.432 

.458 

.480 

.500 

.518 

.25 

.102 

,288 

.405 

.481 

,534 

.574 

.606 

.631 

.652 

.670 

.685 

.699 

.50 

.458 

.697 

.793 

.844 

.875 

.896 

.912 

.923 

.932 

.939 

.945 

.950 

.75 

1.34 

1.40 

1.39 

1.37 

1.35 

1.33 

1.31 

1.30 

1.29 

1.28 

1.27 

1.26 

.90 

2.75 

2.35 

2.13 

3.99 

1.90 

1.82 

1.77 

1.72 

1.68 

1.65 

1.62 

1.60 

.95 

3.92 

3.07 

2.68 

2.45 

2.29 

2.18 

2.09 

2.02 

1.96 

1.91 

1.87 

1.83 

.975 

5.15 

3.80 

3.23 

2.89 

2.67 

2.52 

2.39 

2.30 

2-22 

2.16 

2.10 

2.05 

.99 

6.85 

4.79 

3.95 

3.48 

3.17 

2.96 

2.79 

2.66 

2.56 

2.47 

2.40 

2.34 

.995 

8.  18 

5.54 

4.50 

3.92 

3.55 

3.28 

3.09 

2.93 

2.81 

2.71 

2.62 

2.54 

.999 

11.4 

7.32 

5.79 

4.95 

4.42 

4.04 

3.77 

3.55 

3.38 

3.24 

3.12 

3.02 

.9995 

12.8 

8.10 

6.34 

5.39 

4.79 

4.37 

4.07 

3.82 

3.63 

3.47 

3.34 

3.22 

CO 

.0005 

.0839 

.0350 

.0251 

.016 

.032 

,050 

.069 

.088 

.108 

.127 

.144 

.161 

.001 

.On6 

.0210 

.0281 

,023 

.042 

.063 

.085 

.107 

.128 

.148 

.167 

.185 

.005 

.0^39 

.0250 

.024 

.052 

.082 

,113 

.141 

.168 

.193 

.216 

.236 

.256 

.01 

,0n6 

.010 

.038 

,074 

,111 

.145 

.177 

.206 

.232 

.256 

.278 

.298 

.025 

.0398 

.025 

.072 

.121 

.166 

.206 

.241 

.272 

.300 

.325 

.347 

.367 

.05 

,0239 

.051 

.117 

.178 

.229 

.273 

.310 

.342 

.369 

.394 

.417 

.436 

.10 

.016 

.105 

.195 

.266 

.322 

.367 

.405 

.436 

.463 

.487 

.508 

.525 

.25 

.102 

.288 

.404 

.481 

.535 

.576 

.608 

.634 

.655 

.674 

.690 

.703 

.50 

.455 

.693 

.789 

.839 

.870 

.891 

.907 

.918 

.927 

.934 

.939 

.945 

.75 

1.32 

1.39 

1.37 

1.35 

1.33 

1.31 

1.29 

1.28 

1.27 

1.25 

1.24 

1.24 

.90 

2.71 

2.30 

2.08 

1.94 

1.85 

1.77 

1.72 

1.67 

1.63 

1.60 

1.57 

1.55 

.95 

3.84 

3.00 

2.60 

2.37 

2.21 

2,10 

2.01 

1.94 

1.88 

1.83 

1.79 

1.75 

.975 

5-02 

3.69 

3.12 

2.79 

2.57 

2.41 

2.29 

2.19 

2.11 

2.05 

1.99 

1.94 

.99 

6.63 

4.61 

3.78 

3.32 

3.02 

2.80 

2.64 

2.51 

2.41 

2.32 

2.25 

2.18 

.995 

7.88 

5.30 

4.28 

3.72 

3.35 

3.09 

2.90 

2.74 

2.62 

2.52 

2.43 

2.36 

.999 

10.8 

6.91 

5.42 

4,62 

4.10 

3.74 

3.47 

3.27 

3.10 

2.96 

2.84 

2.74 

.9995 

12.1 

7.60 

5.91 

5.00 

4.42 

4.02 

3.72 

3.48 

3.30 

3.14 

3.02 

2.90 

£5421 


"2 

15 

20 

24 

30 

40 

50 

60 

100 

120 

200 

500 

CO 

120 

.0005 

.199 

.256 

.293 

.338 

.390 

.429 

.458 

.524 

.543 

.578 

.614 

.676 

.001 

.223 

.282 

.319 

.363 

.415 

.453 

.480 

.542 

.568 

.595 

.631 

.691 

.005 

.297 

,356 

.393 

.434 

.484 

.520 

.545 

.605 

.623 

.661 

.702 

.733 

.01 

.338 

.397 

.433 

.474 

.522 

.556 

.579 

.636 

.652 

.688 

.725 

.755 

.025 

.406 

.464 

.498 

.536 

.580 

.611 

.633 

.684 

.698 

.729 

.762 

.789 

.05 

.473 

.527 

.559 

.594 

.634 

.661 

.682 

.727 

.740 

.767 

.785 

.819 

.10 

.560 

.609 

.636 

.667 

.702 

.726 

.742 

.781 

.791 

.815 

.838 

.855 

.25 

.730 

.765 

.784 

.805 

.828 

.843 

.853 

.877 

.884 

.897 

.911 

.923 

.50 

.961 

.972 

.978 

.983 

.989 

.992 

.994 

1.00 

1.00 

1.00 

1-01 

1.01 

.75 

1.24 

1.22 

1.21 

1.19 

1.18 

1.17 

1.16 

1.14 

1.13 

1.12 

1,11 

1.10 

.90 

1.55 

1.48 

1.45 

1.41 

1.37 

1.34 

1.32 

1.27 

1.26 

1.24 

1.21 

1.19 

,95 

1.75 

1.66 

1.61 

1.55 

1.50 

1.46 

1.43 

1.37 

1.35 

1.32 

1*.2S 

1.25 

.975 

1.95 

1.82 

1.76 

1.69 

1.61 

1.56 

1.53 

1.45 

1.43 

1.39 

1.34 

1.31 

.99 

2.19 

2.03 

1.95 

1.86 

1.76 

1.70 

1.66 

1.56 

1.53 

1-48 

1.42 

1.38 

.995 

2.37 

2.19 

2.09 

1.98 

1.87 

1.80 

1.75 

1.64 

1.61 

1-54 

1.48 

1.43 

.999 

2.78 

2.53 

2.40 

2.26 

2.11 

2,02 

1.95 

1.82 

1.76 

1-70 

1.62 

1.54 

.9995 

2.96 

2.67 

2.53 

2.38 

2.21 

2.11 

2.01 

1.88 

1.84 

1.75 

1.67 

1.60 

00 

.0005 

.207 

.270 

.311 

,360 

.422 

.469 

.505 

.599 

.624 

.704 

.304 

1.00 

.001 

.232 

.296 

.338 

.386 

.448 

.493 

.527 

.617 

.649 

.719 

.819 

l.OO 

.005 

.307 

.372 

.412 

.460 

.518 

.559 

.592 

.671 

.699 

.762 

.843 

l.OO 

.01 

.349 

.413 

.452 

.499 

,554 

.595 

.625 

.699 

,724 

.782 

.858 

1.00 

.025 

.418 

.480 

.517 

.560 

.611 

.645 

.675 

.741 

.763 

.813 

.878 

l.OO 

.05 

.484 

.543 

.577 

.617 

.663 

.694 

.720 

.781 

.797 

.840 

.896 

1.00 

.10 

.570 

.622 

.652 

.687 

.726 

.752 

.774 

.826 

.838 

.877 

.919 

1.00 

.25 

.736 

.773 

.793 

.816 

.842 

.860 

.872 

.901 

.910 

.932 

.957 

l.OO 

.50 

.956 

.967 

.972 

.978 

.983 

.987 

.989 

.993 

.994 

.997 

.999 

1.00 

.75 

1.22 

1.19 

1.18 

1.16 

1.14 

1.13 

1.12 

1.09 

1.08 

1.07 

1.04 

1.00 

.90 

1.49 

1.42 

1.38 

1.34 

1.30 

1.26 

1.24 

1.18 

1.17 

1.13 

l.OS 

1.00 

.95 

1.67 

1.57 

1.52 

1.46 

1.39 

1.35 

1.32 

1.24 

1.22 

1.17 

1.11 

1.00 

.975 

1.83 

1.71 

1.64 

1.57 

1.48 

1,43 

1.39 

1.30 

1.27 

1.21 

1.13 

1.00 

.99 

2.04 

1.88 

1.79 

1.7O 

1.59 

1.52 

1.47 

1.36 

1.32 

1.25 

1.15 

1.00 

.995 

2.19 

2.00 

1.90 

1.79 

1.67 

1.59 

1.53 

1.40 

1.36 

1.28 

1.17 

1.00 

.999 

2.51 

2.27 

2.13 

1.99 

1.84 

1.73 

1.66 

1.49 

1.45 

1.34 

1.21 

1.00 

.9995 

2.65 

2.37 

2.22 

2.07 

1.91 

1.79 

1.71 

1,53 

1.48 

1.36 

1.22 

1.00 

C5431 


APPENDIX     71 

RANDOM  NUMBERS 


s  s 

S  8 

G>  s^^ 

xo   ON 

S  SI 

VO   O\ 
O3   O-4 

CD   T*< 
co   co 

w}   O\ 
co   CO 

9  ^ 

^  gs 

00 

39591 

66O82 

48626 

95780 

55228 

87189 

75717 

97042 

19696 

48613 

01 

463O4 

97377 

43462 

21739 

14566 

72533 

60171 

29024 

77581 

72760 

02 

99547 

60779 

22734 

23678 

44895 

89767 

18249 

41702 

35850 

40543 

03 

06743 

63537 

24553 

77225 

94743 

79448 

12753 

95986 

78088 

48019 

04 

69568 

65496 

49033 

88577 

98606 

92156 

08846 

54912 

12691 

13170 

05 

68198 

69571 

34349 

73141 

4264O 

44721 

30462 

35075 

33475 

474O7 

06 

27974 

12609 

77428 

64441 

49008 

60489 

66780 

55499 

80842 

57706 

07 

50552 

20688 

02769 

63O37 

15494 

71784 

70559 

58158 

53437 

46216 

08 

74687 

02033 

98290 

62635 

88877 

28599 

63682 

35566 

03271 

05651 

09 

49303 

76629 

71897 

30990 

62923 

36686 

96167 

11492 

90333 

84501 

10 

89734 

39183 

52026 

14997 

15140 

18250 

62831 

51236 

61236 

09179 

11 

74042 

40747 

O2617 

11346 

01884 

82066 

55913 

72422 

13971 

64209 

12 

84706 

31375 

67053 

73367 

95349 

31074 

36908 

42782 

89690 

480O2 

13 

83664 

21365 

28882 

48926 

45435 

60577 

85270 

02777 

06878 

27561 

14 

47813 

74854 

73388 

11385 

99108 

97878 

32858 

17473 

07682 

20166 

15 

O0371 

56525 

38880 

53702 

09517 

47281 

15995 

98350 

25233 

79718 

16 

81182 

48434 

27431 

55806 

25389 

4O774 

72978 

16835 

65066 

28732 

17 

75242 

35904 

73077 

24537 

81354 

48902 

03478 

42867 

04552 

66034 

18 

96239 

80246 

O70OO 

09555 

55051 

49596 

44629 

88225 

28195 

44598 

19 

82988 

1744O 

85311 

03360 

38176 

51462 

86070 

03924 

84413 

92363 

20 

77599 

29143 

89088 

57593 

60036 

17297 

30923 

36224 

46327 

96266 

21 

61433 

33118 

53488 

82981 

44709 

63655 

64388 

00498 

14135 

57514 

22 

76008 

15045 

45440 

84062 

52363 

18079 

33726 

44301 

86246 

99727 

23 

26494 

76598 

85834 

10844 

5630O 

02244 

72118 

96510 

98388 

80161 

24 

46570 

88558 

77533 

33359 

07830 

84752 

53260 

46755 

36881 

98535 

25 

73995 

41532 

87933 

79930 

14310 

64833 

49020 

70067 

99726 

970O7 

26 

93901 

38276 

75544 

19679 

62899 

11365 

22896 

42118 

77165 

08734 

27 

41925 

28215 

40966 

93501 

45446 

27913 

21708 

01788 

81404 

15119 

28 

80720 

02782 

24326 

41328 

10357 

86883 

80086 

77138 

57072 

121OO 

29 

92596 

39416 

5O362 

04423 

04561 

58179 

54188 

44978 

14322 

97056 

30 

39693 

58559 

45839 

47278 

38548 

38885 

19875 

26829 

86711 

57O05 

31 

86923 

37863 

14340 

30929 

04079 

65274 

03030 

15106 

09362 

82972 

32 

9970O 

79237 

18172 

58879 

56221 

65644 

33331 

87502 

32961 

40996 

33 

60248 

21953 

52321 

16984 

03252 

9O433 

97304 

50181 

71026 

01946 

34 

29136 

71987 

03992 

67025 

31070 

78348 

47823 

11033 

13037 

47732 

35 

57471 

42913 

85212 

42319 

92901 

97727 

04775 

94396 

38154 

25238 

36 

57424 

93847 

03269 

56096 

95028 

14039 

76128 

63747 

27301 

65529 

37 

56768 

71694 

63361 

80836 

30841 

71875 

40944 

54827 

01887 

54822 

38 

70400 

81534 

02148 

41441 

26582 

27481 

84262 

14084 

42409 

62950 

39 

05454 

88418 

48646 

99565 

36635 

85496 

18894 

77271 

26894 

O0889 

40 

80934 

56136 

47063 

96311 

19067 

59790 

08752 

68040 

85685 

83076 

41 

O6919 

46237 

50676 

11238 

75637 

43086 

95323 

52867 

06891 

32089 

42 

00152 

23997 

41751 

74756 

50975 

75365 

70158 

67663 

51431 

46375 

43 

88505 

74625 

71783 

82511 

13661 

63178 

39291 

76796 

74736 

10980 

44 

64514 

80967 

33545 

09582 

86329 

58152 

05931 

35961 

70O69 

12142 

45 

25280 

53007 

99651 

96366 

49378 

8O971 

10419 

12981 

70572 

11575 

46 

71292 

63716 

9321O 

59312 

39493 

24252 

54849 

29754 

41497 

79228 

47 

49734 

50498 

O8974 

05904 

68172 

02864 

10994 

22482 

12912 

17920 

48 

43075 

09754 

71880 

92614 

99928 

94424 

86353 

87549 

94499 

11459 

49 

15116 

16643 

O3981 

06566 

14050 

33671 

03814 

48856 

41267 

76252 

1  Reproduced  from  George  W.  Snedecor,  Everyday  Statistics.  Copyright  1950. 
Published  by  Wm.  C.  Brown  Company,  I>ubuque,  Iowa.  !By  permission  of  the 
author  and  publishers. 


3  S 

10   ON 

»o  to 

s  s 

^8  © 

O   ^* 
J>-  *— 

*0   0\ 
•**--  r— 

s  a 

_ 

£  a 

s  3: 

—  —  —  

*0   ON 

25178 

O  C  1  fi/^. 

41773 

39926 

09843 

29694 

43801 

69276 

44707 

23455 

00 

rrOoVJO 

yj  _L\JO 

ooolo 

oOoOO 

37383 

76832 

37024 

06581 

r\i 

99068 
70983 

30898 
35453 
88359 

42152 
95583 

13923 
12078 
79848 

44987 
04913 
24101 

45122 
06083 
67502 

86515 
06645 
25692 

55836 
93310 
42496 

96165 
40016 

19650 

Ul 

04 

71181 

48289 

03153 

18779 

65702 

03612 

64608 

84071 

47588 

09982 

05 

91555 

87708 

70964 

43346 

27731 
56811 

08725 

75139 

77674 

44816 
82467 

07 

54307 

12188 

58089 

73745 

35569 

97352 

77301 

37684 

6991  R 

HQ 

63631 

23919 

06785 

13891 

89918 

76211 

09362 

34292 

17640 

65907 

Uo 

09 

46832 
05944 
28199 
08391 
29634 

90116 
05600 
13021 

98898 
39310 
59501 
00624 
96568 

49025 
95068 
15124 

97793 
61725 
73005 
33776 
55092 

84954 
44985 
44043 

36775 
81937 
11587 
01505 
31073 

71974 
16820 
97691 
76911 
92371 

15574 
85446 
90415 
45539 
51288 

09184 
51168 
84685 
32181 
33378 

12 
13. 
14 

61509 

18842 

79201 

46451 

68594 

98120 

68110 

91062 

42095 

61839 

15 

87888 

23033 

69837 

65661 

15130 

44649^ 

42515 

83861 

50721 

36110 

16 

94585 

15218 

74838 

61809 

92293 

S5400 

46934 

08531 

70107 

65707 

j.  \j 
17 

82033 

93915 

34898 

79913 

70013 

27573 

39256 

35167 

35070 

47095 

18 

79131 

10022 

82199 

78976 

22702 

37936 

10445 

96846 

84927 

69745 

19 

79344 

39236 

41333 

11473 

15049 

47930 

99029 

97150 

82275 

55149 

20 

15384 

44585 

18773 

89733 

40779 

59664 

83328 

25162 

58758 

17761 

21 

38802 

90957 

32910 

97485 

10358 

88588 

95310 

22252 

19143 

69011 

22 

85874 

18400 

28151 

29541 

63706 

43197 

65726 

94117 

22169 

91806 

23 

26200 

72680 

12364 

46010 

92208 

59103 

60417 

45389 

56122 

85353 

24 

13772 

75282 

81418 

42188 

66529 

47981 

92548 

10079 

68179 

40915 

25 

91876 

07434 

96946 

98382 

97374 

34-4_44. 

17992 

42811 

01579 

48741 

26 

31721 

21713 

83632 

40605 

24227 

53219 

05482 

86768 

53239 

24812 

27 

92570 

53242 

98133 

84706 

78048 

29645 

79336 

66091 

05793 

25922 

28 

02880 

29307 

73734 

66448 

64739 

74645 

29562 

13999 

17492 

49891 

29 

80982 

14684 

31038 

85302 

98349 

57313 

86371 

33938 

10768 

60837 

30 

38000 

43364 

94825 

32413 

46781 

09685 

69058 

56644 

85531 

55173 

31 

14218 

94289 

79484 

61868 

40034 

22546 

68726 

14736 

89844 

13466 

32 

74358 

21940 

40280 

22233 

09123 

49375 

55094 

46113 

54046 

51771 

33 

39049 

14986 

94000 

26649 

13037 

34609 

45186 

89515 

63214 

66886 

34 

48727 

06300 

91486 

67316 

84576 

11100 

37580 

49629 

83224 

46321 

35 

22719 

29784 

40682 

96715 

40745 

57458 

70048 

48306 

50270 

87424 

36 

33980 

36769 

51977 

03689 

79071 

20279 

64787 

48877 

44063 

93733 

37 

23885 

66721 

16542 

12648 

65986 

43104 

45583 

75729 

35118 

58742 

38 

85190 

44068 

78477 

69133 

58983 

96504 

44232 

74809 

25266 

73872 

39 

33453 

36333 

45814 

78128 

55914 

89829 

43251 

41634 

48488 

49153 

40 

98236 

11489 

97240 

01678 

30779 

75214 

80039 

68895 

95271 

19654 

41 

21295 

53563 

43609 

48439 

87427 

88065 

09892 

58524 

43815 

31340 

42 

28335 

79849 

69842 

71669 

38770 

54445 

48736 

03242 

83181 

85403 

43 

95449 

35273 

62581 

85522 

35813 

34475 

97514 

72839 

10387 

31649 

44 

88167 

03878 

89405 

55461 

73248 

48620 

31732 

47317 

06252 

54652 

45 

86131 

62596 

98785 

02360 

54271 

26242 

93735 

20752 

17146 

18315 

46 

71134 

90264 

30126 

08586 

97497 

61678 

81940 

00907 

39096 

02082 

47 

02664 

53438 

76839 

52290 

77999 

05799 

93744 

16634 

84924 

31344 

48 

90664 

96876 

16663 

25608 

67140 

84619 

67167 

13192 

81774 

58619 

49 

[545] 


50 
51 
52 
53 
54 

55 
56 
57 
58 
59 

60 
61 
62 
63 
64 

65 
66 
67 
68 
69 

70 
71 

72 
73 
74 

75 
76 
77 
78 
79 

80 
81 
82 
83 
84 

85 
86 
87 
88 
89 

90 
91 
92 
93 
94 

95 
96 
97 
98 
99 


93873  86558  72524  02542  73184  37905  05882  15596  73646  50798 

08761  47547  02216  48086  56490  89959  69975  04500  23779  76697 

61270  98773  40298  26077  80396  08166  35723  61933  13985  19102 

73758  15578  95748  02967  35122  36539  72822  68241  34803  42457 

17132  32196  60523  00544  73700  70122  27962  85597  36011  79971 

26175  29794  44838  84414  82748  22246  70694  57953  39780  17791 

06004  04516  06210  03536  84451  30767  37928  26986  07396  64611 

34687  73753  36327  73704  61564  99434  90938  03967  97420  19913 

27865  08255  57859  04746  79700  68823  16002  58115  07589  12675 

89423  51114  90820  26786  77404  05795  49036  34686  98767  32284 

99030  80312  69745  87636  10058  84834  89485  08775  19041  61375 

02852  54339  45496  20587  85921  06763  68873  35367  42627  54973 

10850  42788  94737  74549  74296  13053  46816  32141  02533  25648 

38301  18507  33151  69434  80103  02603  61110  89395  67621  67025 

48181  95478  62739  90148  00156  09338  44558  53271  87549  45974 

23098  23720  76508  69083  56584  90423  21634  35999  09234  95116 

25104  82019  21120  06165  44324  77577  15774  44091  69687  67576 

22205  40198  86884  28103  57306  54915  03426  66700  45993  36668 

64975  05064  29617  40622  20330  18518  45312  57921  23188  82361 

58710  75278  47730  26093  16436  38868  76861  85914  14162  21984 

12140  72905  26022  07675  16362  34504  47740  39923  04081  03162 

73226  39840  47958  97249  14146  34543  76162  74158  59739  67447 

12320  86217  66162  70941  58940  58006  80731  66680  02183  94678 

41364  64156  23000  23188  64945  33815  32884  76955  56574  61666 

97881  80867  70117  72041  03554  29087  19767  71838  80545  61402 

88295  87271  82812  97588  09960  06312  03050  77332  25977  18385 

95321  89836  78230  46037  72483  87533  74571  88859  26908  55626 

24337  14264  30185  36753  22343  81737  62926  76494  93536  75502 

00718  66303  75009  91431  64245  61863  16738  23127  89435  45109 

38093  10328  96998  91386  34967  40407  48380  09115  59367  49596 

87661  31701  29974  56777  66751  35181  63887  95094  20056  84990 

87142  91818  51857  85061  17890  39057  44506  00969  32942  54794 

60634  27142  21199  50437  04685  70252  91453  75952  66753  50664 

73356  64431  05068  56334  34487  78253  67684  69916  63885  88491 

29889  11378  65915  66776  95034  81447  98035  16815  68432  63020 

48257  36438  48479  72173  31418  14035  84239  02032  40409  11715 

38425  29462  79880  45713  90049  01136  72426  25077  64361  94284 

48226  31868  38620  12135  28346  17552  03293  42618  44151  78438 

80189  30031  15435  76730  58565  29817  36775  64007  47912  16754 

33208  33475  95219  29832  74569  50667  90569  66717  46958  04820 

19750  48564  49690  43352  53884  80125  47795  99701  06800  22794 

62820  23174  71124  36040  34873  95650  79059  23894  58534  78296 

95737  34362  81520  79481  26442  37826  76866  01580  83713  94272 

64642  62961  37566  41064  69372  84369  92823  91391  61056  44495 

77636  60163  14915  50744  95611  99346  39741  04407  72940  87936 

43633  52102  93561  31010  11299  52661  79014  17910  88492  60753 

93686  41960  61280  96529  52924  87371  34855  67125  40279  10186 

23775  33402  28647  42314  51213  29116  26243  40243  32137  25177 

91325  64698  58868  63107  08993  96000  66854  11567  80604  72299 

58129  44367  31924  73586  24422  92799  28963  36444  01315  10226 


[5461 


ro  31209  83677  99115  94024  09286  58927  24078  16770 

58108  29344  11825  51955  50618  99753  02200  50503  32466  5OO55 

71545  42326  66429  93607  55276  85482  24449  41764  19884  46443 

93303  90557  79166  90097  01627  96690  77434  06402  05379  59549 

36731  37929  13079  83036  31525  35811  59131  65257  03731  §670? 


«nnc   nH£o   !229i   846°8   2339°   30433   °8240   85136   ^0060   43651 


«nnc   no 

65995  94208  68785  04370  44192  91852  01129  28739  08705 

09309  02836  10223  90814  92786  96747  46014  54765  76001 

63812  47615  1722°  27942  11785  4"33  03923  35432 


95407  95006  95421  20811  76761  47475  58865  06204  36543  81O02 

22789  87011  61926  97996  10604  80855  48714  52754  98279  96467 

96783  18403  36729  18760  30810  73087  94565  68682  15792  60020 

68933  05665  12264  23954  01583  75411  04460  83939  66528  22576 

68794  13000  20066  98963  93483  51165  63358  12373  13877  37580 

40537  31604  60323  51235  65546  85117  15647  09617  73520  48525 

41249  42504  91773  81579  02882  74657  73765  10932  74607  83825 

08813  84525  30329  33144  76884  89996  07834  67266  96820  15128 

46609  30917  29996  10848  39555  09233  58988  82131  69232  76762 

68543  69424  92072  57937  05563  80727  67053  35431  OO881  56541 

09926  84219  30089  08843  24998  27105  18397  79071  40738  73876 

30515  76316  49597  37900  98604  05857  51729  19006  15239  27129 

21611  26346  04877  71584  55724  39616  64648  36811  60915  34108 

47410  83767  56454  96768  27001  83712  01245  27256  57991  75758 

18572  31214  41015  64110  61807  72472  78059  69701  78681  17356 

28078  02819  02459  33308  96540  15817  78694  81476  87856  99737 

56644  50430  34562  75842  67724  02918  55603  55195  88219  39676 

27331  48055  18928  47763  61966  64507  06559  81329  29481  03660 

32080  21524  32929  07739  00836  39497  94476  27433  96857  52987 

27027  69762  65362  90214  89572  52054  43067  73017  87664  03293 

56471  68839  09969  45853  72627  71793  49920  64544  71874  74053 

22689  19799  18870  49272  74783  38777  76176  40961  18089  32499 

71263  82247  66684  90239  67686  48963  30842  59354  33551  87966 

64084  57386  89278  27187  52142  96305  87393  80164  95518  82742 

23121  10194  09911  37062  43446  09107  47156  70179  00858  92326 

78906  48080  76745  65814  51167  87755  66884  12718  14951  47937 

87257  26005  21544  37223  53288  72056  96396  67099  49416  91891 

39529  98126  33694  29025  94308  24426  63072  51444  04718  49891 

89632  11606  87159  89408  06295  31055  15530  46432  49871  37982 

23708  98919  14407  53722  58779  92849  04176  24870  56688  25405 

51445  46758  42024  27940  64237  10086  95601  53923  85209  79385 

23849  65272  24743  39960  27313  99925  29743  87270  05773  21797 

78613  15441  34568  57398  25872  61792  94599  60944  90908  38948 

90694  27996  94181  87428  41135  29461  72716  68956  67871  72459 

96772  86829  36403  40087  67456  21071  39039  91937  45280  00066 

24527  40701  56894  73327  00789  97573  09303  41704  05772  95372 

31596  70876  46807  06741  29352  23829  52465  00336  24155  61871 

31613  99249  1726O  05242  19535  52702  64761  66694  06150  13820 

02911  09514  50864  80622  20017  59019  43450  75942  08567  40547 

02484  74068  04671  19646  41951  05111  34013  57443  87481  48994 

69259  75535  73007  15236  01572  44870  53280  25132  70276  87334 


[5471 


50 
51 

52 
53 
54 

55 
56 
57 
58 
59 

60 
61 
62 
63 
64 

65 
66 
67 
68 
69 

70 

71 
72 
73 
74 

75 
76 
77 
78 
79 

80 
81 
82 
83 
84 

85 
86 
87 
88 
89 

90 
91 
92 
93 
94 

95 
96 
97 
98 
99 


APPENDIX  81 

CONTROL  CHART  CONSTANTS 

1  Reprinted  from  ASTM  Manual  on  Quality  Control  of  Materials,  Special 
Technical  Publication  15-C,  Table  B2,  American  Society  for  Testing  Materials, 
Philadelphia,  1951.  By  permission  of  the  authors  and  publishers. 


[5481 


6 


s 


1 


oa  10  oa  T-I 

CO  01  O4  oj 


vOt-          •&  VO  CN  ^ 

-rH  *•—      •<•*<  -^H  os  *-^  to 

CO*-        t^O^v 


v£>  ,-1  00  O  VO 


-~  _  ^  A 

OOOO 


SO  \O  rf  co 

_^  *^  co  oo  es 

O  O  T-H  T-H  csj 


OOOO 


SO  -rH  00  Os  CO 

to  co  o  oq  -^ 
o"  o  o"  o  o* 


^  O\  CN  •<* 

sOr-^OsO 

co  co  co  -* 


CSi  Os 

10  to 


OOOOO 


£2  29  °° 


>O  co  J>- 
CO  O\  -^  ON  CO 


O 
lO 


ON  VO  •«—  i  CO 


IOIOLOIO1O 


--^.^._ 
OOOO 


O\COOOCOON      ^  o  VD  CM 


CM  ox 

*-t  o 


o  c>  o  o  o 


O  10  Os 

o  o"  o"  o* 


vo  co  CM  r—  o\ 

rf  OS  I-H  O  r*< 
Os  VO  1O  CO  O3 


s  co  »o  O       TH  i>-  r-.  T-I  i 

Os  co  OO        co  OO  -«*i  T-I  J 


-  oo  c<»  r>- 

-r-f  O\  V5 


OOOOO 


OOOOO         OOOOO 


C<lCsIC<lCSICO  COCOCOCOCO 


COCOCOCOCO 


OO  Os  CO  to  «— i 
*•—  •«— ( to  Os  co 
i-^  CO  CO  CO  Os 


Standard  Deviation 


Cha 


a? 


t-^OQVOOs 
vO'O'O 


OOO>O 


O  OO  IO  Os  ^ 

CO  *H  CO  CO  OO 


OO  VD  03  t^  O 

OST-i 


o  o  o  o  o      o  o  o  o  o      o  o  o  o  o 


PC) 


•rJHiOOJiO         T—  it^.co 
oOOOoCt—         l>-  vO  SO  S 


*—  »  *—  1  CO  * 

sOT^CNJ 

LO  iO  to  iO  ^*« 


-«^V-)t-~OsC*3 
Ol'«-HOOsOS 

-*  r^  -^  CO  rO 


ft 


OOOO 


O  -^H  ^-4  05  04 

o  o*  o  o"  o" 


05  co  co  co  ^ 
O0*o"oo" 


OOOOO 


10  to  to  to  to 

OOOOO 


t-ooiooo       iooio 


co^t^-ONt-^ 
iooOc>4t--co 
-r-^osototo 

OOOOO 


'—  "  Os 
co 
O 


<oooc<i.r--co 


OOO 


trt       t 

II 


O-r-t 
OsOs 


Os  O  CO  O    co  ^-*  ^O  Os  Os 
10  -^-t  10  O\    <M  IO  f-  O\  i-H 
- 


OsOs         OvOsOsOsOs         Os^OsOsOs 


CO  to  O> 
CO  to  1^ 

\O\OsOV 
OsOvOs 


Av 


J0404*-—         CO  -^  ff  CO  O         CO  so  r^  cO  01         ^-f  O  Os  CO  OO        JC^.  so  so  IO  to 
)  O  J>-  to         T^TfcOCOCO          O4OJOIO4OA         OJ  O?  T— I  T— I  T— i         -«-H  ^— i  ^-t  *-*-»-( 

o  o"  o  cs  o 


.. 

t--  co  oo  to       -^  es>  T-I  o  O       osos 


Os  O5  r*-  O5  Os 
SO  (^O  so  so  sO 
O*OOO"O 


•«-HOOOs         OsCOOOOOt^- 

^H*  ^  *-I      ^4  ^4  i-i  T-I  o      o  o  o  o"  o" 


CO  1>-  CO  ^—  I 


so  ^o  so  vo 


•I1? 


t 


tO         sot—  CO  Os  O 


15491 


APPENDIX    91 

NUMBER  OF  OBSERVATIONS  FOR 
f-TEST  OF  MEAN2 

1  Reproduced  from  Table  E  of  Owen  L.  Davies,   The  Design  and  Analysis  of 
Industrial   Experiments,    second   ed.,    Oliver   and    Boyd,    Edinburgh,    1956.    By 
permission  of  the  author  and  publishers. 

2  The  entries  in  this  table  show  the  number  of  observations  needed  in  a  i-test 
of  the  significance  of  a  mean  in  order  to  control  the  probabilities  of  errors  of  the 
first  and  second  kinds  at  a  and  j8,  respectively. 


[550] 


Level  of  f-Test 

0.01 

0.02 

0.05 

0.1 

Single-Sided  Test 

«=0.005 

«=0.01 

a-0  025 

Double-Sided  Teat 

a-0.01 

«=0.02 

a^O.05 

a=0.05 
«=0.1 

*~ 

0.01  0,05  0.1     0.2     0.5 

0.010.05    0.1     0.2    0.5 
i  

0.01  0.05  0.1    0.2     0.5 

0.01  0.05  0.1    0.2    0.5 

0.05 

i 

0.10 

0.05 

0.15 

0.10 

122 

0.15 

0.20 

139 

99 

70 

0.20 

0.25 

110 

90 

128      64 

139     101      45 

0.25 

0,30 

134      78 

115      63 

119      90      45 

122      97      71      32 

0.30 

0.35 

125       99      58 

109      85      47 

109      88      67      34 

90      72      52      24 

0.35 

0.40 

115    *97      77      45 

101       85      66      37 

117      84      68      51      26 

101      70      55      40      19 

0.40 

0,45 

92     ;77      62      37 

110      81      68      53      30 

93      67      54      41      21 

80      55      44      33      15 

0.45 

0.50 

100      75      63      51      30 

90      66      55      43      25 

76      54      44      34      18 

65      45      36      27      13 

0.50 

0.55 

83      63      53      42      26 

75      55      46      36      21 

63      45      37      28      15 

54      38      30      22      11 

0.55 

0.60 

71      53      45      36      22 

63      47      39      31      18 

53      38      32      24      13 

46      32      26      19        9 

0.60 

0.65 

61      46      39      31      20 

55      41      34      27      16 

46      33      27      21      12 

39      28      22      17        8 

0.65 

0.70 

53      40      34      28      17 

47      35      30      24      14 

40      29      24      19      10 

34      24      19       15        8 

0.70 

0.75 

47      36      30      25      16 

42      31      27      21      13 

35      26      21       16        9 

30      21      17      13        7 

0.75 

0.80 

41      32      27      22      14 

37      28      24      19      12 

31      22      19      15        9 

27      19      15      12        6 

0.80 

0.85 

37      29      24      20      13 

33      25      21      17      11 

28      21      17      13        8 

24      17      14      11        6 

0.85 

Value  of 

^.90 

34      26      22      18      12 

29      23      19      16      10 

25      19      16      12        7 

21       15      13       10        5 

0.90 

s 

0.95 

31      24      20      17      11 

27      21      18      14        9 

23      17      14      11        7 

19      14      11        9        5 

0.95 

*>=? 

1.00 

28      22      19      16      10 

25      19      16      13        9 

21      16      13      10        6 

18      13      11        85 

1.00 

1.1 

24      19      16      14        9 

21      16      14      12        8 

18      13      11        9        6 

15      11        9        7 

1.1 

1.2 

21      16      14      12        8 

18      14      12      10        7 

15      12      10        8        5 

13      10        8        6 

1.2 

1,3 

18      15      13      11        8 

16      13      11        9        6 

14      10        9        7 

11        8        7        6 

1.3 

1.4 

16      13      12      10        7 

14      11      10        9        6 

12        9        8        7 

10        8        7        5 

1.4 

1.5 

15      12      11        9        7 

13      10        9        8        6 

11        8        7        6 

976 

1.5 

1.6 

13      11       10        8        6 

12      10        9        7        5 

10        8        7        6 

866 

1.6 

1.7 

12      10        9        8        6 

11        9        8        7 

9765 

865 

1.7 

1.8 

12      10        9        86 

10        8        7        7 

876 

7        6 

1.8 

1.9 

11        9        8        7        6 

10        8        7        6 

866 

7        5 

1.9 

2.0 

10        8        8        7        5 

9776 

765 

6 

2.0 

2.1 

10        8        7        7 

8766 

7        6 

6 

2.1 

2.2 

9876 

8765 

7        6 

6 

2.2 

2.3 

9776 

866 

6        5 

5 

2.3 

2.4 

8776 

766 

6 

2.4 

2.5 

8766 

766 

6 

2.5 

3.0 

7665 

655 

5 

3.0 

3.5 

655 

5 

3.5 

4.0 

6 

4.0 

C5511 


APPENDIX    1  Ol 

NUMBER  OF  OBSERVATIONS  FOR 
f-TEST  OF  DIFFERENCE  BETWEEN 
TWO  MEANS2 

1  Reproduced  from  Table  E.I  of  Owen  L.  Davies,  The  Design  and  Analysis  of 
Industrial  Experiments,  second  ed'.,  Oliver  and  Boyd,  Edinburgh,  1956.  By  per 
mission  of  the  author  and  publishers. 

2  The  entries  in  this  table  show  the  number  of  observations  needed  in  a  Z-test 
of  the  significance  of  the  difference  between  two  means  in  order  to  control  the 
probabilities  of  the  errors  of  the  first  and  second  kinds  at  a.  and  ft  respectively. 
"It  should  be  noted  that  the  entries  in  the  table  show  the  number  of  observa 
tions  needed  in  each  of  two  samples  of  equal  size.77 


I552I 


Level  of  f-Test 

0.01 

0.02 

0.05 

0.1 

Single-Sided  Test 

<x=0.005 

a^O.Ol 

<*=*0.025 

a=0.05 

Double-Sided  Test 

a=0.01 

a=0.02 

a=0.05 

a=0.1 

0- 

0.01  0.05   0.1    0.2    0.5 

.01  0.05    0.1    0.2    0.5 

.01  0.05   0.1    0.2    0.5 

0.01  0.05   0.1    0.2    0.5 

0.05 

0.05 

0.10 

0.10 

0.15 

0.15 

0.20 

137 

0.20 

0.25 

124 

88 

0.25 

0.30 

123 

87 

61 

0.30 

0.35 

110 

90 

64 

102      45 

0.35 

0.40 

85 

70 

100      50 

108      78      35 

0.40 

0.45 

118      68 

101      55 

105      79      39 

108      86      62      28 

0.45 

0.50 

96      55 

106      82      45 

106      86      64      32 

88      70      51      23 

0.50 

0.55 

101      79      46 

106      88      68      38 

87      71      53      27 

12      73      58      42      19 

0.55 

0.60 

101      85      67      39 

90      74      58      32 

04      74      60      45      23 

89      61      49      36      16 

0.60 

0.65 

87      73      57      34 

04      77      64      49      27 

88      63      51      39      20 

76      52      42      30      14 

0.&5 

0.70 

100      75      63      SO      29 

90      66      55      43      24 

76      55      44      34      17 

66      45      36      26      12 

0.70 

0.75 

88      66      55      44      26 

79      58      48      38      21 

67      48      39      29      15 

57      40      32      23      11 

0.75 

0.80 

77      58      49      39      23 

70      51      43      33      19 

59      42      34      26      14 

50      35      28      21      10 

0.80 

0.85 

69      51      43      35      21 

62      46      38      30      17 

52      37      31      23      12 

45      31      25      18       9 

0.85 

Vaiue  of 

0.90 

62      46      39      31      19 

55      41      34      27      15 

47      34      27      21      11 

40      28      22      16        8 

0.90 

0.95 

55      42      35      28      17 

50      37      31      24      14 

42      30      25      19      10 

36      25      20      15        7 

0.95 

B--5- 

1.00 

50      38      32      26      15 

45      33      28      22      13 

38      27      23      17        9 

33      23      18      14        7 

1.00 

or 

1.1 

42      32      27      22      13 

38      28      23      19      11 

32      23      19      14        8 

27      19      15      12        6 

1.1 

1.2 

36      27      23      18      11 

32      24      20      16        9 

27      20      16      12        7 

23      16      13      10        5 

1.2 

1.3 

31      23      20      16      10 

28      21      17      14        8 

23      17      14      11        6 

20      14      11        9        5 

1.3 

1.4 

27      20      17      14        9 

24      18      15      12        8 

20      15      12      10        6 

17      12      10        8        4 

1.4 

1.5 

24      18      15      13        8 

21      16      14      11        7 

18      13      11        9        5 

15      11        9        7       4 

1.5 

1.6 

21       16      14      11        7 

19      14      12      10        6 

16      12      10        8        5 

14      10        8        6        4 

1.5 

1.7 

19      15      13      10        7 

17      13      11        9        6 

14      11        9        7        4 

12        9        7        6        3 

1.7 

1.8 

17      13      11      10        6 

15      12      10        8        5 

13      10        8        6        4 

11        8        7        5 

1.8 

1,9 
2.0 

16      12      11        9        6 
14      11      10        8        6 

14      11        9        8        5 
13      10        9        7        5 

12        9        7        6        4 
11        8        7        6        4 

10        7        6        5 
9764 

1.9 
2.0 

2.1 
2.2 
2.3 
2.4 
2.5 

13      10        9        8        5 
12      10        8        7        5 
11        9        8        7        5 
11        9        8        6        5 
10        8        7        6        4 

12        9        8        7        5 
11        9        7        6        4 
10        8        7        6        4 
10        8        7        6        4 
97654 

10        8        6        5        3 
9765 
9765 
8654 
8654 

8654 
8654 
7554 
7544 
6543 

2.1 
2.2 
2.3 
2.4 
2.5 

3.0 
3.5 

86654 
65543 

76543 
6544 

6544 
5443 

543 

4        3 

3.0 
3.5 
4.0 

4.0 

6544 

5443 

443 

4 

[5531 


APPENDIX    1 


NUMBER  OF  OBSERVATIONS  REQUIRED 
FOR  THE  COMPARISON  OF  A 
POPULATION  VARIANCE  WITH  A 
STANDARD  VALUE  USING  THE     -TEST 


V 

0  =  0.01 

0  =  0.05 

).01 

0=0.5 

0  =  0.01 

0  =  0.05 

.05 
0  =  0.1 

0  =  0.5 

1 

42,240 

1,687 

420.2 

14.58 

24,450 

977.0 

243.3 

8.444 

2 

458.2 

89.78 

43.71 

6.644 

298.1 

58.40 

28.43 

4.322 

3 

98.79 

32.24 

19.41 

4.795 

68.05 

22.21 

13.37 

3.303 

4 

44,69 

18.68 

12.48 

3.955 

31.93 

13,35 

8.920 

2.826 

5 

27.22 

13.17 

9.369 

3.467 

19.97 

9.665 

6.875 

2.544 

6 

19.28 

10.28 

7.628 

3.144 

14.44 

7.699 

5.713 

2.354 

7 

14.91 

8.524 

6.521 

2.911 

11.35 

6,491 

4.965 

2.217 

8 

12.20 

7.352 

5.757 

2.736 

9.418 

5,675 

4.444 

2.112 

9 

10.38 

6.516 

5.198 

2.597 

8.103 

5.088 

4.059 

2.028 

10 

9.072 

5.890 

4.770 

2.484 

7.156 

4.646 

3.763 

1.960 

12 

7.343 

5.017 

4.159 

2.312 

5,889 

4.023 

3.335 

1.854 

15 

5.847 

4.211 

3.578 

2.132 

4.780 

3.442 

2.925 

1.743 

20 

4.548 

3.462 

3.019 

1.943 

3.802 

2.895 

2.524 

1.624 

24 

3,959 

3.104 

2.745 

1.842 

3.354 

2.630 

2.326 

1.560 

30 

3.403 

2.752 

2.471 

1.735 

2.927 

2.367 

2.125 

1.492 

40 

2.874 

2.403 

2.192 

1.619 

2.516 

2.103 

1.919 

1.418 

60 

2.358 

2.046 

1.902 

1.490 

2.110 

1.831 

1.702 

1.333 

120 

1.829 

1.661 

1.580 

1.332 

1.686 

1.532 

1.457 

1.228 

CO 

1.000 

1.000 

1.000 

1.000 

1.000 

1.000 

1.000 

1.000 

1  Reproduced  from  Table  G  of  Owen  L.  Davies,  The  Design  and  Analysis  of 
Industrial    Experiments,    second   ed.,    Oliver   and    Boyd,    Edinburgh,    1956.    By 
permission  of  the  author  and  publishers. 

2  The  entries  in  this  table  show  the  value  of  the  ratio  R  of  the  population 
variance  <r2  to  a  standard  variance  &l  which  will  be  undetected  with  probability 
0  in  a  x2-test  at  the  100<*  per  cent  significance  level  of  an  estimate  s2  of  cr2  based 
on  v  degrees  of  freedom. 


[554] 


APPENDIX    1  21 


NUMBER  OF  OBSERVATIONS  REQUIRED 
FOR  THE  COMPARISON  OF  TWO 
POPULATION  VARIANCES  USING 
THE  F-TEST2 


V 

0=0.01 

0=0.05 

L 
0=0.1 

0=0.5 

0-0.01 

a=0,05 
0=0.05   0=0.1 

0=0.5 

0=0.01 

«=0.5 

0=0.05  0=0.1 

0=0.5 

1 

16,420,000 

654,200 

161,500 

4,052 

654,200 

26,070       6,436 

161.5 

4,052 

161.5 

39.85 

1.000 

2 

9,801 

1,881 

891.0 

99.00 

1,881 

361.0       171.0 

19.00 

99.00 

19.00 

9.000 

1.000 

3 

867.7 

273.3 

158.8 

29.46 

273.3 

86.06       50.01 

9.277 

29.46 

9.277 

5.391 

1.000 

4 

255.3 

102.1 

65.62 

15.98 

102.1 

40.81       26.24 

6.388 

15.98 

6.388 

4.108 

1.000 

5 

120.3 

55.39 

37.87 

10.97 

55.39 

25,51       17.44 

5.050 

10.97 

5.050 

3.453 

1.000 

6 

71.67 

36.27 

25.86 

8.466 

36.27 

18.35       13.09 

4.284 

8.466 

4.284 

3.056 

1.000 

7 

48.90 

26.48 

19.47 

6.993 

26.48 

14.34       10.55 

3,787 

6.993 

3.787 

2.786 

1.000 

8 

36.35 

20.73 

15.61 

6.029 

20.73 

11.82       8.902 

3.438 

6.029 

3.438 

2.589 

1.000 

9 

28.63 

17.01 

13.06 

5.351 

17.01 

10.11       7,757 

3.179 

5.351 

3.179 

2.440 

1.000 

10 

23.51 

14.44 

11.26 

4.849 

14.44 

8.870       6.917 

2.978 

4.849 

2.978 

2.323 

1.000 

12 

17.27 

11.16 

8.923 

4.155 

11.16 

7.218      5.769 

2.687 

4.155 

2.687 

2.147 

1.000 

15 

12.41 

8.466 

6.946 

3.522 

8.466 

5,777      4.740 

2.404 

3.522 

2.404 

1.972 

1.000 

20 

8.630 

6.240 

5.270 

2.938 

6.240 

4.512       3.810 

2.124 

2.938 

2.124 

1,794 

1.000 

24 

7.071 

5.275 

4.526 

2.659 

5,275 

3.935      3.376 

1.984 

2.659 

1.984 

1.702 

1.000 

30 

5.693 

4.392 

3,833 

2.386 

4.392 

3.389       2.957 

1.841 

2.386 

1.841 

1.606 

1.000 

40 

4.470 

3.579 

3.183 

2.114 

3.579 

2.866       2.549 

1.693 

2.114 

1.693 

1.506 

1.000 

60 

3.372 

2.817 

2.562 

1.836 

2.817 

2.354      2.141 

1.534 

1.836 

1.534 

1.396 

1.000 

120 

2.350 

2.072 

1.939 

1.533 

2.072 

1.828       1.710 

1.352 

1.533 

1.352 

1.265 

1.000 

00 

1,000 

1.000 

1.000 

1,000 

1.000 

1.000       1.000 

1.000 

1.000 

1.000 

1.000 

1.000 

1  Reproduced  from  Table  H  of  Owen  L.  Davies,  The  Design  and  Analysis  of 
Industrial  Experiments,  second  ed.,  Oliver  and  Boyd,  Edinburgh,  1956.  By  per 
mission  of  the  author  and  publishers. 

2  The  entries  in  this  table  show  the  value  of  the  ratio  R  of  two  population 
variances  vl/<r\  which  will  be  undetected  with  probability  /3  in  a  variance  ratio 
test  at  the  lOOce  per  cent  significance  level  of  the  ratio  si/ 'si  of  estimates  of  the 
two  variances,  both  being  based  on  v  degrees  of  freedom. 


[555] 


CRITICAL  VALUES  OF  r  FOR  THE 
SIGN  TEST2 


n 

1% 

5% 

10% 

25% 

n 

1% 

5% 

10% 

25% 

1 

46 

13 

15 

16 

18 

2 

47 

14 

16 

17 

19 

3 

O 

48 

14 

16 

17 

19 

4 

0 

49 

15 

17 

18 

19 

5 

0 

0 

50 

15 

17 

18 

20 

6 

0 

0 

1 

51 

15 

18 

19 

20 

7 

O 

0 

1 

52 

16 

18 

19 

21 

8 

0 

0 

1 

1 

53 

16 

18 

20 

21 

9 

0 

1 

1 

2 

54 

17 

19 

20 

22 

10 

o 

1 

1 

2 

55 

17 

19 

2O 

22 

11 

0 

1 

2 

3 

56 

17 

20 

21 

23 

12 

1 

2 

2 

3 

57 

18 

20 

21 

23 

13 

1 

2 

3 

3 

58 

18 

21 

22 

24 

14 

1 

2 

3 

4 

59 

19 

21 

22 

24 

15 

2 

3 

3 

4 

60 

19 

21 

23 

25 

16 

2 

3 

4 

5 

61 

20 

22 

23 

25 

17 

2 

4 

4 

5 

62 

20 

22 

24 

25 

18 

3 

4 

5 

6 

63 

20 

23 

24 

26 

19 

3 

4 

5 

6 

64 

21 

23 

24 

26 

2O 

3 

5 

5 

6 

65 

21 

24 

25 

27 

21 

4 

5 

6 

7 

66 

22 

24 

25 

27 

22 

4 

5 

6 

7 

67 

22 

25 

26 

28 

23 

4 

6 

7 

8 

68 

22 

25 

26 

28 

24 

5 

6 

7 

8 

69 

23 

25 

27 

29 

25 

5 

7 

7 

9 

70 

23 

26 

27 

29 

26 

6 

7 

8 

9 

71 

24 

26 

28 

30 

27 

6 

7 

8 

10 

72 

24 

27 

28 

30 

28 

6 

8 

9 

10 

73 

25 

27 

28 

31 

29 

7 

8 

9 

10 

74 

25 

28 

29 

31 

30 

7 

9 

10 

11 

75 

25 

28 

29 

32 

31 

7 

9 

10 

11 

76 

26 

28 

30 

32 

32 

8 

9 

10 

12 

77 

26 

29 

30 

32 

33 

8 

10 

11 

12 

78 

27 

29 

31 

33 

34 

9 

10 

11 

13 

79 

27 

30 

31 

33 

35 

9 

11 

12 

13 

80 

28 

30 

32 

34 

36 

9 

11 

12 

14 

81 

28 

31 

32 

34 

37 

10 

12 

13 

14 

82 

28 

31 

33 

35 

38 

10 

12 

13 

14 

83 

29 

32 

33 

35 

39 

11 

12 

13 

15 

84 

29 

32 

33 

36 

40 

11 

13 

14 

15 

85 

30 

32 

34 

36 

41 

11 

13 

14 

16 

86 

30 

33 

34 

37 

42 

12 

14 

15 

16 

87 

31 

33 

35 

37 

43 

12 

14 

15 

17 

88 

31 

34 

35 

38 

44 

13 

15 

16 

17 

89 

31 

34 

36 

38 

45 

13 

15 

16 

18 

90 

32 

35 

36 

39 

* 

*  For  values  of  n  larger  than  90,  approximate  values  of  r  may  be  found  by  taking  the 
nearest  integer  less  than  (n— 1)/2— -k^/n-\-l,  where  k  is  1.2879,  0.980O,  0.8224,  0.5752 
for  the  1,  5,  10,  25%  values,  respectively. 

1  Reproduced  from  W.  J.  Dixon  and  F.  J.  Massey,  Jr.,  An  Introduction  to 
Statistical  Analysis,  McGraw-Hill  Book  Company,  Inc.,  New  York,  1951,  p.  324. 
By  permission  of  the  authors  and  publishers. 


APPENDIX    1  41 


TABLE  OF  CRITICAL  VALUES  OF  T  IN 
THE  WILCOXON  SIGNED  RANK  TEST2 


_evel  of  Significance  for 

One-Tailed  Test 

.025 

.01 

.005 

L,evel  of  Significance  for 

Two-Tailed  Test 

n 

.05 

.02 

.01 

6 

0 



— 

7 

2 

0 

— 

8 

4 

2 

0 

9 

6 

3 

2 

10 

8 

5 

3 

11 

11 

7 

5 

12 

14 

10 

7 

13 

17 

13 

10 

14 

21 

16 

13 

15 

25 

20 

16 

16 

30 

24 

20 

17 

35 

28 

23 

18 

40 

33 

28 

19 

46 

38 

32 

20 

52 

43 

38 

21 

59 

49 

43 

22 

66 

56 

49 

23 

73 

62 

55 

24 

81 

69 

61 

25 

89 

77 

68 

* 

*  For  n>25,  T  is  approximately  normally  distributed  with  mean  «(w+l)/4  and  vari 
ance  «(«+!)  (2n  +  l)/24. 

Adapted  from  Table  I  of  F.  Wilcoxon,  Some  Rapid  Approximate  Statistical 
,  American  Cyanamid  Co.,  Stanford,  Conn.,  1949,  p.  13.  By  permuwion 


TueSofen  in  the  table  are  critical  values  associated  -ith  selected 
values  of  n.  Any  value  of  T  which  is  less  than  or  equal  to  the  tabulated  value 
is  significant  at  the  indicated  level  of  significance. 

[5571 


APPENDIX    151 

TABLE  OF  CRITICAL  VALUES  OF  r  IN 
THE  RUN  TEST2 


TABLE  1 


2 

3 

4 

5 

6 

7 

8 

9 

10 

11 

12 

13 

14 

15 

16 

17 

18 

19 

20 

2 

2 

2 

2 

2 

2 

2 

2 

2 

2 

3 

2 

2 

2 

2 

2 

2 

2 

2 

2 

3 

3 

3 

3 

3 

3 

4 

2 

2 

2 

3 

3 

3 

3 

3 

3 

3 

3 

4 

4 

4 

4 

4 

5 

2 

2 

3 

3 

3 

3 

3 

4 

4 

4 

4 

4 

4 

4 

5 

5 

5 

6 

2 

2 

3 

3 

3 

3 

4 

4 

4 

4 

5 

5 

5 

5 

5 

5 

6 

6 

7 

2 

2 

3 

3 

3 

4 

4 

5 

5 

5 

5 

5 

6 

6 

6 

6 

6 

6 

8 

2 

3 

3 

3 

4 

4 

5 

5 

5 

6 

6 

6 

6 

6 

7 

7 

7 

7 

9 

2 

3 

3 

4 

4 

5 

5 

5 

6 

6 

6 

7 

7 

7 

7 

8 

8 

8 

10 

2 

3 

3 

4 

5 

5 

5 

6 

6 

7 

7 

7 

7 

8 

8 

8 

8 

9 

11 

2 

3 

4 

4 

5 

5 

6 

6 

7 

7 

7 

8 

8 

8 

9 

9 

9 

9 

12 

2 

2 

3 

4 

4 

5 

6 

6 

7 

7 

7 

8 

8 

8 

9 

9 

9 

10 

10 

13 

2 

2 

3 

4 

5 

5 

6 

6 

7 

7 

8 

8 

9 

9 

9 

10 

10 

10 

10 

14 

2 

2 

3 

4 

5 

5 

6 

7 

7 

8 

8 

9 

9 

9 

10 

10 

10 

11 

11 

15 

2 

3 

3 

4 

5 

6 

6 

7 

7 

8 

8 

9 

9 

10 

10 

11 

11 

11 

12 

16 

2 

3 

4 

4 

5 

6 

6 

7 

8 

8 

9 

9 

10 

10 

11 

11 

11 

12 

12 

17 

2 

3 

4 

4 

5 

6 

7 

7 

8 

9 

9 

10 

10 

11 

11 

11 

12 

12 

13 

18 

2 

3 

4 

5 

5 

6 

7 

8 

8 

9 

9 

10 

10 

11 

11 

12 

12 

13 

13 

19 

2 

3 

4 

5 

6 

6 

7 

8 

8 

9 

10 

10 

11 

11 

12 

12 

13 

13 

13 

20 

2 

3 

4 

5 

6 

6 

7 

8 

9 

9 

10 

10 

11 

12 

12 

13 

13 

13 

14 

1  Adapted  from  Frieda  S.  Swed  and  C.  Eisenhart,  "Tables  for  testing  random 
ness  of  grouping  in  a  sequence  of  alternatives/7  Ann.  Math.  Stat.,  VoL  14,  1943, 
pp.  83-86.  By  permission  of  the  authors  and  publishers. 

2  The  values  of  r  given  in  Tables  1  and  2  are  various  critical  values  of  r  associ 
ated  with  selected  values  of  wi  and  n2.  For  the  one-sample  run  test,  any  value 
of  r  which  is  equal  to  or  less  than  the  value  shown  in  Table  1  or  equal  to  or 
greater  than  the  value  shown  in  Table  2  is  significant  at  the  5  per  cent  level. 
J*!or  the  two-sample  run  test,  any  value  of  r  which  is  equal  to  or  less  than  the 
value  shown  in  Table  1  is  significant  at  the  5  per  cent  level. 


[558: 


TABLE  2 


23456 

7 

8 

9 

10 

11 

12 

13 

14 

15 

16 

17 

18 

19 

20 

2 

3 

4 

9   9 

5 

9  10  10 

11 

11 

6 

9  10  11 

12 

12 

13 

13 

13 

13 

7 

11  12 

13 

13 

14 

14 

14 

14 

15 

15 

15 

8 

11  12 

13 

14 

14 

15 

15 

16 

16 

16 

16 

17 

17 

17 

17 

17 

9 

13 

14 

14 

15 

16 

16 

16 

17 

17 

18 

18 

18 

18 

18 

18 

10 

13 

14 

15 

16 

16 

17 

17 

18 

18 

18 

19 

19 

19 

20 

20 

11 

13 

14 

15 

16 

17 

17 

18 

19 

19 

19 

20 

20 

20 

21 

21 

12 

13 

14 

16 

16 

17 

18 

19 

19 

20 

20 

21 

21 

21 

22 

22 

13 

15 

16 

17 

18 

19 

19 

20 

20 

21 

21 

22 

22 

23 

23 

14 

15 

16 

17 

18 

19 

20 

20 

21 

22 

22 

23 

23 

23 

24 

15 

15 

16 

18 

18 

19 

20 

21 

22 

22 

23 

23 

24 

24 

25 

16 

17 

18 

19 

20 

21 

21 

22 

23 

23 

24 

25 

25 

25 

17 

17 

18 

19 

20 

21 

22 

23 

23 

24 

25 

25 

26 

26 

18 

17 

18 

19 

20 

21 

22 

23 

24 

25 

25 

26 

26 

27 

19 

17 

18 

20 

21 

22 

23 

23 

24 

25 

26 

26 

27 

27 

20 

17 

18 

20 

21 

22 

23 

24 

25 

25 

26 

27 

27 

28 

[5591 


APPENDIX    161 

TABLE  OF  CRITICAL  VALUES  OF  D  IN 
THE  KOLMOGOROV-SMIRNOV 
GOODNESS  OF  FIT  TEST2 


Sample  Size 
(n) 

Level  of  Significance 

for  £>= 

Maximum  |  F(x)  —Sn(x 

01 

.20 

.15 

.10 

.05 

.01 

1 

.900 

.925 

.950 

.975 

.995 

2 

.684 

.726 

.776 

.842 

.929 

3 

.565 

.597 

.642 

.708 

.828 

4 

.494 

.525 

.564 

.624 

.733 

5 

.446 

.474 

.510 

.565 

.669 

6 

.410 

.436 

.470 

.521 

.618 

7 

.381 

.405 

.438 

.486 

.577 

8 

.358 

.381 

.411 

.457 

.543 

9 

.339 

.360 

.388 

.432 

,514 

10 

.322 

.342 

.368 

.410 

.490 

11 

.307 

.326 

.352 

.391 

.468 

12 

.295 

.313 

.338 

.375 

.450 

13 

.284 

.302 

.325 

.361 

.433 

14 

.274 

.292 

.314 

.349 

.418 

15 

.266 

.283 

.304 

.338 

.404 

16 

.258 

.274 

.295 

.328 

.392 

17 

.250 

.266 

.286 

.318 

.381 

18 

.244 

.259 

.278 

.309 

.371 

19 

.237 

.252 

.272 

.301 

.363 

20 

.231 

.246 

.264 

.294 

.356 

25 

.21 

.22 

.24 

.27 

.32 

30 

.19 

.20 

.22 

.24 

.29 

35 

.18 

.19 

.21 

.23 

.27 

Over  35 

1.07 

1.14 

1.22 

1.36 

1.63 

V^                 VZ                VZ                 Vi                v^ 

1  Adapted  from  F.  J.  Massey,  Jr.,  "The  Kolmogorov-Smirnov  test  for  goodness 
of  fit/'  Jour.  Amer.  Stat.  Assn.,  Vol.  46,  1951,  pp.  68-78.  By  permission  of  the 
author  and  publisher. 

2  The  values  of  D  given  in  the  table  are  critical  values  associated  with  selected 
values  of  n.  Any  value  of  D  which  is  greater  than  or  equal  to  the  tabulated  value 
is  significant  at  the  indicated  level  of  significance. 


[560] 


APPENDIX    1?i 

PERCENTAGE  POINTS  OF  PSEUDO  f 
AND  f  STATISTICS 

TABLE  1-Table  for  Testing  the  Significance  of  the  Deviation  of  the  Mean 
of  a  Small  Sample  (of  Size  «)  From  Some  Pre-Assigned  Value 


n 

P 

0.95 

0.975 

0.99 

0,995 

0.999 

0.9995 

2 

3.196 

6.353 

15.910 

31.828 

159.16 

318-31 

3 

0.885  — 

1.304 

2.111 

3.008 

6.77 

9.58 

4 

.529 

0.717 

1.023 

1.316 

2.29 

2.85+ 

5 

.388 

.507 

0.685+ 

0.843 

1.32 

1,58 

6 

0.312 

0.399 

0.523 

0.628 

0.92 

1.07 

7 

.263 

.333 

.429 

.507 

.71 

0.82 

8 

.230 

.288 

.366 

.429 

.59 

.67 

9 

.205  — 

.255+ 

.322 

.374 

,50 

.57 

10 

.186 

.230 

.288 

.333 

.44 

.50 

11 

0.170 

0.210 

0.262 

0.302 

0.40 

0.44 

12 

.158 

.194 

.241 

.277 

.36 

.40 

13 

.147 

.181 

.224 

.256 

.33 

,37 

14 

.138 

.170 

.209 

.239 

.31 

.34 

15 

.131 

.160 

.197 

.224 

.29 

,32 

16 

0.124 

0.151 

0.186 

0.212 

0.27 

0.30 

17 

.118 

.144 

.177 

.201 

.26 

-28 

18 

.113 

.137 

.168 

.191 

.24 

.26 

19 

,108 

.131 

.161 

.182 

.23 

.25+ 

20 

.104 

.126 

.154 

.175- 

.22 

.24 

P  {TJ  <  value  in  table}  =  p 

1  Table  1  is  reproduced  with  the  permission  of  Professor  E.  S.  Pearson  from 
E.  Lord,  "The  Use  of  the  Range  in  Place  of  the  Standard  Deviation  in  the  t 
Test/'  Biometrika,  XXXIV  (1947),  66. 

Table  2  is  reproduced  with  the  permission  of  Professor  E.  S.  Pearson  from 
E.  Lord,  "The  Use  of  the  Range  in  Place  of  the  Standard  Deviation  in  the  t 
Test,7'  Biometrika,  XXXIV  (1947),  66. 

Table  3  is  reproduced  from  J.  E.  Walsh,  "On  the  Range-Midrange  Test  and 
Some  Tests  With  Bounded  Significance  Levels/'  Ann.  Math.  Stat.,  XX  (1949), 
257.  By  permission  of  the  author  and  publishers. 

Table  4  is  reproduced  from  R.  F.  Link,  "On  the  Ratio  of  Two  Ranges/7  Ann. 
Math.  Stat.,  XXI  (1950),  112.  By  permission  of  the  author  and  publishers. 


15611 


TABLE  2-Table  for  Testing  the  Significance  of  the  Difference  Between  the 
Means  of  Two  Small  Samples  of  Equal  Size  n 


n 

P 

0.95 

0.975 

0.99 

0.995 

0.999 

0.9995 

2 

1.161 

1.713 

2.776 

3.958 

8.90 

12.62 

3 

.487 

.636 

.857 

1.046 

1.63 

2.09 

4 

.322 

.406 

.523 

.618 

.87 

.99 

5 

.246 

.306 

.386 

.448 

.60 

.67 

6 

.202 

.249 

.310 

.357 

.47 

.51 

7 

.173 

.213 

.262 

.300 

.38 

.42 

8 

.153 

.186 

.229 

.260 

.33 

.36 

9 

.137 

.167 

.204 

.232 

.29 

.32 

10 

.125 

.152 

.185 

.209 

.26 

.29 

11 

.116 

.140 

.170 

.192 

.24 

.26 

12 

.107 

.130 

.157 

.177 

.22 

.24 

13 

.100 

.121 

.147 

.165 

.20 

.22 

14 

.094 

.114 

.138 

.155 

.19 

.21 

15 

.089 

.108 

.130 

.146 

.18 

.19 

16 

.085 

.102 

.123 

.139 

.17 

.18 

17 

.081 

.097 

.118 

.132 

.16 

.17 

18 

.077 

.093 

.112 

.126 

.15 

.17 

19 

.074 

.089 

,108 

.121 

.15 

.16 

20 

.071 

.086 

.103 

.116 

.14 

.15 

<  value  in  table}  = 


C5621 


TABLE   3-Cumulative  Percentage  Points  for 


Sample  Size 

P      •                     j>      .99             p  —  .995 

2 

3.16 

6.35 

15.91 

31.83 

3 

.90 

1.30 

2.11 

3.  02 

4 

.55 

.74 

1.04 

1.37 

5 

.42 

.52 

.71 

.85 

6 

.35 

.43 

.56 

.66 

7 

.30 

.37 

.47 

.55 

8 

.26 

.33 

.42 

.47 

9 

.24 

.30 

.38 

.42 

10 

.22 

.27 

.35 

.39 

p  —  P  {7-3  <  value  In  the  table} 
TABLE  4-Substitute  F-Ratio,  Cumulative  Percentage  Points 


Sample 

Sample  Size  for  Numerator 

Size  for 

Denominator 

2 

3 

4 

5 

6 

7 

8 

9 

10 

2 

12.7 

19.1 

25 

28 

29 

31 

32 

34 

36 

3 

3.19 

4.4 

5.0 

5.7 

6.2 

6.6 

6.9 

7.2 

7.4 

4 

2,02 

2.7 

3.1 

3.4 

3.6 

3.8 

4.0 

4.2 

4,4 

5 

1.61 

2.1 

2,4 

2.6 

2.8 

2.9 

3.0 

3.1 

3,2 

6 

1.36 

1.8 

2,0 

2.2 

2.3 

2.4 

2.5 

2.6 

2.7 

7 

1.26 

1.6 

1.8 

1.9 

2.0 

2.1 

2.2 

2.3 

2.4 

8 

1.17 

1.4 

1.6 

1.8 

1.9 

1.9 

2.0 

2.1 

2,1 

9 

1.10 

1.3 

1.5 

1.6 

1.7 

1,8 

1.9 

1.9 

2.0 

10 

1.05 

1.3 

1.4 

1.5 

1.6 

1.7 

1.8 

1.8 

1.9 

p=.975 

Sample 

Sample  Size  for  Numerator 

Denominator 

2 

3 

4 

5 

6 

7 

8 

9 

10 

2 

25.5 

38.2 

52 

57 

60 

62 

64 

67 

68 

3 

4.61 

6.3 

7,3 

8.0 

8.7 

9.3 

9.8 

10.2 

10.5 

4 

2.72 

3.5 

4.0 

4.4 

4.7 

5.0 

5.2 

5.4 

5.6 

5 

2.01 

2.6 

2.9 

3.2 

3.4 

3.6 

3.7 

3.8 

3.9 

6 

1.67 

2.1 

2.4 

2.6 

2,8 

2.9 

3.0 

3.1 

3.2 

7 

1.48 

1.9 

2.1 

2.3 

2.4 

2.5 

2.6 

2.7 

2.8 

8 

1.36 

1.7 

1.9 

2.0 

2.2 

2.3 

2.3 

2.4 

2.5 

9 

1.27 

1.6 

1.8 

1.9 

2.0 

2.1 

2.1 

2.2 

2.3 

10 

1.21 

1.5 

1.6 

1.8 

1.9 

1.9 

2.0 

2.0 

2.1 

[5631 


=  .99 


Sample 

Sample  Size  for  Numerator 

Size  for 
Denominator 

2 

3 

4 

5 

6 

7 

8 

9 

10 

2 

63.7 

95 

125 

140          150 

150 

160 

160 

160 

3 

7.37 

10 

12 

13 

14 

15 

15 

16 

17 

4 

3.83 

5.0 

5.5 

6.0 

6.4 

6.7 

7.0 

7.2 

7.5 

5 

2.64 

3.4 

3.8 

4.1 

4.3 

4.6 

4.7 

4.9 

5.0 

6 

2.16 

2.7 

3.0 

3.2 

3.4 

3.6 

3.7 

3.8 

3.9 

7 

1,87 

2.3 

2.6 

2.8 

2.9 

3.0 

3.1 

3.2 

3.3 

8 

1.69 

2.1 

2.3 

2,4 

2.6 

2.7 

2.8 

2.8 

2.9 

9 

1.56 

1.9 

2.1 

2.2 

2.3 

2.4 

2.5 

2.6 

2.6 

10 

1.47 

1.8 

1.9 

2,1 

2.2 

2.2 

2.3 

2,4 

2.4 

p=.995 

Sample 

Sample  Size  for  Numerator 

Size  for 

Denominator 

2 

3 

4 

5 

6 

7 

8 

9 

10 

2 

127 

191 

230 

250 

260 

270 

280 

290 

290 

3 

10.4 

14 

17 

18 

20 

21 

22 

23 

25 

4 

4.85 

6.1 

7.0 

7.6 

8.1 

8.5 

8.8 

9.3 

9.6 

5 

3.36 

4.1 

4.6 

4.9 

5.2 

5.5 

5.7 

5.9 

6.1 

6 

2.67 

3.1 

3.5 

3.8 

4.0 

4.1 

4.3 

4.5 

4.6 

7 

2.28 

2.7 

2.9 

3.1 

3.3 

3.5 

3.6 

3.7 

3.8 

8 

2.03 

2.3 

2.6 

2.7 

2.9 

3.0 

3.1 

3.2 

3.3 

9 

1.87 

2.1 

2.4 

2.5 

2.6 

2.7 

2.8 

2.9 

3.0 

10 

1.75 

2.0 

2.2 

2.3 

2.4 

2.5 

2.6 

2.6 

2.7 

P{Ri/R*< value  in  the  table} 


[564] 


INDEX 


A,  alternative  hypothesis,  108 

a,  probability  of  type  I  error,  107 

Abbreviated  Doolittle  method,  177 

Absolute  value,  18 

Acceptable  quality  level  (AQL),  486 

Acceptance 

number,  485 

region,  108 

sampling,  477,  485 
Accuracy,  104 
Ackoff,  R.  L.,  15 
Additive  law  of  probability,  3 1 
Additive  property 

chi-square,  82 

normal,  80 

Additivity,  test  for,  338 
Adjusted  value  of 

chi-square,  117 

Yf  in  regression,  199 
Alger,  P.  L.,  8,  9,  14 
Alternative  hypothesis  (A),  108 
American     Management    Association, 

14,  15 

American    Society  for    Testing    Mate 
rials,  498,  548 
American  Standards  Association,  Inc., 

498 
Analysis  of  covariance,  200,  437 

adjusted  means,  442,  446,  450,  453 

assumptions,  438 

completely  randomized  design,  439 

factorial,  452 

Latin  square  design,  449 

multiple,  457 

randomized  complete  block  design, 
444 

uses  of,  437 

variance   of    adjusted    means,    442, 

446,  450,  453 
Analysis    of    enumeration    data    (see 

Enumeration  data) 
Analysis  of  regression,  159 
Analysis  of  variance 

among  means,  133 

among  and  within  groups,  279 

assumptions,  281 

between  and  within  groups,  279 

comparing  individual  means,  303 

completely  randomized  design,  278 


component  of  variance   model,  318, 

324 

computational  procedure,  279 
degrees  of  freedom,  281 
disproportionate  subclass  numbers, 

423 

efficiency,  298 
factorials,  316 

fixed  effects  model,  282,  318,  324 
Graeco-Latin  square  design,  410,  414 
group  comparison,  279 
homogeneity  of  variances,   test  for, 

136 

hypotheses,  284 
Incomplete  block  design,  417 
individual  comparisons,  303 
individual  degrees  of  freedom,  303 
in  regression,  166 
Latin  square  design,  412 
mixed  model  319,  325 
Model  I,  fixed  treatment  effects,  282, 

318,  324 
Model  II,  random  treatment  effects, 

282,  318,  324 

Model  III,  mixed  model,  319,  325 
nonconformity    to    assumed    model, 

338 

N-way  classification,  335 
one-way  classification,  279 
proportionate  subclass  numbers,  421 
random  effects  model,  282,  318,  324 
randomized  complete  block  design, 

363 

in  regression,  166 

relation  to  regression  analysis,  340 
relative  efficiency,  300,  373 
relative  information,  300 
selected  treatment  comparisons,  303, 

376 

split-plot  design,  415 
subsampli;ng,  288,  335,  368 
table,  280 

three-way  classification,  321 
two-way  classification,  316 
transformations,  340 
Andersen,  S.  L.,  338,  361 
Anderson,  H.  E.,  434 
Anderson,  J.  A.,  15 


[565] 


566 


INDEX 


Anderson,  R.  L,.,  43,  105,  157,  194,  221, 

313,  360,  408,  434,  465 
ANOVA  (AnalyMs  of  variance),  278 
Anscombe,  F.  JL,  275,  338,  360,  434 
AOQ,  average  outgoing  quality,  487- 
AOQL,  average  outgoing  quality  limit, 

488 

AOV  (Analysis  of  variance),  278 
Approximations 

binomial  to  hypergeo metric,  74 
normal  to  binomial,  76,  116 
Poisson  to  binomial,  75 
AQL,,  acceptable  quality  level,  486 
Arbitrary  origin,  54 
Arcsine  transformation,  340 
Arithmetic  mean,  53 
Arnofi%  E>  L.,  15 
Array,  47 

ASN,  average  sample  number,  141,  489 
Association,  measures  of,  222 
Asterisk 

*Significant  at  <x=«0.05,  343 
**Significant  at  a  =  0.01,  343 
Attribute  data  (see  Enumeration  data) 
Attributes 

acceptance  sampling  by,  485 
double  sampling,  485,  488 
multiple  sampling,  485,  490 
sampling  inspection  by,  485 
sequential  sampling,  485,  490 
single  sampling,  485 
Average    (see    Mean,    Median,    Mode, 

etc.),  47 

Average  outgoing  quality  (AOQ),  487 
Average        outgoing        quality        limit 

(AOQL,),  488 

Average  sample  number    (ASN),    141, 
489 


bf  number  of  blocks,  364 

6,  sample  regression  coefficient,  164 

£,  population  regression  coefficient,  164 

£,  probability  of  type  II  error,  107 

Balancing,  251,  268 

Bancroft,  T.  A.,  43,  105,  157,  408,  434, 
465 

Barbacki,  S,,  275 

Bard,  J.  C.,  9,  15 

Bartlett,  M.  S.,  157,  338,  340,  361,  457, 
465 

Bartlett's  test  for  equality  (homogene 
ity)  of  variances,  136,  338 

Bayes3  theorem,  32 

Bazovsky,  I.,  498,  510 

Beatty,  G.  H.,  100,  106 

Bechhofer,  R.  EL,  31O,  361 

Bernhard,  F.  L.,  10,  15 


Bernoulli  trials,  32 
Beta 

distribution,  39 

function,  22 
Beveridge,  W.  L  B.,  15 
Bias,  104 

Bicking,  C.  A.,  265,  266,  271,  275,  434 
Bimodal  distribution,  59 
Bingham,  R.  S.,  Jr.,  275 
Binomial 

approximation    to     hyper  geometric, 
74 

chi-square  approximate  test,  117 

confidence  intervals,  94 

distribution,  38,  74 

estimation,  94 

expansion,  21 

expected  value,  38 

mean,  38,  74 

normal  approximation,  76,  116 

Poisson  approximation,  75 

population,  74 

probabilities,  32 

probability  function,  38,  74 

standard  deviation,  38,  74 

tests  of  hypotheses,  115,  128 

variance,  38,  74 
Birnbaum,  Z.  W.,  474 
Biserial  correlation,  231 
Bivariate  normal  population,  198,  225 
Blocking,  251,  268,  270 
Blocks 

incomplete,  417 

randomized  complete,  268,  363 
Blum,  J.  R.,  475 

Bowker,  A.  H.,  100,  105,  113,  114,  115, 
121,   123,   124,   157,  221,  243,  498, 
504,  510 
Box,  G.  E.  P.,  277,  338,  361,  434,  435, 

436,  502,  510 

Breipohl,  A.  M.,  503,  504,  510 
Bross,  I.  D.  J.,  275 
Brown,  G.  W.,  473,  475 
Brownlee,  K.  A.,  43,  105,  157,  221,  243, 

275,  361,  408,  435,  475 
Brunk,  H.  D.,  475 
Bryan,  J.  G.,  43,  106,  158 
Budne,  T.  A.,  275,  435 
Bur  man,  J".  P.,  436 
Buros,  O.  K.,  15 
Burr,  I.  W.,  75,  86,  498 
Bush,  G.  P.,  15 


C,  coefficient  of  contingency,  232 
c-chart 

center  line,  480 

control  limits,  480 


INDEX 


567 


example  of,  483 
Calabro,  S.  R.,  510 
Calvert,  R.  L.,  94,  105 
Caplan,  F.,  276 
Carroll,  M.  B.,  435 
Causes,  chance,  477 
Center  line  (for  control  charts),  478 
Central 

limit  theorem,  72 
moments  (definition),  36 
tendency  (measures  of),  53 
Chance  variable,  30 
Chapanis,  A.  R.  E.,  15 
Chapin,  F.  S.,  276,  435 
Chew,  V.,  221,  276,  341,  361,  408,  415, 

435 
Chi-square 

addition  theorem,  82 

adjusted  value,  117 

definition,  81 

degrees  of  freedom,  81 

density  function,  40,  82 

distribution,  37,  40,  81 

expected  value,  40 

heterogeneity,  128 

interaction,  129 

mean,  40 

parameter  of,  81 

pooled,  128 

reproductive  property,  82,  128 

small  expected  numbers,  127 

t^able,  523 

tests, 

Bartlett's  test,  136 

binomial,  117 

contingency  tables,  129 

for  independence,  129 

goodness  of  fit,  126,  466 

homogeneity  of  variances,  136 

multinomial,  124 

Poisson  data,  125 

rXc  table,  129 

variance    of    normal    population, 

114 

total,  128 
variance,  40 

Chorafas,  E>.  N.,  498,  510 
Churchman,  C.  W.,  15 
Clark,  C.  R,,  486,  488,  498 
Class 

frequency,  49 
interval,  47 

end  points,  47 
midpoint,  47 
Classification 
AT- way,  335 
one-way,  279 
three-way,  321 


two-way,  316 
Clopper,  C.  J.,  94,  105 
Cochran,   W.    G,,  261,   276,   303,   338, 

361,  374,  408,  435,  457,  465,  475 
Coding,  54,  62,  239 
Coefficient 

contingency,  232 

correlation,  222 

regression,  164 
Coefficient  of 

alienation,  225 

biserial  correlation,  231 

concordance,  233 

contingency,  232 

correlation,  222 

determination,  225 

linear  correlation,  224 

multiple  correlation,  228 

non-determination,  225 

partial  correlation,  228 

partial  regression,  229 

product-moment  correlation,  225 

tetrachoric  correlation,  232 

variation  (CF),  64 
Coefficients 

for  orthogonal  polynomials,  194,  314 

for  selected  treatment  comparisons, 

262 

Cohen,  M.  R.,  15 
Colcord,  C,  C.,  529 
Combinations 

of  n  things  taken  r  at  a  time,  20 

treatment,  252 

Combining  experimental  results,  128 
Comparison,  261 
Comparison  of 

means,  262,  310 

percentages,  131 

proportions,  131 

standard  deviations,  123 

variances,  123 
Comparisons 

among  means,  262,  306,  310 

designed,  262,  306 

group 

equal  size,  280 
unequal  size,  279 

individual,  262,  306,  310 

orthogonal,  263,  306,  376 

selected,  262,  306 
Complement,  17 
Completely  randomized  design,  268 

analysis  of  co variance,  439 

analysis  of  variance,  280 

assumptions,  281 

calculations,  279 

comparison   of  selected  treatments, 
306 


568 


INDEX 


Completely  randomized  design  (cont'd) 
components  of  variance,  285,  298 
definition  of,  278 
degrees  of  freedom,  281 
examples  of,  278 
expected  mean  squares,  282,  298 
factorials,  316 

individual  degrees  of  freedom,  304 
relative  efficiency,  300 
standard  error  of  a  treatment  mean, 

285 

subsampling,  288,  335 
*-test,  288 
table,  280 

tests  of  hypotheses,  284,  327 
variance  of  a  treatment  mean,  285 
Component  of  variance 
definition  of,  237 
estimate  of,  299 

Component  of  variance  model,  282 
Co  nuponents  of  variance,  299 
Composite  design,  424 
Concomitant 

information,  159 
variable,  159 

Concordance,  coefficient  of,  233 
Conditional 

distribution,  35 
probability,  31 
Confidence 
coefficient,  90 

interval,  definition  of,  88,  89 
interval  statement,  91 
limits 

definition,  90 
one-sided,  90 
two-sided,  90 
limits  for 

binomial  parameter,  94 

contrast,  309 

correlation  coefficient,  226 

difference  between  two  means,  95 

fraction,  94 

intercept  of  straight  line,  171 

mean,  univariate,  90,  92 

mean,  in  regression,  171 

percentage,  94 

proportion,  94 

ratio  of  two  standard  deviations, 

97 

ratio  of  two  variances,  97 
regression  coefficient,  170 
slope  of  straight  line,  170 
standard  deviation,  93 
variance,  93 
Confounded,  246 
Confounding,  248 
Connor,  W.  S.,  277,  435 


Consistent  estimator,  87 
Consumer's  risk,  487 
Contingency 

coefficient  of,  232 
tables,  129 

2X2,  or  fourfold,  131,  132 
rXc,  129 
JV-way,  131 
Continuity,     correction    for,    77,     116, 

131 
Continuous 

distribution,  33 
distribution  function,  33 
probability  density  function,  33 
random  variable,  30 
variable,  30 
Contrast,  261,  306,  376 
Control 

lack  of,  478 
local,  246,  250 
quality,  477 
state  of,  478 
statistical,  478 
Control  charts 
center  line,  480 
definition  of,  477 
factors  for  computing  limits,  549 
interpretation  of,  479 
limits,  definition  of,  480 
types 

c-chart    (for    number   of    defects), 

479,  483 
2>-chart     (for    fraction    defective), 

479,  481 

JE-chart  (for  ranges  or  variability), 
_      479,  481 

X-chart  (for  averages),  479,  481 
Coons,  Irma,  465 

Correction  for  continuity,  77,  116,  131 
Correlated  observations,  96,   121,   159, 

222 

Correlation 
analysis,  222 

between  mean  and  variance,  340 
biserial,  231 
coefficient,  222 
definition,  36 
homotypic,  235 
index,  223 
intraclass,  235 
methods,  159,  222 
multiple,  227 
partial,  228 
product-moment,  36 
rank,  233 
ratio,  229 

sample  coefficient,  224 
several  samples,  227 


INDEX 


569 


simple,  223 
spurious,  236 
sums  and  differences,  238 
tetrachoric,  232 
Coutie,  G.  A.,  434 
Co  variance 

analysis  of,  200,  437 
definition  of,  36 

Cox,  D.  R.,  261,  273,  274,  276,  435,  465 
Cox,  G.  M.,  15,  261,  276,  361,  408,  435, 

457,  465 

Cramer,  H.,  28,  30,  43 
Crampton,  E.  W.,  457,  458,  465 
Critical  region,  109 
Crow,  E.  L.,  498 
Crump,  S,  L.,  276 
Cumulative  distribution,  33,  52 
Cumulative  distribution  function 
continuous,  33 
definition  of,  33 
discrete,  33 
properties,  33 
Cumulative  frequency 
distribution,  49,  52 
table,  52 

Cumulative  relative  frequency,  52 
Curve  fitting,  160 
Curvilinear  regression,  190 
CV,  coefficient  of  variation,  64 


D,  difference  between  paired  observa 
tions,  96,  121  m 
D,   statistic  in  the    Kolmogorov-Smir- 

nov  test,  471 

table  of  critical  values,  560 
Daniel,  C.,  435 
Data,  missing,  390,  412 
David,  H.  A.,  338,  361 
Davies,  G.  R.,  210 

Davies,  O.  L.,  157,  167,  221,  261,  276, 
361,  408,  418,  419,  425,  435,  550, 
552,  554,  555 
Davis,  F.  A.,  498 
DeBaun,  R.  M.,  435 
Decile  limits,  57 
Decisions 

correct,  108 
incorrect,  108 
Defect,  483 
Defective 
item,  481 

proportion,  94,  115,  481 
Defects,  control  chart  for  (see  c-chart), 

483 
d.f.,    degrees   of  freedom,   61,   81,   281, 

329 


Degrees  of  freedom  (d.f.),  61,  81,  281, 

329 

DeLury,  D.  B.,  276 
Deming,  L.  S.,  529 
Density  function,  33 
chi-square,  40 
continuous  probability,  33 
exponential,  39 
F,  40 

gamma,  39 
normal,  39 

probability  (p.d.f.),  33 
relation  to  cumulative  function,  33 
standard  normal,  39 
t,  40 

Weibull,  39 

Dependent  variable,  159 
Derman,  C.,  43 
Design 

completely  randomized,  268,  278 
composite,  424 
experimental,  244 
advantages,  271 

check  list  of  pertinent  points,  265 
disadvantages,  271 
examples  of  approach  to,  266 
nature  and  value  of,  244 
principles  of,  246 
steps  in,  264 
of  experiments,  244 
Graeco-Latin  square,  410 
incomplete  block,  417 
Latin  square,  410 
random  balance,  425 
randomized     complete     block,     268, 

363 

split  plot,  271,  415 
Descriptive  statistics,  44 
Deter  minant 

calculations,  24 
definition,  23 
Deviation 

from  mean,  61 
from  regression,  162 
standard,  36 

Dichotomous  data  (see  Binomial) 
Dick,  I.  D.,  339,  362 
Difference  between 

paired  observations  (Z>),  96,  121 
two  means,  81,  95,  119,  288 
two  proportions,  132 
Differences  u 

among  correlation  coefficients,  226 
among  means,  133,  279 
between  means,  81,  95,  119,  288 
among  variances,  136 
Differential  response,  258 


57O 


INDEX 


Digits 

random,  46 

table  of,  544 
Discrete 

distribution  function,  33 

probability  function,  33 

random  variable,  30 
Dispersion,  measures  of,  60 
Disproportionate     subclass     numbers. 

423 
Distribution 

beta,  39 

binomial,  38,  74 

bivariate,  34 

bivariate  normal,  198,  225 

chi-square,  37,  40,  81 

conditional,  35 

continuous,  33 

cumulative,  33 

difference  between  two  means,  81 

discrete,  33 

empirical,  49 

expected  value  of,  35 

exponential,  39 

F,  37,  40,  84 

frequency,  47 

function,  33 

gamma,  39 

Gaussian,  39 

geometric,  38,  79 

hypergeometric,  38,  73 

joint,  34 

of    linear    combinations    of    random 
variables,  80 

marginal,  34 

of  mean,  80 

multinomial,  78 

negative  binomial,  38,  79 

normal,  39 

Poisson,  37,  38 

probability,  33 

of  s2  and  s,  83 

sampling,  70 

of  standard  deviation,  83 

standard  normal,  37,  39 

"Student's"  t,  37,  40,  83 

of  sum  of  squares,  82 

t,  37,  40,  83 

of  variance,  83 

Weibull,  39 
Distribution-free 

methods,  466 

tests,  466 

tolerance  limits,  100 
Dixon,  W.  J.,   1O5,   157,  221,  243,  338, 

361,  408,  475,  529,  556 
Dodge,  H.  F.,  498 
Doolittle  solution,  abbreviated,  177 


Double  sampling,  485,  488 

advantages  of,  489 

nature  of,  488 
Duffett,  J.  R.,  276 
Dummer,  G.  W.  A.,  498,  510 
Dummy  treatments,  374 
Duncan,  A.  J.,  498 
Duncan,  D.  B.,  310,  361 
Dunnett,  C.  W.,  310,  361 
Dykstra,  O.,  Jr.,  435 


E*,  correlation  ratio,  229 

EVOP,  evolutionary  operation,  501 

c,  error  variable,  161 

Economy,  resource,  245 

Effect,  definition  of,  257 

Effects 

fixed,  282 
interaction,  257 
main,  257 
random,  282 
Efficiency 

relative,  300,  373 
statistical,  245 
Efficiency  of 

Latin  square,  414 
randomized  complete  block,  375 
range,  63 
Eisenhart,  C.,  100,  106,  158,  221,  338, 

361,  476,  499,  558 
Element,  17 

Empirical  distribution,  49 
English,  T.  S.,  147 
Enumeration  data 

acceptance  sampling  plans,  485 
binomial 

correction  for  continuity,  77,  116, 

131 

estimation,  94 
normal  approximation,  76 
Poisson  approximation,  75 
tests  of  hypotheses 
exact,  116 

normal  approximation,  116 
chi-square  approximation,   117 
control  charts  (c  and  p),  479 
double  dichotomy  (2X2  table) 
correction  for  continuity,  131 
independence,    approximate    test, 

131 

independence,  exact  test,  132 
goodness  of  fit  test,  126 
hypergeometric,  73 

binomial  approximation,  74 
independence  in  two-way  tables 
2X2,  131 
rXc,  129 


INDEX 


571 


multinomial,  78,  124 

chi-square  approximation,  124 
percentages  or  proportions,  74,  115 
comparison  of  two,  131 
comparison  of  k,  128 
proportions,  74,  115 

confidence  intervals,  94 
tests,  115,  128,  131 
Poisson,  75,  124 

tests,  124 
Equality 

of  means,  119,  133 
of  variances,  123,  136 
Equally  likely  events,  29 
Equations 

estimating,  163 
linear,  24,  164 
normal,  164 
predicting,  163 
regression,  163 

simultaneous,  24,  162,  164,  178 
Error 

off  estimate,  170 
experimental,  246,  247 
partitioning  of,  340,  376 
reduction  of,  248 
in  independent  variable,  198 
kind 

Type  I,  107 
Type  II,  107 
sampling,  291 
standard 

of  contrast,  309 
of  estimate,  170 
of  mean,  91,  285,  366 
of  regression  coefficient,  170 
Estimate 
biased,  87 
interval,  88 
minimum  variance,  87 
point,  87 
unbiased,  87 
Estimates 
best,  87 

maximum  likelihood,  89 
repeatability  of,  104 
Estimating  equation,  163 
Estimation,  87 

confidence  interval 

for    difference    between    means    of 

two  normal  populations,  95 
for  mean  of  a  normal  population, 

90 

for  a  proportion,  94 
for     slope     and     intercept     of     a 

straight  l£ne,  170 

for  standard  deviation  of  a.  normal 
population,  93 


for  variance  of  a  normal  popula 
tion,  93 
statement,  91 

point 

maximum  likelihood,  88 
method  of  least  squares,  88 
method  of  moments,  88 
minimum  chi-square,  88 
Estimator 

consistent,  87 

definition,  61 

interval,  88 

minimum  variance,  87 

point,  87 

properties,  87 

unbiased,  61,  70,  87 
Estimators 

maximum  likelihood,  89 

methods  of  obtaining,  88 
Event,  29 

Evolutionary  operation  (EVOP),  501 
EVOP  (evolutionary  operation),  501 
Expectation 

mathematical,  33 

of   mean   squares    in    ANOVA,    282, 

328 

Expected  frequency,  127 
Expected  mean  squares,  282,  328 
Expected  numbers,   117,  124,  127,  130 
Expected  value 

definition,  33,  35 

notation,  33,  35 
Expected  values  of  (or  for) 

beta  distribution,  39 

binomial  distribution,  38,  74 

chi-square  distribution,  40 

constant,  36 

constant  times   a  random  variable, 
36 

continuous  random  variable,  36 

discrete  random  variable,  36 

exponential  distribution,  39 

F- distribution,  40 

gamma  distribution,  39 

geometric  distribution,  38 

hypergeo metric  distribution,  38,  74 

linear  combinations  of  random  vari 
ables,  80 

negative  binomial  distribution,  38 

normal  distribution,  39 

Poisson  distribution,  38 

^-distribution,  40 
Weibull  distribution,  39 
Experiment 

definition  of,  30 

design  of,  244 
Experimental  design 

nature  of,  4,  244 


572 


INDEX 


Experimental  Design  (continued) 

principles  of,  246 

purpose  of,  245 

steps  in,  264 
Experimental 

error,  246,  247 

units,  247 
Experimentation,  4 
Exponential 

distribution,  39 

regression,  190,  194 
Extrapolation,  173 
Extreme  values,  84 
Ezekiel,  M.,  196,  221 


F-ratio,  less  than  unity,  301 
/'"-variate 

definition  of,  83 

degrees  of  freedom,  83 

distribution,  37,  40,  84 

expected  value,  40 

parameters  of,  83 

table  of,  529 

/,  observed  frequency,  5O 
Factor 

definition,  254 

level,  254 

symbolism,  254 
Factorial 

design  (see  Factorials) 

experiment  (see  Factorials) 

notation  (nl),  20 
Factorials 

advantages,  272 

analysis  of  covariance,  452 

calculations,  317,  322 

completely  randomized  design,  316 

component  of  variance  model,   318, 
324 

confounding,  420 

definition  of,  254 

degrees  of  freedom,  329 

disadvantages,  272 

effects,  257 

expected  mean  squares,  327 

fixed  effects  model,  318,  324 

fractional,  419 

interaction 

definition  of,  257 
effects,  257 

main  effects,  257 

meaning  of,  254 

mixed  model,  319,  325 

Model 

I  (fixed),  318,  324 

II  (random),  318,  324 

III  (mixed),  319,  325 


notation,  255 

response  curves,  336 

standard  error  of  effects,  309 

subsampling,  335 

terminology,  254 

tests  of  hypotheses,  327 

variance  of  effects,  309 

without  replication,  418 
Failure  rate,  509 
Fattu,  1ST.  A.,  475 
Federer,  W.  T.,  261,  276,  310,  361,  408, 

435,  465 

Feigenbaum,  A.  V.,  498,  510 
Festinger,  L,.,  475 

Fiducial  limits  (see  Confidence  limits) 
Finite  population  correction  factor,  71 
Finney,  D.  J.,  261,  276,  435,  465 
First  kind  of  error,  107 
Fisher,  R.   A.,  88,   105,   226,   243,  275, 

276,  435,  528 

Fisher's  exact  test  in  2X2  table,  132 
Fitting 

of  constants,  179 

of  curves,  160 
Flagle,  C.  D.,  15 
Fractile,  36 

Fractional  factorials,  419 
Fraction  defective 

control  chart  for,  481 

estimation  of,  94 

tests  of  hypotheses  about,  115 
Fraser,  D.  A.  S.,  475 
Freedman,  P.,  15 

Freedom,  degrees  of,  61,  81,  281,  329 
Freeman,  H.  A.,  499 
Frequency 

class,  47 

cumulative,  49 

distribution,  47 

histogram,  49 

observed  (/),  50 

polygon,  49 

relative,  48 

relative  cumulative,  52 

table,  50 

tally,  47 

Freund,  J.  E.,  105,  157 
Friedman,  M.,  499 
Friedman,  L.,  16 
Fry,  Thornton,  C.,  127,  472 
Function 

beta,  22 

cumulative  distribution,,  33 

density,  33 

distribution,  33 

gamma,  21 

likelihood,  73,  89 

OC,  108,  485 


operating  characteristic,  108    485 
power,   108 

probability  density,  33 
regression,  159 
response,  160 
Functional  relations,  159 

O 

Oamma 

distribution,  39 

function,  21 
Geffner,  J.,  16 
Generalized  interaction,  329 
Geometric  distribution.  38    79 
Gilbert,  S.,  276 
Gompertz  curve,  190 
Good,  CX  V.,  15 
Goode,  H,  H,,  15 
Goodness  of  fit  tests 

chi-square,  126 

in  regression  (lack  of  fit),  188 

Kolmogorov-Smirnov,  471 
Gosset,  W.  S.,  277 
Graeeo-Latin  squares,  410,  414 
Grandage,  A.  H.  E.,  188,  221 
Grant,^  E.  L_,  477,  484,  499 
Graphical  representation,  51    341 
Graybill,  F.  A.,  28,  221,  341,  361 
Greek  alphabet,  511 
Griffin,  1ST.,  498,  510 
Grouping,  251,  268 
Growth  curve  (see  Exponential) 
Gryna,  F.  M.,  Jr.,  15,  499,  51O 

H 

H.  hypothesis  (or  null  hypothesis).  1OS 

Hader,  H.  J.,  188,  221 

Hald,  A.,  43,  94,  105,  157,  523,  529 

Hamaker,  H.  CX,  339,  361 

Hartley,  H.  O.,  85,  86,  310,  361 

Hastay,  M.  W,,  1OO,  106,  158,  499 

Hattery,  L.  H,,  15 

Hawley,  G,  O.,  15 

Hay,  W.  A.,  435 

Hazard  rate,  43,  509 

Heterogeneity 

chi-square,  128 

of  variances,  339 
Heterogeneous  variances,  339 
Heteroschedasticity,  339 
Hill  way,  T.,  15 
Histogram,  frequency,  49 
Homogeneity  of  variances 

assumption  of,  95,  119,  133,  169,  197, 
282,  338 

test  for,  136 

Homogeneous  variances,  197 
Homoscedasticity,  197 


INDEX 

Homotypic  correlation,  235 
Hopkins,  J.  W.,  457,  458,  465 
Horton,  W.  H.,  435 


573 


Hunter    J.  S.,  188,  189,  221,  276,  277 
434,436,502,510  ' 

Huntsberger,  D.  V.,  43,  105,  158 
Mutt,   F.  B.,  6,   15 
Hypergeometric  distribution,  38    73 

binomial  approximation,  74 

use  in  acceptance  sampling    486 

use  in  Fisher's  exact  test,  132 
Hypotheses 

alternative,  108 

composite,  109 

null,  108 

simple,  109 

tests  of,  107 

I 

Identities,  20 

Incomplete  block  design,  417 

Independence 

in  a  two-way  table,  129 

mutual,  31 

pair  wise,  31 

statistical,  31 
Independent 

events,  31 

variables,  159 
Index  of  summation,  19 
Indifference  quality,  4B6 
Individual  comparisons,  262,  303    376 
Inference,  statistical,  87,  107 
Inferences  about  populations,  87,   107 
Information,  relative,  300 
Inspection 

by  attributes,  485 

by  variables,  485 

sampling,  485 

Interaction,  definitions  of,  257 
Intersection,  17 
Interval 

class,  47 

confidence,  88 

estimate,  88 
Intervals  for,  confidence 

difference  of  two  means,  95 

mean,  90 

proportion,  94 

ratio  of  two  variances,  97 

standard  deviation,  93 

variance,  93 

Intraclass  correlation,  235 
Inverse  of  a  matrix,  24 
Investigations,  experimental,  244 
Iterative  procedure,  393 


574 


INDEX 


Jeffreys,  H.,  15,  276 
Jevons,  W.  S.,  15 
Johnson,  P.  CX,  15 
Joint 

distribution.,  34 

probability,  31 

probability  function,  34 

probability  density  function,  34 
Juran,  J.  NL,  499 

K 

Kemeny,  J.  G.,  28,  43 

Kempthorne,    O.,   221,   261,    264,   266, 

267,  276,  277,  335,  341,  361,  408, 

436,  465 
Kendall,  M.  G.,  193,  221,  233,  234,  243, 

475 

Keuls,  M.,  310,  361 
Kimball,  G.  E.,  16 
Klein,  M.,  43 
Kolmogorov-Smirnov  test 
discussion,  338,  471 
table  of  critical  values,  560 
Koopmans,  L.  H.,  486,  488,  498 
Kramer,  C.  Y.,  221,  310,  361 
Kruskal,  W.  H.,  475 


X,  parameter  of  Poisson  distribution,  38 
Lack  of  fit,  188 

test  for  (in  regression),  188 
Latin  square,  410 
Latin  square  design 

analysis  of  covariance,  449 

analysis  of  variance,  412 

assumptions,  411 

calculations,  411 

comparison   of  selected   treatments, 
414 

components  of  variance,  414 

degrees  of  freedom,  412 

examples  of,  410 

expected  mean  squares,  412    • 

individual  degrees  of  freedom,  414 

missing  observations,  412 

relative  efficiency,  414 

response  curves,  414 

subsampling,  414 

test  of  hypothesis,  412 
Law  of  large  numbers,  72 
Laws   of  probability,   31 
Least  significant  difference  (LSD),  310 
Least  squares,  method  of,  88,  161 
Leone,  F.  C.,  276 
Level  of  factor,  254 
Level  of  significance,  108 


Lewis,  W.,  147 

Lieberman,    G.   J.,   86,    100,    105,    113, 
114,  115,  121,  123,   124,  157,  221, 
243,  486,  498,  499,  504,  510 
Likelihood  function,  73,  89 
Limits 

confidence  (see  Confidence  limits) 

control  (see  Control  limits),  480 

decile,  57 

natural  tolerance,  98,  502 

percentile,  57 

quartile,  57 

of  summation,  19 

specification,  502 

tolerance,  98,  502 
Linear 

combination  of  random  variables,  80 

correlation,  coefficient  of,  224 

equations,  24 

model,  164 

regression,  164 

relationship,  164 

statistical  model,  164 
Linearity 

assumption  of,  164 

of  regression,  164 

test  for,  188 
Link,  R.  P.,  510,  561 
Lipow,  M,,  499,  510 
Livermore,  P.  E.,  105,  157 
Lloyd,  D.  K.,  499,  510 
Local  control,  246,  250 
Location,  measure  of,  55 
Logarithmic  transformation,  340 
Logic,  3 

Logistic  curve,  190 
Lord,  E.,  510,  561 
Lot 

acceptance     procedures     based      on 
sampling  from,  485 

sampling  from,  485 

tolerance  percent  defective  (LTPD) 

487 

LSD,  least  significant  difference,  310 
LTPD,  lot  tolerance  percent  defective, 

487 

Lush,  J.  L.,  6,  15 
Luszki,  M.  E.  B.,  15 

M 

M,  sample  median,  55 
MO,  sample  mode,  58 
MR,  sample  midrange,  55 
At,  population  mean,  36 
Machol,  K.  E.,  15 
Main 

effect,  257 

plot,  415 


INDEX 


Mandelson,  J.,  271,  276 
Mann,  H.  B.,  475 
Marginal 

distribution,  34 

probability,  35 

probability  density  function,  35 
Massey,  F.  J.,  105,  157,  221,  243,  338, 

361,  408,  475,  529,  556,  560 
Mathematical 

concepts,  17 

expectation,  33 

model,  163 
Matrices 

addition  of,  23 

multiplication  of,  23 
Matrix 

definition  of,  22 

determinant  of,  23 

dimension  of,  22 

diagonal,  23 

identity,  23 

inverse  of,  24 

nonsingular,  24 

null,  23 

square,  23 

symmetric,  23 

transpose  of,  23 
Maxfield,  M.  W.,  498 
Maximum  likelihood,  88 
McAfee,  N.  J.,  15,  499,  510 
McNemar,  Q.,  231,  232,  233,  243 
Mean 

adjusted,  442,  446,  450 

arithmetic,  53 

confidence  interval  for,  91,  92 

confidence  limits  for,  91 

correction  for,  165 

definition  of,  53 

deviation  from,  53 

difference,  96,  121 

distribution  of,  80 

estimation  of,  70,  90 

of  probability  distribution   (see  Ex 
pected  value) 

population  (/*),  36 

sample  (jy),  53 

square,  134 

standard  error  of,  91 

tests  concerning,  113 

time  between  failure   (MTBF),   509 
Means 

differences  among  several,   133,   279 

difference  between  two,  95,  119,  288 
Measure 

of  location,  55 

of  position,  55 

of  precision,  60 
Median 


575 


definition,  36 

sample  (M),  55 

tests,  473 

Merrington,  M.,  529 
Method  of  least  squares,  88    161 
Midrange  (MR),  55 
Military  Standards 

105C,  499 

414,  499 

Miller,  D.  W.,  16 
Miller,  I.,  105,  157 
Minimum  variance  estimator,  87 
Missing  data  or  plot 

Latin  square,  412 

randomized  complete  block,  390 
Mitscherlich  curve,  190 
Mixed  model,  319,  325 
MO  (mode),  58 
Mode 

absolute,  58 

definition,  36 

relative,  58 

sample  (MO),  58 
Model 

analysis  of  variance,  282,  318,  324 

component  of  variance,  282,  318,  324 

for   completely   randomized    desien 
281  ' 

for  covariance 

completely      randomized      design 

438 

Latin  square  design,  438 
randomized    complete    block,    de 
sign,   438 

two-factor  factorial  in  an  R.CB  de 
sign,  438 

for  factorials,  316 

for  Latin  square,  411 

for  randomized  complete  block,  364 

for  regression,  163 

linear,  281 

mathematical,  163 

mixed,  319,  325 

statistical,  163 
Molina,  E.  C,,  512 
Moments 

definition,  36 

of  population,  36 

of  sample,  7O 
Mood,  A.  M.,  43,  78,  86,  105,  131,  158, 

276,  473,  475 
Morse,  P.  M.,  16 
Moses,  L>  E.,  475 
Mosteller,  F.,  43,  499 
MR  (midrange)  55 
MTBF    (mean   time   between   failure) 

509 
Muench,  J.  O.,  94,  105 


576 


INDEX 


Multino  mial 

distribution,  78 

population,  78 

tests  of  hypotheses,  124 
Multiple 

correlation,  227 

covariance,  457 

regression,  178 

sampling,  485,  490 
Multiplication 

of  matrices,  23 

of  probabilities,  31 

Multiplicative  law  of  probability,  31 
Murphy,  R.  B.,  101,  105 
Mutual  independence,  31 

N 

N,  size  of  lot  or  population  or  stratum, 

71,  73,  485 
N  (/*>  <0  >  normally  distributed  -with  mean 

fjt,  and  standard  deviation  o>  80 
nt  sample  size,  46 

no,  average  group  size  in  ANOVA,  283 
v,  degrees  of  freedom,  81 
v,  parameter  in  x2-<iiatribution,  81 
v,  parameter  in  ^-distribution,  83 
v,  parameter  in  ^-distribution,  83 
Nagel,  E.,  15 

National  Applied  Mathematics  Labo 
ratories,  245 
National  Bureau  of  Standards,  43,  74, 

86,  276,  436 
Natural  tolerance  limits,  98,  502 

relation  to  specification  limits,  502 
Near  stationary  region,  424 
Negative  binomial  distribution,  38,  79 
Negative  correlation,  224 
Newman,  D.,  310,  361 
Nominal  value,  502 
Noncentral  ^-distribution,  113 
Non-linear 

functions  of  random  variables,  505 
regression,  190 
Nonparametric 
methods,  466 
tests 

goodness-of-fit  test,  126 
Kolmogorov-Smirnov  test,  471 
median  tests,  473 
run  tests,  470 
sign  test,  466 
signed  rank  test,  468 
Wilcoxon  signed  rank  test,  468 
Normal  approximation 
to  binomial,  76,  116 
to  distribution  of  sample  means,  72 
Normal  distribution 

approximation  to  binomial,  76,   116 


approximation     to     distribution     of 
sample  means,  72 

areas,  table  of,  517 

bivariate,   198,  225 

central  limit  theorem,  72 

density  function  of,  39 

expected  value  of,  39 

equation  of,  39 

mean  of,  39 

parameters  of,  39 

percentage  points  of,  517 

reproductive  property,  80 

standard,  39 

standard  deviation  of,  39 

standardized,  39 

table  of  areas,  517 

tolerance  limits,  98 

variance  of,  39 
Normal  equations,  164 
Normal  population,  bivariate,  198,  225 
Normal  random  variables 

linear  combinations  of 
distribution  of,  80 
mean  of,  80 
variance  of,  80 
Normality 

assumption  of,  338 

test  for,  338 

Normalizing  transformations,  340 
Notation,  18 
Nottingham,  R.  B.,  276 
Null 

hypothesis,  109 

set,  17 
Numbers 

disproportionate,  423 

equal,  280 

expected,  117,  124,  127,  130 

proportionate,  421 

random 

table  of,  544 
use  of,  46 

unequal,  279 


Observations,  paired,  96,  121,  368 
Objectionable  quality  level  (OQL),  487 
OC  curve  (see  Operating  characteristic 

curve),  110,  487 

OC  function  (see  Operating  character 
istic  function),   108,   485 
Ogive  curve,  52 
Olds,  E.  G.,  234,  243 
Olmstead,  P.  S.,  475 
Operating  characteristic  curve 

in  acceptance  sampling,  487 

in  tests  of  hypotheses,  110 


INDEX 


577 


Operating  characteristic  function,  108, 
485 

definition  of,  108 

graph   of    (see   Operating  character 
istic  curve),  110,  487 
OQL,  objectionable  quality  level,  487 
Order  statistics,  84 
Orthogonal 

comparisons,  263,  306,  376 

contrasts,  263,  306,  376 

polynomial,  192,  313 

polynomial  coefficients,  194,  314 
Ostle,  B.,  15,  16,  101,  105 
Outcome  of  an  experiment,  30 
Owen,    D.   B.,   86,    100,    101,    105,    106, 
486,  499 


P  (...),  probability,  29 

p  chart 

center  line,  480 
control  limits,  480 
example  of,  481 
interpretation  of,  481 

n,  product  sign,  20 

Paired  observations,  96,  121,  368 

Pairing,  96,  121,  239,  368 

Pair  wise  independence,  31 

Parabolic  regression,  191 

Parameters  of 

beta  distribution,  39 
binomial  distribution,  38,  74 
chi-square  distribution,  40,  81 
exponential  distribution,  39 
F  distribution,  40,  83 
gamma  distribution,  39 
geometric  distribution,  38,  79 
hypergeo metric  distribution,  38,   73 
multinomial  distribution,  78 
negative    binomial   distribution,    38, 

79 

normal  distribution,  39 
Poisson  distribution,  38 
standard  normal  distribution,  39 
t  distribution,  40,  83 
,  Weibull  distribution,  39 

Partial 

correlation,  228 
regression  coefficient,  229 

Path  of  steepest  ascent,  424 

Paull,  A.  E.,  372,  373,  408 

Peach,  P.,  276 

Pearson,  E.  S.,  85,  86,  94,  105,  276,  561 

Pearson,  K.,  231,  232,  243,  517 

Percentage  points 

for  chi-square  distribution,  523 

for  F  distribution,  529 

for  normal  distribution,  517 


for  Poisson  distribution,  512 

for    standard    normal    distribution, 

517 

for  t  distribution,  528 
Percentages,  estimation  and  testing  of 
(see    Binomial    and    Multinomial 
distributions) 
Percentile,  57 
Percentile  limits,  57 
Permutations,  20 
Plackett,  R,  L.,  436 
Planning  of  experiments,  244 
Plot,  split  or  sub,  415 
Plot,  missing,  390,  412 
Point  estimate,  87 
Poisson  data,  124 
distribution,  37,  38 

approximation  to  binomial,  75 

equation  for,  38 

expected  value  of,  38 

mean  of,  38 

probability  function,  38 

relation  to  c-chart,  483 

standard  deviation  of,  38 

table  of  cumulative  probabilities, 

512 

tests  of  hypotheses,  124 
variance  of,  38 
Polygon,  frequency,  49 
Polynomial 

orthogonal,  192,  313 
regression,  192 
second  degree,  191 
Pomerans,  A.  J.  (see  Taton,  R.),  16 
Pooled 

chi-square,  128 
estimate  of  variance,  95 
Popper,  K.  R.,  16 
Population 

binomial,  38,  73,  74 
bivariate  normal,  198,  225 
correlation  coefficient,  36,  225 
definition,  44 
finite,  71 
mean,  36 
median,  36 
mode,  36 
moments,  36 
multinomial,  78 
normal,  39,  80 
Poisson,  38,  124 

product     moment     correlation     co 
efficient,    36 
standard  deviation,  36 
standard  normal,  39 
variance,  36 
Position,  measure  of,  55 
Positive  correlation,  224 


578 


INDEX 


Power 

function,  108 
of  a  test,  108 
Prater,  N.  H.,  184,  221 
Precision 

definition  of,  104 
measure  of,  60 

Predicted  (or  estimated)  value,  163 
Predicting   (estimating)   equation,   163 
Prediction     (or    regression)     equation, 

163 

Prediction  interval,  172 
Preliminary  tests  of  significance,  371 
Presentation  of  results,  341 
Principle  of  least  squares,  161 
Principles  of  experimental  design,  246 
Probabilities,  binomial,  32 
Probability 

additive  law,  31 

classical  definition,  29 

compound,  31 

concepts,  30 

conditional,  31,  35 

continuous  distribution,  33 

cumulative      distribution      function 

(c.d.f.),  33 
definition  of,  29 
density  function  (p.d.f.),  33 
distribution,  33 
discrete  distribution,  33 
estimates  of,  48 
function  (p.f.)»  33 
independence  (definition),  31 
joint,  34 

laws  of  (general),  31 
marginal,  34 
multiplicative  law,  31 
notation,  29 
operations  with,  31 
properties  (general),  30 
relative  frequency  definition,  30 
total,  31 

Tchebycheffs     inequality     to     esti 
mate,  71 

unconditional  (see  Marginal),  34 
Probability  distribution 

associated  with  a  random  variable, 

33 
continuous 

cumulative  distribution,  33 
expected  value  of,  35 
mean  of,  36 
moments  of 

about  mean,  36 
about  origin,  36 
standard  deviation  of,  36 
variance  of,  36 
definition  of,  33 


discrete 

cumulative  distribution,  33 
expected  value  of,  35 
mean  of,  36 
moments  of 

about  mean,  36 
about  origin,  36 
standard  deviation  of,  36 
variance  of,  36 
Procedures,  test,  111 
Process,  control  of,  477 
Producer's  risk,  487 
Product  moment,  36 

Product  moment  correlation  coef 
ficient,  36 

in  the  population,  36 
in  the  sample,  225 
Product  sign  (H),  20 
Properties  of 

arithmetic  mean.,  53 
coefficient  of  variation,  64 
estimators,  87 
mean,  53 
median,  56 
midrange,  55 
mode,  58 
probability,  30 
range,  60 

standard  deviation,  60 
variance,  60 

Proportion,  estimation  and  tests  of  (see 
Binomial  and  Multinomial  dis 
tributions) 

Proportion  defective,  94,  115,  481 
Proportionate  subclass  numbers,  421 
Purcell,  W.  R.,  276 


Quality,  477 
Quality  control,  477 
Quality  control  charts,  477 

c-chart,  479,  483 

examples  of,  479 

nature  of,  477 

Z?-chart,  479,  481 

.ft-chart,  479,  481 

5r-chart,  479,  481 
Quartile  limits,  57 

Quenouille,  M.  H.,  261,  277,  361,  409, 
436,  465 


R  chart 

center  line,  480 
control  limits,  480 
example  of,  483 
interpretation  of,  483 


INDEX 


579 


R,   coefficient   of   multiple   correlation, 

228 

R,  sample  range,  60 
R2,  correlation  index,  223 
r,  linear  correlation  coefficient,  224 
r,  sample  correlation  coefficient,  224 
r,  statistic  in  the  run  test,  470 
table  of  critical  values,  558 
r,  statistic  in  the  sign  test,  467 

table  of  critical  values,  556 
n,    sample    intraclass    correlation    co 
efficient,  236 

rsj    Spearman's    rank    correlation    co 
efficient,  233 

7*12.3,  partial  correlation  coefficient,  228 
rXc  table,  129 

p,   population   product-moment   corre 
lation  coefficient,  36 
PI,     population    intraclass    correlation 

coefficient,  236 
Random 

balance,  425 
numbers,  46 

table  of,  544 
Random  sample,  45 
Random  sampling  distributions,  70 
chi-square,  81 

difference  between  two  means,  81 
F,  83 

linear  combination  of   random  vari 
ables,  80 
mean,  80 

sample  proportion,  74 
standard  deviation,  83 
t,  83 

variance,  83 

Random  sampling  numbers,  46 
Random  variable,  definition  of,  30 
Randomization 
concept,  246 

use  in  experimental  design,  249 
Randomized  blocks,  363 
Randomized    complete    block    design, 

268,  363 

analysis  of  covariance,  444 
analysis  of  variance,  364 
assumptions,  364 
calculations,  364 
comparison  of  selected  treatments, 

376 

components  of  variance,  373 
definition,  363 
degrees  of  freedom,  365 
efficiency,  375 
expected  mean  sqxiares,  365 
factorial  treatments,  380 
individual  degrees  of  freedom,   376, 
380 


missing  observations,  390 

relative  efficiency,  373 

response  curves,  380 

standard  error  of  treatment   mean, 

366 
subdivision    of    experimental    error, 

376 

subsampling,  368 
^-test  (paired  observations),  368 
tests  of  hypotheses,  366 
variance  of  treatment  mean,  366 
Range,  47 

charts,  480,  481 

definition  of,  60 

distribution  of,  85 

sample,  60,  84 

tests  using,  500 

use  as  an  estimator  of  the  standard 

deviation,  63 

Rank  correlation,  233,  466 
Rao,  C.  R.,  436 
Ratio,  correlation,  229 
Ratner,  R.  A.,  254,  277 
Reduction  in  sum  of  squares,  187 
Region 

acceptance,  108 
critical,  109 
rejection,  109 
Regression 

abbreviated  Doolittle  method,  177 
analysis,  159 
analysis  of  variance,  166 
assumptions,  168 
average  within  groups,  202 
both  variables  subject  to  error,  199 
cautions  about,  160 
coefficient,  164 
in  comparisons,  312 
and  correlation,  223 
curvilinear,  190 
definition,  159 
deviations  from,  162 
in  each  group,  201 
equation,  163 
curvilinear,  190 
linear,  164 
multiple,  178 
polynomial 

orthogonal,  192 
second  degree,  191 
estimates,  170 
exponential,  190,  194 
general  remarks,  186 
Gompertz,  190 
graphical  interpretation,  162 
of  group  means,  202 
inverse  prediction,  176 
lack  of  fit,  188 


580 


INDEX 


Regression  (continued) 

least  squares,  161 

linear,  164 

logistic,  190 

methods,  159 

Mitscherlich,  190 

multiple,  178 

multiple  linear,  178 

nonlinear,  190 

normal  equations,  164 

orthogonal  polynomials,  192 

parabolic,  191 

partitioning  sum  of  squares,  164 

polynomial,  190 

pooled  coefficient,  202 

quadratic,  191 

reduction  in  sum  of  squares,  187 

relation    to    analysis    of    covariance, 
438 

relation  to  analysis  of  variance,  340 

several  samples,  201 

second  degree,  191 

second  order  models,  191 

simple  linear,  164 

tests  of  hypotheses,  174 

uses,  205 

weighted,  197 

Rejectable  quality  level  (RQLf),  487 
Rejection  region,  109 
Relative 

efficiency,  300 

frequency,  30,  48 

definition  of  probability,  30 

information,  300 

standard  deviation,  64 
Relevant  factors  in  an  experiment,  245 
Reliability,  105,  477,  506 
Replicate,  381 
Replication,  128,  246,  270 
Repetition,  246 
Research,  1 

Resource  economy,  245 
Response 

curve,  312 

function,  160 

surface  techniques,  424 

variable,  159 
Riordan,  J.,  28 
Risk 

consumer's,  487 

producer's,  487 

Robertson,   W.  H.,  74,  86,  133,   158 
Romig,  H.  G.,  43,  74,  86,  498 
Rourke,  R.  E.  K.,  43 
Roy,  R.  H.,  15 

RQL,  rejectable  quality  level,  487 
Run  test 

discussion,  470 


table  of  critical  values,  558 
Ryerson,  C.  M.,  15,  499,  510 


s,  sample  standard  deviation,  61 
szy  sample  variance,  60 
#61?    standard    error    of    regression    co 
efficient,  170 

SE,  standard  error  of  estimate,  170 
sx,  standard  error  of  sample  mean,  91 
Sxi-Xz,  standard  error  of  difference  be 
tween  two  sample  means,  95 
Sr,  standard  error  of  predicted  value  of 
dependent  variable  in  a  regression 
analysis,  170 
T"!,  summation  sign,  19 
cr,  population  standard  deviation,  36 
<**}  population  variance,  36 
<TXY,  population  covariance,  36 
Saaty,  T.  L.,  14,  16 
Sample 

coefficient  of  variation,  64 

correlation  coefficient,  224 

definition  of,  45 

drawing  of,  46 

judgment,  45 

mean,  53 

median,  55 

midrange,  55 

mode,  58 

moments,  70 

obtaining  of,  46 

point,  29 

probability,  45 

random,  45 
simple,  45 
stratified,  46 
systematic,  46 

range,  60,  84 

regression  coefficient,  164 

simple  random,  45 

size,  136 

space,  29 

standard  deviation,  61 

stratified  random,  46 

systematic  random,  46 

variance,  60 
Sampling 

acceptance,  477,  485 
by  attributes,  485 
by  variables,  485 

distribution,  70 

double,  488 

elements  of,  44 

error,  291 

inspection,  485 

judgment,  45 

multiple,  490 


INDEX 


581 


multistage,  297 

plans,  477 

probability,  45 

random,  45,  73 

sequential,  490 

simple,  45 

single,  485 

stratified  random,  46 

systematic  random,  46 

without  replacement,  46,  73 

with  replacement,  46,  74 
Sasieni,  M.,  16 
Satterthwaite,    F.    E.,    277,    302,    329, 

361,  425,  436 
Savage,  I.  R.,  475 
Scates,  D.  E.,  15 
Scatter  diagram,  162 
Scattergram,  162 
Scheffe,  H.,  310,  311,  361,  475 
Schenck,  J.,  Jr.,  16 
Schneider,  A.  M.,  435 
Science,  2 

Scientific  method,  2,  4 
Second  kind  of  error,  107 
Seder,  L,  A.,  499 

Selected  comparisons,  261,  303,  376 
Sequential 

fitting  of  constants,  179 

probability  ratio  test,  140 

sampling,  485,  490 

tests  of  hypotheses,  140 
Series,  20 
Set 

complementary,  17 

definition  of,  17 

null,  17 

number  of  elements  in,  18 

of  all  possible  outcomes  of  an  experi 
ment,  30,  264 

theory,  17 

universal,  17 
Shainin,  D.,  277 
Shewhart,  W.  A.,  477,  499 
Siegal,  S.,  475 
Sigma 

53,  summation  sign,  19 

<r,  population  standard  deviation,  36 
Sign  test 

discussion,  466 

table  of  critical  values,  556 
Signed  rank  test 

discussion,  468 

table  of  critical  values,  557 
Significance  level,  108 
Significance  test  (see  Hypothesis,  test 

of),  107 
Significance  tests 

acceptance  regions,  108 


chi-square  tests,  114,  117,  125,  126, 

129,  137 
for  coefficients  of  fitted  straight  line, 

174 

for  comparison  of  two  means,  119 
in  contingency  tables,  129 
critical  regions,  109 
distribution-free,  466 
for  equality  of  means  when   obser 
vations  are  paired,  121 
for  equality  of  two  percentages,  132 
F  test  (see  also  ANOVA  and  Regres 
sion),  123 

goodness  of  fit,   126,  471 
for  independence,  129 
Kolmogorov-Smirnov,  471 
that  mean  of  normal  population  has 

specified  value 
one-sided,  113 
two-sided,  113 
about  mean  of  Poisson  distribution, 

124 

that  means  of  two   normal  popula 
tions  are  equal 
one-sided,  122 
two-sided,  119 

that  means  of  k  >  2  normal  popula 
tions  are  equal,  133 
median  tests,  473 
nonparametric  tests,  466 
preliminary,  371 
procedures  (general),  107 
about     proportion     or     proportions, 

115,  132 

ratio  of  two  variances,  123 
in  regression,  174,  192 
rejection  regions,  109 
run  test,  470 
sample  size,  136 
sign  test,  466 
signed  rank  test,  468 
t  test,  113,  119,  121,  174 
that  variance  of  normal  population 

has  specified  value 
one-sided,  115 
two  sided,  114 

that  variances  of  k>2  normal  popu 
lations  are  equal,  136 
variance  ratio  test,  123 
Wilcoxon  signed  rank  test,  468 
Simon,  L.  E.,  499 
Simple  random  sample,  45 
Single  sampling,  485 
Sinkbaek,  S.  A.,  523 
Size  of  sample,  136 
Slope  (of  straight  line),  174 
Smirnov,  N.  V.,  475 
Smith,  B.  Babington,  233,  243 


582 


INDEX 


Smith,  H.  Fairfield,  465 
Smith,  K.,  475 

Snedecor,  G.  W.,  16,  94,  106,  158,  220, 
221,  243,  277,  339,  362,  409,  436, 
544 

Snell,  J.  L.,  28,  43 
Space,  17,  29 
Spearman,  C.,  233 

Spearman's     rank     correlation     coeffi 
cient,  233 
Specification  limits 
definition,  502 
relation      to      statistical      tolerance 

limits,  502 
Split  plot  (subplot) 
definition  of,  415 
design,  271,  415 

square  root  transformation,  340 
Squares,  least,  88,  161 
Standard  deviation 

of  beta  distribution,  39 

of  binomial  distribution,  38,  74 

of  chi-square  distribution,  40 

confidence  interval  for,  93 

definition,  36 

distribution  of  sample,  83 

estimated  by  sample  range,  63 

of  exponential  distribution,  39 

of  F  distribution,  40 

of  gamma  distribution,  39 

of  geometric  distribution,  38 

of  hyper  geometric   distribution,   38, 

74 

of  linear  function   of   random  vari 
ables,  80 
of  mean,  80,  91 

of  negative  binomial  distribution,  38 
of  normal  distribution,  39 
of  Poisson  distribution,  38 
of  population,  36 
of  sample,  61 
significance  test  for,  114 
of  standard  normal  distribution,  39 
of  t  distribution,  40 
of  Weibull  distribution,  39 
Standard  error  of 
comparison,  309 
contrast,  309 

difference  between  two  means,  95 
estimate,  170 
mean,  91 

predicted  value  in  regression,  170 
regression  coefficient,  170 
sample  mean,  91 
treatment  comparison,  309 
treatment  mean,  285,  366 
Y,  170 
Standard  normal  distribution,  37 


definition,  39 
table  of  areas,  517 
Starr,  M.  K.,  16 
Statistic,  test,  109 
Statistical 
control,  >78 

design  of  experiments,  244 
estimation,  87 
independence,  31 
inference,  87,  107 
model,  163 
quality  control,  477 
regularity,  30 
tests  of  hypotheses,  107 
tolerance  limits,  98 
Statistical  Research  Group,  Columbia 

University,  106,  158,  499 
Statistics 

definition  (as  a  science),  3 
definition  (plural  of  statistic),  2 
descriptive,  44 
order,  84 

relation  to  probability,  17 
relation  to  research,  3 
relation  to  scientific  method,  3 
role  in  research,  1 
scope  of,  f2 

some  fields  of  application,  5 
Steepest  ascent 
direction  of,  424 
path  of,  424 

Steps  in  designing  experiments,  264 
Straight  lines   (see  Regression),    164 
Strata,  46 
Stratification,  46 
Stratified  random  sample,  46 
Stratum,  46 

"Student"  (W.  B.  Gosset),  277 
"Student's"  t  (see  t  distribution),   37, 

40,  83 
Subclass  numbers 

disproportionate,  423 
proportionate,  421 
Subplot,  415 
Subsample,  288 
Subsampling,  288 
Subset,  18 
Sum 

algebraic,  19 
definition,  19 

of  products,  corrected,  164,  441 
of  squares,  corrected,  165 
of  matrices,  23 
of  products,  164 
of  squares,  162 
Summation 
index,  19 
notation,  19 


INDEX 


583 


sign    (),  19 
Swed,  Frieda  S.,  476,  558 
Symbolism  (see  Notation),  18 
Symbols  (see  Notation),  18 
Symmetric  matrix,  23 
System  reliability,  506 


T,  statistic  in  the  signed  rank  test,  468 

table  of  critical  values,  557 
T,  general  symbol  for  total,  135,   236, 

279 
t  distribution 

degrees  of  freedom,  83 
density  function,  40,  83 
expected  value,  40 
mean,  40 
noncentral,  113 
percentage  points,  528 
relation  to  normal  distribution,  83 
standard  deviation,  40 
table  of  values,  528 
variance,  40 
t  random  variable 
definition,  83 
degrees  of  freedom,  83 
density  function,  40,  83 
table  of  critical  values,  528 
Tables 

analysis  of  covariance,  440 
analysis  of  variance,  134,  166,  236, 

280 

contingency,  129 
frequency,  50 
rXc,  129 
2X2,  131 

chi-square  (Appendix  4),  523 
control    chart    constants    (Appendix 

8),  548 
D      in      Kolmogorov-Smirnov      test 

(Appendix  16),  560 
F  (Appendix  6),  529 
normal,  standard  (Appendix  3),  517 
number  of  observations  (see  "size  of 

sample"),  550-555 
orthogonal  polynomial  coefficients 
in     analysis     of     variance     (Table 

11.26),  314 

in  regression  (Table  8.19),  194 
Poisson  (Appendix  2),  512 
pseudo    F-statistics    (Appendix    17), 

561 
pseudo    2-statistics    (Appendix    17), 

561 

r  in  run  test  (Appendix  15),  558 
r  in  sign  test  (Appendix  13),  556 
random  numbers  (Appendix  7),  544 


size   of    sample    (Appendices    9-12), 

550-555 
X2  test  of  variance  (Appendix  11), 

554 
F  test  of   ratio   of   two   variances 

(Appendix  12),  555 
t  test  of  mean  (Appendix  9),  550 
t  test   of    difference   between   two 

means  (Appendix  10),  552 
standard  normal  (Appendix  3),  517 
T  in  signed  rank  test  (Appendix  14), 

557 

t  (Appendix  5),  528 
TaUy,  47 
Taton,  R.,  16 

Taylor  series  expansion,  505 
TchebychefFs  inequality,  71,  466 
Technique,  experimental,  248 
Test  procedures,  111 
Test  statistic,  109 
Tests  of  hypotheses 

analysis  of  covariance,  442 
analysis  of  variance,  133 
Bartlett's,  136 
binomial,  115,  132 
contingency  table,  129 
correlation  coefficient,  225 
definition,  107 

difference  between  two  means,  119 
distribution-free,  466 
goodness  of  fit,  126 
homogeneity  of  variances,  136 
Kolmogorov-Smirnov,  338,  471 
lack  of  fit,  188 
mean,  113 
multinomial,  124 
nonpara metric,  466 
notation  for,  108 
paired  observations,  121 
Poisson,  124 
preliminary,  371 
regression,  174 
sequential,  140 
standard  deviation,  114,  123 
variance,  114,  123 
Tests  of  significance   (see  Significance 

tests) 

Tetrachoric  correlation,  232 
Thomas,  G.  B.,  43 
Thompson,  C.  M.,  529 
Thompson,  G.  L,.,  28,  43 
Tinker,  M.  A,,  392,  393,  409 
Tischer,  R.  G.,  16 
Tolerance 
factors,  100 
limits 

distribution-free,  100,  466 
for  normal  distribution,  98 


584 


INDEX 


Tolerances 

engineering,  502 

natural,  98,  502 

statistical,  98,  502 
Total 

chi-square,  128 

probability,  31 
Transformations 

arcsine,  340 

general  discussion,  338 

logarithmic,  340 

reciprocal,  340 

square  root,  340 
Treatment 

combination,  252 

comparison,  261,  306,  310,  376 

contrast,  261,  306,  376 

definition  of,  252 
Treloar,  A.  E.,  231,  232,  233,  243 
Trials,  Bernoulli,  32 
Truitt,  J.  T.,  465 
Tukey,  J.  W.,  277,  310,  338,  340,  360, 

362,  436,  475 
Type  I  error,  107 
Type  II  error,  107 
2X2  contingency  tables,  131 

U 

Unbiased  estimator,  61,  70,  87 
Uniformity  data,  374 
Union,  17 

Unit,  experimental,  247 
United    States    Government,    Depart 
ment  of  Defense,  499 
Universal  set,  17 
Universe  (definition),  44 


V(X),  population  variance  (  —  cr^3),  36 
^(-XT),  sample  variance  (=5^),  60 
Value 

absolute,  18 

expected,  33,  35 
Values,  extreme,  84 
Variability  (see  also  Dispersion),  49 

measures  of,  60 
Variable 

chance,  30 

concomitant,  159 

continuous,  30 

dependent,  159 

discrete,  30 

independent,  159 

random,  30 

response,  159 

Variables  sampling  plans,  485 
Variance 

analysis  of  (see  Analysis  of  variance) 


of  beta  distribution,  39 

of  binomial  distribution,  38,  74 

of  chi-square  distribution,  40 

components  of,  237,  299 

confidence  limits  for,  93 

of  contrast,  309 

definition  of,  36 

distribution  of  sample,  83 

estimation  of,  93 

of  exponential  distribution,  39 

of  F  distribution,  40 

of  gamma  distribution,  39 

of  geometric  distribution,  38 

homogeneity  of,  136,  338 

homogeneous,  197 

of  hypergeo metric    distribution,   38, 
74 

of    linear    combination    of    random 
variables,  80 

of  negative  binomial  distribution,  38 

of  normal  distribution,  39 

of  Poisson  distribution,  38 

pooled  estimate  of,  95 

population,  36 

ratio,  97,  123 

sample,  60 

of  sample  mean,  70 

significance    tests    concerning,     114, 
123,  136,  338 

of  standard  normal  distribution,  39 

of  t  distribution,  40 

tests  about,  114,  123,  136,  338 

of  treatment 

comparison,  309 
contrast,  309 
mean,  285,  366 

unbiased  estimator  of,  61,  70 

of  Weibull  distribution,  39 
Variation 

coefficient  of,  64 

measures  of,  60 
Vector,  definition  of,  22 
Venn  diagram,  17 

W 

w,  weight,  197 

MJ,  width  of  class  interval,  54,  56,  59 

Wadsworth,  G.  P.,  43,  106,  158 

Wald,  A.,  499 

Walker,  H.  M.,  16 

Wallace,  H.  A.,  220 

Wallis,  W.  A.,  100,  106,  158,  475,  499 

Walsh,  J.  E.,  510,  561 

Ward,  G.  C.,  339,  362 

Weibull  distribution,  39 

Weight,  w,  197 

Weighted  regression,  197 

Weissberg,  A.,  100,  106 


INDEX 


585 


Western  Electric  Company,  Inc.,   499 

Whitney,  D.  R.,  475 

Whitney,  F.  L.,  16 

Whitesitt,  J.  E.,  28,  43 

Wiesen,  J.  M.,  16,  486,  488,  499 

Wilcoxon,  F.,  476,  557 

Wilcoxon  signed  rank  test 

discussion,  468 

table  of  critical  values,  557 
Wilkinson,  G.  N.,  465 
Wilks,  S.  S.,  9,  16,  101,  106,  476 
Wilson,  E.  B.,  Jr.,  16,  277 
Wilson,  K.  B.,  435 
Wishart,  J.,  276,  447,  455,  465 
Wolfowitz,  J.,  499 
Woo,  T.  L.,  230,  243 
Worthing,  A.  G.,  16 


X,  independent  variable  in  regression, 

__       159 

X,  sample  mean,  53 

X  chart 

center  line,  480 


control  limits,  480 
example  of,  479 
interp£etation  of,  479 
=  X  ~X,  deviation  from  mean,  61 


Y,  dependent  variable  in  regression, 
159 

Fadj.,  adjusted  value,  199 

If,  estimated  (or  predicted)  value,  161 

Y ",  value  of  Y  estimated  from  regres 
sion,  161 

F  — F",  deviation  from  regression,   161 

Yaspan,  A,,  16 

Yates,  253,  261,  277,  301,  335,  362,  394, 
409,  435,  528 

Youden,  W.  J.,  277,  436 

Youle,  P.  V.,  435 

Young,  S.,  435 


Zelen,  M.,  277,  435,  465 
Zucker,  J.,  276 
Zwerling,  S.,  15,  499,  510 


This  very  usable  text  Is  a  multiple- 
purpose  book  designed  specifically  for 
the  following  areas: 

descriptive  statistics  /  mathematical 
statistics  /  statistical  methods  /  the 
design  and  analysis  of  experiments 

Here  is  an  up-to-the-minute  statistics 
book  ideal  for  an  understanding  of  both 
the  classical  and  the  newly  developed 
statistical  tools  currently  being  used  in 
all  fields  of  research. 


BERNARD  OSTTJE,  Professor  of  Engineering  at 
Arizona  State  University,  includes  operation 
research,  quality  control,  reliability,  the  sta 
tistical  design  of  experiments,  and  the  static 
tical  analysis  of  data  in  his  principal  areas  o 
teaching  and  research.  He  also  serves  as  spe 
cial  consultant  for  several  industrial  firms. 
Earlier,  Dr.  Ostle  was  Supervisor  of  the  Sta 
tistics  Section  In  the  Reliability  Department 
of  Sandia  Corporation,  Albuquerque,  New 
Mexico.  From  1952  to  1957  he  was  Profes 
sor  of  Mathematics,  Agricultural  Experiment 
Station  Statistician,  and  Director  of  the  Sta 
tistical  Laboratory  at  Montana  State  College. 
Prior  to  that  he  taught  statistics  at  Iowa 
State  University  and  the  University  of  Min 
nesota.  In  addition,  he  served  as  Special  Lec 
turer  in  Statistics  at  the  University  of  New 
Mexico,  1958—60,  and  as  Special  Lecturer  in 
Statistics  for  the  National  Science  Founda 
tion  Summer  Institute  at  Oklahoma  State 
University  hi  ^Q59. 

He  received  his  J8.A.  decree  from  the  Uni-  „ 
versit\  of  British  Columbia  LI  1945  with 
honors  in  rnat'ie  it.-.ics,  and  the  TVv.A.  degree 
in  economics  from  i  he  sciine  institution  in 
1916.  He  was  awarded  the  Ph.D.  degree  in 
statietk,^  i, t  Iowa  State  Uni\ersity  in  1949. 
He  is  a  mem-ber  of  the  American  Society  for 
Quality  Control,  the  American  Statistical 
Association,  the  Operations  Research  Society 
~*  \ ***„,*  AlnK:  p;  TViii.  Ph"  Kappa  Phi, 


CD 


1  O6  O73