Skip to main content

Full text of "Naval research logistics quarterly"

See other formats


4/y^fr 


■-s 


NAVAL  B€S£flfiCH 
LOGISTICS 


DECEMBER  1980 
VOL.  27,  NO.  4 


OFFICE     OF     NAVAL     RESEARCH 

NAVSO  P-1278 


<tf?7-6 


NAVAL  RESEARCH  LOGISTICS  QUARTERLY 


EDITORIAL  BOARD 

Marvin  Denicoff,  Office  of  Naval  Research,  Chairman  Ex  Officio  Members 

Murray  A.  Geisler,  Logistics  Management  Institute 
W.  H.  Marlow,  The  George  Washington  University 


Thomas  C.  Varley,  Office  of  Naval  Research 
Program  Director 


Seymour  M.  Selig,  Office  of  Naval  Research 
Managing  Editor 


MANAGING  EDITOR 

Seymour  M.  Selig 
Office  of  Naval  Research 
Arlington,  Virginia  22217 


ASSOCIATE  EDITORS 


Frank  M.  Bass,  Purdue  University 

Jack  Borsting,  Naval  Postgraduate  School 

Leon  Cooper,  Southern  Methodist  University 

Eric  Denardo,  Yale  University 

Marco  Fiorello,  Logistics  Management  Institute 

Saul  I.  Gass,  University  of  Maryland 

Neal  D.  Glassman,  Office  of  Naval  Research 

Paul  Gray,  Southern  Methodist  University 

Carl  M.  Harris, Center  for  Management  and 

Policy  Research 
Arnoldo  Hax,  Massachusetts  Institute  of  Technology 
Alan  J.  Hoffman,  IBM  Corporation 
Uday  S.  Karmarkar,  University  of  Chicago 
Paul  R.  Kleindorfer,  University  of  Pennsylvania 
Darwin  Klingman,  University  of  Texas,  Austin 


Kenneth  O.  Kortanek,  Carnegie-Mellon  University 
Charles  Kriebel,  Carnegie-Mellon  University 
Jack  Laderman,  Bronx,  New  York 
Gerald  J.  Lieberman,  Stanford  University 
Clifford  Marshall,  Polytechnic  Institute  of  New  York 
John  A.  Muckstadt,  Cornell  University 
William  P.  Pierskalla,  University  of  Pennsylvania 
Thomas  L.  Saaty,  University  of  Pittsburgh 
Henry  Solomon,  The  George  Washington  University 
Wlodzimierz  Szwarc,  University  of  Wisconsin,  Milwaukee 
James  G.  Taylor,  Naval  Postgraduate  School 
Harvey  M.  Wagner,  The  University  of  North  Carolina 
John  W.  Wingate,  Naval  Surface  Weapons  Center,  White  Oak 
Shelemyahu  Zacks,  Virginia  Polytechnic  Institute  and 
State  University 


The  Naval  Research  Logistics  Quarterly  is  devoted  to  the  dissemination  of  scientific  information  in  logistics  and 
will  publish  research  and  expository  papers,  including  those  in  certain  areas  of  mathematics,  statistics,  and  economics, 
relevant  to  the  over-all  effort  to  improve  the  efficiency  and  effectiveness  of  logistics  operations. 

Information  for  Contributors  is  indicated  on  inside  back  cover. 

The  Naval  Research  Logistics  Quarterly  is  published  by  the  Office  of  Naval  Research  in  the  months  of  March,  June, 
September,  and  December  and  can  be  purchased  from  the  Superintendent  of  Documents,  U.S.  Government  Printing 
Office,  Washington,  D.C.  20402.  Subscription  Price:  $11.15  a  year  in  the  U.S.  and  Canada,  $13.95  elsewhere.  Cost  of 
ndividual  issues  may  be  obtained  from  the  Superintendent  of  Documents. 

The  views  and  opinions   expressed   in   this  Journal    are  those  of  the  authors  and  not  necessarily  those  of  the  Office 

of  Naval  Research. 

ssuance  of  this  periodical  approved  in  accordance  with  Department  of  the  Navy  Publications  and  Printing  Regulations, 
P-35  (Revised  1-74). 


STORAGE  PROBLEMS  WHEN  DEMAND  IS  "ALL  OR  NOTHING"* 

D.  P.  Gaver  and  P.  A.  Jacobs 

Department  of  Operations  Research 

Naval  Postgraduate  School 

Monterey,  California 

ABSTRACT 

An  inventory  of  physical  goods  or  storage  space  (in  a  communications  sys- 
tem buffer,  for  instance)  often  experiences  "all  or  nothing"  demand:  if  a 
demand  of  random  size  D  can  be  immediately  and  entirely  filled  from  stock  it 
is  satisfied,  but  otherwise  it  vanishes.  Probabilistic  properties  of  the  resulting 
inventory  level  are  discussed  analytically,  both  for  the  single  buffer  and  for 
multiple  buffer  problems.    Numerical  results  are  presented. 


1.   INTRODUCTION 

The  usual  storage  or  inventory  problems  involve  demands  imagined  to  occur  randomly, 
and  to  be  capable  of  reducing  any  available  stock  to  zero,  or  even  beyond,  when  backordering  is 
permitted.  Yet  in  many  situations  at  least  one  component  of  total  demand  is  "all  or  nothing;" 
that  is,  it  reduces  inventory  only  if  it  can  be  entirely  satisfied  by  the  inventory  present,  and  oth- 
erwise seeks  another  supplier.    Here  are  examples. 

(a)  A  manufacturer's  warehouse  is  filled  with  a  certain  item  at  the  beginning  of  the  sel- 
ling season;  let  /  denote  the  initial  inventory.  Suppose  that  demands  occur  as  follows:  a  mes- 
sage is  sent  requesting  that  7),  items  be  shipped  from  inventory,  but  only  if  the  entire  order  can 
be  filled.  That  is,  the  demand  is  satisfied  if  7^  ^  /,  in  which  case  inventory  level  is  reduced  to 
7(1)  =  7  —  D\\  while  if  D[  >  I  the  inventory  remains  unchanged  and  7(1)  =  /.  Allowing  for 
no  replenishment,  the  second  demand,  of  size  D2,  interacts  with  inventory  7(1),  so  that  it  is 
filled  if  D2  <  7(1),  but  is  not  placed  if  D2  >  /(l).  The  process  continues  along  these  lines 
until  the  selling  season  is  over  and  there  are  no  more  demands. 

(b)  A  buffer  storage  device  used  to  contain  messages  prior  to  their  batch  transmission 
has  capacity  7.  Messages  of  length  (7),,  /  =  1,2,  . . .}  approach  the  buffer  successively,  and  are 
admitted  on  an  "all  or  nothing"  basis,  just  as  was  true  of  demands  for  physical  inventory  in  (a) 
above.  Once  again  rejection  will  occur,  and  more  frequently  to  large  demands  (messages)  than 
to  short  ones. 

(c)  A  system  of  many  buffer  storage  devices  is  used  to  contain  messages  prior  to  their 
batch  transmission.  Each  buffer  has  capacity  /.  Messages  of  length  [Dn  i  =  1,2,  . . .}  approach 
the  device  and  are  successively  admitted  to  the  first  buffer  until  there  is  a  demand  that  exceeds 
its  remaining  capacity.    The  first  buffer  is  left  forever  and  the  demand  that  exceeds  the  first 

"This  research  was  supported  by  the  National  Science  Foundation  under  NSF  ENG  77-09020.  ENG  79-01438  and  MCS 
77-07587,  and  by  the  Office  of  Naval  Research  under  Contract  Task  NR042-411. 


530  DP   GAVER  AND  PA.  JACOBS 

buffer,  plus  successive  demands,  applies  to  the  second  buffer  until  one  occurs  that  exceeds  the 
remaining  capacity.  This  demand  then  applies  to  the  third  buffer,  and  so  on.  As  a  result  there 
will  be  some  unused  capacity  in  each  buffer.  For  a  similar  problem  see  the  paper  of  Coffman, 
Hofri,  and  So  [2].  For  related,  although  not  identical  formulations,  see  Cohen  [3],  Gavish  and 
Schweitzer  [6],  and  Hokstad  [7]. 

In  Section  2  we  will  discuss  some  models  for  the  situations  in  Examples  (a)  and  (b).  We 
compute  such  items  as  the  distribution  of  the  amount  of  inventory  left  at  some  time  t  and  the 
distribution  of  the  times  of  successive  unsatisfied  demands. 

In  Section  3  we  next  consider  a  model  for  Example  (c),  and  derive  equations  for  the  lim- 
iting distribution  of  used  capacity  of  a  buffer  and  the  expected  used  capacity  of  a  buffer.  It 
seems  to  be  difficult  to  obtain  simple  analytic  solutions  to  these  equations,  but  certain  illustra- 
tive numerical  results  are  provided. 

2.  THE  ONE-BUFFER  INVENTORY  PROBLEM 

Suppose  that  demands  for  available  stock  occur  according  to  a  compound  Poisson  process: 
if  N,  is  the  number  of  demands  that  occur  in  (0,  /],  then  [N,\t  ^  0}  is  a  stationary  Poisson 
process  with  rate  \;  the  sizes  of  successive  demands  [D,}  are  independent  with  common  distri- 
bution F.  Assume  that  there  are  no  replenishments  of  inventory.  Let  {/,;  t  ^  0}  denote  the 
stochastic  process  describing  available  inventory  at  time  /,  and  let  {/(«);  n  =  0, 1,  ...}  be  the 
stochastic  process  of  available  inventory  following  the  «th  demand.  It  is  apparent  from  our 
assumptions  that  both  {/,}  and  {/(«)}  are  Markov  processes. 

2.1    Functional  Equations  for  the  Amount  of  Available  Inventory 

Let 

(2.1)  4>Ut)  =  E[e~sl'] 

be  the  Laplace  transform  of  the  available  inventory  at  time  t.   Similarly,  let 

</,(*,«)  =  E[-*nn)]. 

Properties  of  the  available  inventory  can  be  studied  in  terms  of  <f>  and  <//.  It  may  be  shown  by 
using  conditional  expectations  that  0  satisfies  the  following  differential  equation. 

(2.2)  4J..x£[.-*JoV_  „,(*)]. 

Further,  >//  satisfies  the  following  difference  equation 

(2.3)  if,U  n  +  \)  =  ^{s,n)  +  E  \e~snn)  J"o' "  (esx  -  l)F(rfx)]. 

Differentiation  with  respect  to  s  at  s  =  0,  or  a  direct  conditional  probability  argument,  now  pro- 
duce equations  for  £[/,]  and  £[/(«)]: 

(2.4)  -^  £[/,]  -  -\E  [/o''x  F(dx)\ 
and 


£•[/(«  +  1)]  =  E[I(n)]  -  E  [fo'(n)  x  F(dx)\ 


In  general,  no  explicit  solutions  for  the  expected  values  are  available,  but  a  simple  lower 
bound  results  from  rewriting  (2.4)  as  follows. 


ALL  OR  NOTHING  STORAGE  PROBLEMS 


(25)  * 


£[/,]  =  -\E  \It  f \'  ±F{dx)\ 
0    It 


>  -\E[I,FU,)] 

>  -\FU)E[It], 

from  which  one  sees  that 

(2.6)  £[/,]  >  /exp  [-\F(I)t] 

and  similarly 

£[/(«)]  ^  /[l  -  -F(/)]", 
so  the  expected  available  inventory  declines  by  at  most  an  exponential  rate. 

2.2.    Explicit  Solution  When  the  Demand  Distribution  is  Uniform 

Although  Equation  (2.2)  seems  to  be  quite  intractable  for  most  demand  distributions,  it 
can  be  solved  completely  when  Fis  uniform: 


F(x)  = 


-      0  ^  x  <  c, 

c 

1         C    >   X 

and  c  >  /.   In  this  case  (2.2)  can  be  expressed  as 

(2.7)  M  =  ,£L^r/'(^_1)^l 

Bt  Jo  c 


-  KE 


-si  ~sll    I    1 

1  -  e     '        e      I,\ 


In  other  words,  <f>  satisfies  a  first-order  (quasi)  linear  partial  differential  equation  with  ini- 
tial condition  0(s,  0)  =  e~sI.   Standard  procedures  (Sneddon  [8])  easily  yield  the  solution 
,-  Qx  1  -<f>(s,t)   _    1  -  exp  [-(s  +  (X/c)t)I] 

a8)  5  "  ,  +  (X/C)t 

which  gives  the  desired  transform.    Passage  to  the  limit  as  s  — •  0  in  (2.8)  shows  that 

(2.9)  £[/]=l-exp[-(Xf/c)/] 

(K/c)t 

This  formula  can  also  be  derived  by  first  finding  an  expression  for  the  /cth  moment  of  /,,  and 
then  employing  a  Taylor  series  argument. 

In  order  to  invert  the  transform  in  (2.8)  note  that 

(2.10)  f'e -«P{I,  >x}dx=   l~f<*0  =    l-exp[-(5  +  (X//c))/] 

^o  s  5  -I-  (\/c)r 

which  is  the  transform  of  a  truncated  exponential  distribution.    Thus,  by  the  unicity  theorem 
for  Laplace  transforms, 


532  DP.  GAVER  AND  PA.  JACOBS 

Iexp[-(X//c)x]      0  ^  x  <  /, 
0  J<* 

Note  that  the  distribution  of  /,  is  absolutely  continuous  in  the  interval  (0,  /)  but  that  there  is  a 
jump  at  /corresponding  to  the  occurrence  of  no  demand  less  than,  or  equal  to,  /in  (0,  t]: 

(2.12)  P{I,=  l)  =  exp[-XtU/c)}. 

2.3.   The  Expected  Number  of  Satisfied  Demands 

Supposing  that  an  initial  inventory,  or  storage  capacity,  /  prevails,  it  is  of  interest  to  com- 
pute the  probability  that  a  demand  is  satisfied,  and  the  expected  number  of  demands  satisfied  in 
an  interval  of  length  t.  First  notice  that  if  a  demand  of  size  D{t)  appears  at  time  /,  at  which 
moment  /,  is  available,  then 

P[D(t)  <  I,\I,}  =  FU.) 

is  the  conditional  probability  that  the  demand  is  satisfied.    When  F  is  uniform,  as  is  presently 
true,  we  may  remove  the  condition  to  find  that 


P[D(t)  ^  /,}  =  E[F(I,)}- 


U\  =   1  -exp[-0u/c)/1 

\  c\  \t 


If  S(t)  is  the  number  of  demands  satisfied  during  the  time  interval  (0,  f],  then  since  demands 
arrive  according  to  a  Poisson  process  with  rate  A., 


(2.13) 


E[S(t)}  =  X  J'E[FUu)]du  =  X  r;i-exp[-(X»/c)/] 

Jo  Jo  \u 


where  E\(-)   is  an  exponential  integral;   Abramowitz  and  Stegun   [1],  and  y  —  0.5112...  is 
Euler's  constant. 

2.4   The  Time  of  the  First  Unsatisfied  Demand  and  the  Amount 
of  Unused  Inventory  at  that  Time 

As  before  F  is  the  common  distribution  function  of  the  successive  demands.    Now  let  t 
be  the  time  of  the  first  unsatisfied  demand.   Then 

P{t  >  t\N,=  n)  =  P{D]  <  /,  2)2<  ?--'£| D„  <  I -D\ £>„_,} 

=  F{n){I) 
where  F(n)  denotes  the  /rth  convolution  of  Fwith  itself.   Hence, 

(2.14)  P{t  >  /}  =  £  e-K'-^I^F{")  (/). 


Explicit  expressions  for  the  distribution  of  t  can  be  obtained  in  some  cases.    If  F  is  uni- 
form on  [0,  c]  with  c  ^  /,  then 

(2.15) 


ALL  OR  NOTHING  STORAGE  PROBLEMS  533 

where  70(z)  is  a  modified  Bessel  function  of  the  first  kind  of  the  zeroth  order.   In  this  case, 

(2.16)  E[t\  =  -j-  exp{//c}  =  ^-exp{//2F[D]}. 
If  Fis  exponential  with  mean  l/>u. ,  then 

(2.17)  P{r  >  t)  =  £  *-*'&¥■  £  e-»l&£- 

n=0  rt-        k=n  k- 

and 

/     1 


(2.18)  E[t]=  f  [1  +^7]  = 


1"'  £[/>]!■ 


Note  that  if  /is  small  relative  to  F[/J>],  then  the  expected  time  to  first  unsatisfied  demand 
when  Fis  exponential  will  be  greater  than  the  expected  time  when  Fis  uniform.  However,  for 
/large  relative  to  E[D]  the  expected  time  for  /"exponential  will  be  less  than  the  expected  time 
when  Fis  uniform. 

Let  Yn  be  the  amount  of  inventory  present  at  the  time  of  the  «th  unsatisfied  demand. 
Then  for  0  ^  a  <  / 

(2.19)  P\YX  >  I-  a)  =  fj  R(dy)FU-y) 
where 

(2.20)  R(y)=  £  F(n)(v) 

and 

(2.21)  1(1-  y)  =  1-  FU-  y). 

Again,  explicit  expressions  for  the  distribution  of  Yx  can  be  obtained  for  some  distribu- 
tions F.   If  Fis  uniform  on  [0,  c]  for  c  >  /,  then 

(2.22)  P{YX  >  /-  a)  =  1  ■ 
If  Fis  a  truncated  exponential 


(2.23)  Fix)  - 


1  -  e~»x         .   . 

7     X    <    /, 

1  -  <?""' 
1  x  >  I, 


then 

(2.24)  P{YX  >  /  -  a]  =  1  -  [e^a  -  e^1]  [1  -  e^T1  exp  [fia[l  -  e^T1}. 
If  Fhas  an  exponential  distribution  with  mean  1//a,  then 

(2.25)  P{YX  >  I-  a}  =  e^{'-a). 

In  this  last  case,  the  distribution  function  of  Yn  can  be  computed  by  induction  quite  easily  and 

(2.26)  P{Yn  >  I  -  a}  =  *-«*(/-•>. 


534  DP.  GAVER  AND  PA.  JACOBS 

Hence,  when  Fis  expjnential 

(2.27)  E[Y„]  -  —  [1  -  frnft1. 

np, 

In  principle,  similar  results  can  be  obtained  for  other  distributions,  but  we  have  found  no  sim- 
ple expressions. 

2.5.    Inventory  Costs  and  Policies 

There  are  at  least  three  monetary  quantities  which  affect  the  profitability  of  an  inventory 
policy  over  a  fixed  interval  of  time  (0,  /]:  the  selling  price,  /r,  the  storage  cost,  a;  and  the  cost 
of  lost  demands,  b.  If  the  storage  cost  a  is  charged  just  on  the  basis  of  /  (something  like  ware- 
house size)  then  the  total  expected  profit  in  (0,/]  is 

Z(/)  =  pil  -  £■[/,])  -  al  -  bSU) 

=  (p-  a)I-  p\-t\      [l-exp[-(A//c)/]] 


■  b\y  +  In  \—l\  +  £,| 


HI 


for  the  case  of  uniformly  distributed  demands;  see  ((2.9)  and  (2.13)).  One  can  numerically 
find  the  maximum  expected  profit  for  this  case;  nothing  explicit  seems  to  be  available. 

3.   THE  MANY-BUFFER  STORAGE  PROBLEM 

In  this  section  we  will  study  a  model  for  the  situation  of  Example  (c)  in  Section  1.  Mes- 
sages are  successively  admitted  to  the  «th  buffer  until  there  is  a  message  length  that  exceeds 
the  remaining  capacity  of  the  buffer.  The  total  amount  of  this  message  is  put  in  the  (n  +  1)5/ 
buffer  and  the  «th  buffer  is  left  forever.  Successive  messages  are  then  put  in  the  (n  +  1)  st 
buffer  until  there  is  a  message  whose  length  exceeds  the  remaining  capacity  of  the  (n  +  I)  st 
buffer;  this  message  is  put  in  the  in  +  2)  nd  buffer  and  so  on. 

Let  /  denote  the  common  capacity  of  the  buffers  and  D,  denote  the  length  of  message  /. 
Assume  {/),}  is  a  sequence  of  independent  identically  distributed  random  variables  with  distri- 
bution F  having  a  density  function  /such  that  fix)  >  d  >  0  for  x  6  [0,  /].    Let  R  ix)  =  £ 

Fin)  (x)  be  the  renewal  function  associated  with  F.  If  F(I)  <  1,  then  we  will  assume  that  an 
incoming  message  to  the  currently  used  nth  buffer  of  length  greater  than  /  is  sent  to  the 
in  +  \)st  buffer;  when  it  cannot  fit  into  the  (n  +  \)st  buffer,  then  it  is  "banished,"  i.e.,  sent  to 
some  other  set  of  buffers.  The  next  message,  however,  will  try  to  enter  the  (n  +  l)st  buffer. 
If  this  message  has  length  greater  than  /  it  is  banished  and  the  following  message  will  try  to 
enter  the  in  +  \)st  buffer;  all  messages  of  length  exceeding  /will  be  banished  until  one  appears 
that  is  smaller  than  /  and  it  will  be  the  first  entry  in  buffer  in  +  1). 

This  model  has  been  studied  for  demand  distributions  Fwith  F(I)  =  1  by  Coffman  et  al. 
[2].  Their  approach  was  to  study  the  Markov  process  describing  the  total  amount  of  inventory 
or  space  consumed  in  successive  buffers  or  bins.  Here  we  study  the  process  {/.„},  where  L„  is 
the  size  of  the  demand  that  first  exceeds  the  remaining  capacity  of  the  nth  buffer; 
[L„;  n  =  1,2,  . ..}  is  a  Markov  process.   Let 

K{x,  [0,v])  =  P{Ln  +  x  ^  y\Ln  =  x). 


ALL  OR  NOTHING  STORAGE  PROBLEMS 


535 


Note  that 

P[LX  ^  y}  =  K(0,   [0,y]) 

is  the  same  as  the  sum  of  the  forward  and  backward  recurrence  times  at  time  /  for  a  temporal 
renewal  process  with  interrenewal  distribution  F,  see  Feller  [5].   Thus  for  y  <  / 

(3.1)  H,{y)  =  P[LX  <  *}-/,*<&)  [Fiy)  -  FU  -  z)\. 

Note  that  for  y  <  I 

X 


(3.2) 


Hence, 


Kix,  [0,y])  - 


(3.3)        Kix,dy)  = 


R(dz)[F(y)-F(I  -  x-  z)] 
if  x  <  I  -  y; 

Jo    XR(dz)[F(y)-F(I  -  x  -  z)\ 
if  /  -  y  <  x  <  I; 

ji'_R(dz)[F(y)-F(I  -  z)} 
if  x  >  I. 


[R(I  -  x)  -  RU  -  x  -  y)]F(dy)     if  x  <  I  -  y, 
R(I   -  x)  F(dy)  if  /  >  x  >  I  - 

kiy)  F(dy)  -  JJ  Ridz)  f(y  -  z)  +  R(dy)F(y) 

.    if  x  =  I  -  Y, 
[RU)  -  RU  -  y)]  F(dy)  if  x  >  I. 


Note  that  for  some  0  <  a  <  b  <  /,  there  exists  a  8  >  0  such  that  for  all  x 

KHx.dy)  ^  8  for  y  €  [a,  b] 

where  K2(x,  dy)  =  J      K(x,  dz)  K(z,  dy).    Hence,  hypothesis  D'  on  page  197  of  Doob  [4] 
satisfied.   Thus,  if 

K"(x,  A)  =  P[Ll+„  6  A\LX=  x) 

for  all  Borel  subsets  A,  then 

(3.4)  lim  K"{x,A)  =  H{A) 
exists  and  further  the  convergence  is  geometric 

\K"(x,A)  -  H(A)  |  <  ay" 
for  some  positive  constants  a  and  y,  y  <  1  for  all  A. 

Now  let 

//„(*)  =  P{L„  €  [0,x]\LQ=0}. 
Then  a  renewal  argument  can  be  used  to  show  that  for  x  ^  / 

(3.5)  ff„+iOc)-  Jj[xHn  *  R(dy)  [Fix)-  F(f  -  y)} 

+  [1  -  //„(/)]  f'_xR(dy)  [Fix)  -  FU  -  y)\. 


536  DP  GAVER  AND  PA.  JACOBS 

Taking  limits  as  //  — ►  °o  it  is  seen  that  the  distribution  Hix)  satisfies  the  following  equation  for 

(3.6)  Hix)  =  f^  H  *  R  idy)  [Fix)  -  F(I  -  y)] 

+  [1  -  //(/)]  ff   .  R  {dy)  [F(x)  -  Fil  -  y)]. 
Equations  (3.1)  and  (3.6)  can  be  simplified  for  certain  specific  distributions  F. 

3.1  Exponential  Demands 

For  the  exponential  distribution  with  mean  1  and  x  ^  /the  equations  are 

(3.7)  //,(*)  =  1  -  e~x-  xe~x 
and 

(3.8)  Hix)  =  xe~xHil)  +  //,(*)  -  e~x  J*  H(I  -  x  +  u)  du. 

3.2  Uniform  Demands 

For  the  uniform  distribution  on  [0,  e)  with  c  >  /they  simplify  to 
(3.9) 
and 

(3.10)  Hix)  -  -exp  -  (/  -  x)\  fQ    \\p\-  -u\Hiu)du 

x\    C'        f       1     ) 

—  I  J     exp u\  Hiu)du 

+  [1  -  //(/)]  Hy(x), 

for  x  ^  I.    Similar  expressions  hold  for  x  >  /,  but  they  are  unimportant  in  the  present  con- 
text. 

Equations  (3.6),  (3.8)  and  (3.10)  do  not  seem  to  yield  explicit  answers.    As  a  result,  we 
have  solved  (3.8)  and  (3.10)  numerically  by  iteration  using  the  system  of  equations 

(3.11)  //„  +  ,(*)  =  xe-xHniI)  +  H\ix)  -  e~x  jj  //„(/  -  x  +  u)du 
with  H\  as  in  (3.7)  and 

(3.12)  //„+,U)  =  -exp  -(/-x)     {'  %xp \--u \Hniu)du 

c  c  \  Jo  c 


:-H„iI)  -  -exp  -/Ml  -  -     f 'exp  -  -  u\Hniu)du 
c         [  c    )\  cjJo  I       c    J 


+  [1  -  //„(/)]  Hxix) 

with  H\  as  in  (3.9).    For  the  cases  carried  out  the  convergence  is  rapid;  after  n  =  5  iterations, 
very  little  change  is  noted  and  convergence  has  occurred  for  most  practical  purposes. 


ALL  OR  NOTHING  STORAGE  PROBLEMS  537 

Next  let  Yn  be  the  amount  of  storage  space  used  in  the  «th  bin;  the  distribution  of  Y„  is 
denoted  by  G„(x),  and 

G(x)  =  lim  P{Yn  ^  x}  =  lim  G„(x) 
is  the  long-run  distribution.    By  probabilistic  arguments  and  (3.4) 

(3.13)  G{x)  =  fjff  *  R{dy)  F(I  -  y)  +  [1  -//(/)]  f^R  (dy)  F(I  -  y) 

where  F(I  —  y)  =  1  —  F(I  —  y)  and  the  long-run  average  expected  capacity  of  a  bin  that  is 
actually  used  is 

A  =   f   xG(dx). 

Jo 

For  the  case  in  which  Fis  exponential  with  unit  mean 

(3.14)  A  =  I  -  [\-  //(/)]  [1  -  <?-1  -  e~l  J"Q  exH(x)dx. 
For  the  case  in  which  fis  uniform  on  [0,  c]  with  c  ^  / 


-2   (*'//(«) 

Jo 


)du  +  exp  —  / 
I  c 

c'       (      1    1 

1     exp u\  H{u)du 

JO                  c     1 

c  exp  — 1\  +  c 
\c  \ 

+    — /  +  c  exp  — 1\  -  c\ 

(3.15) 


+  //(/)    2/- 

Numerical  solutions  were  obtained  for  Equations  (3.14)  and  (3.15)  by  first  computing  the  pro- 
babilities H„(x),  n=  1,2,  ...,  10  iteratively  from  (3.7)  and  (3.11)  for  the  exponential 
demand  case,  and  from  (3.9)  and  (3.12)  for  the  case  of  uniform  demands.  Our  technique  was 
simply  to  discretize  x:  x, ■■  =  jh,  h  =  I/N,  N  being  the  number  of  x- values  at  which  H„(x)  is 
evaluated  (values  of  A/ from  200-1200  were  utilized  in  order  to  obtain  two-significant  digit  accu- 
racy). The  integrals  were  then  approximated  by  a  summation,  i.e.  Simpson's  rule.  Having  the 
values  of  H„(xj)  it  is  possible  to  calculate  those  of  Hn+\(xj),  and  from  these  the  values  of 
G„(x)  and  the  mean  usage,  E[Y„],  may  be  calculated  by  numerical  integration.  In  the  case  of 
exponential  demand  very  simple  upper  and  lower  bounds  were  obtainable;  such  bounds  were 
not  tight  enough  to  be  useful  for  the  uniform  case. 

The  following  table  summarizes  the  numerical  results.  We  have  compared  demand  distri- 
butions that  result,  as  nearly  as  possible,  in  the  same  probability  that  an  initial  demand  on  an 
empty  bin  will  be  rejected.  We  have  tabulated  the  expected  level  to  which  the  bin  is  filled.  It 
is  interesting  that  the  limiting  bin  occupancy  is  0.75  when  a  uniform  demand  over  the  range  of 
the  bin  size  is  experienced.  This  result  has  been  obtained  analytically  by  Coffman  et  al.  [2];  in 
that  paper  simple  and  elegant  analytical  expressions  for  G  and  H  also  appear  for  this  case.  The 
considerable  similarity  of  the  numbers  in  the  rows  of  the  table  is  notable;  apparently  the  long- 
run  bin  occupancy  is  only  slightly  larger  than  is  that  of  the  first  bin,  and  the  occupancy  experi- 
enced for  uniform  demand  is  only  slightly  larger  than  for  exponential.  Further  investigations  to 
examine  the  reasons  for  this  insensitivity  would  seem  to  be  of  interest. 

ACKNOWLEDGMENTS 

D.  P.  Gaver  wishes  to  acknowledge  the  hospitality  of  the  Statistics  Department,  University 
of  Dortmund,  West  Germany,  where  he  was  a  guest  professor  during  the  summer  of  1977,  and 
where  part  of  this  work  was  carried  out. 


538 


DP.  GAVER  AND  PA.  JACOBS 


Expected  Fraction  of  Bin  Filled 

ifn-E[Y„]  +  I) 

Rejection  Probability 

Exponential  Demand 

Uniform  Demand 

F(I) 

A                            /oo 

/i 

/oo 

0.00 

-             - 

0.76 

0.75 

0.05 

0.74            0.75 

0.74 

0.74 

0.10 

0.69            0.70 

0.72 

0.72 

0.15 

0.65            0.66 

0.68 

0.69 

0.20 

0.60            0.62 

0.64 

0.66 

0.25 

0.56            0.58 

0.60 

0.62 

REFERENCES 


[1]  Abramowitz,  M.  and  I.  A.  Stegun,  Handbook  of  Mathematical  Functions.  National  Bureau  of 
Standards,  AMS  55,  Washington,  D.C.  (1965). 

[2]  Coffman,  E.G.,  Jr.,  M.  Hofri  and  K.  So,  "A  Stochastic  Model  of  Bin-Packing,"  Technical 
Report,  TR-CSL-7811,  Computer  Systems  Laboratory,  University  of  California,  Santa  Bar- 
bara, California  (1978),  (submitted  for  publication  to  a  technical  journal). 

[3]  Cohen,  J.W.,  "Single-Server  Queues  with  Restricted  Accessibility."  Journal  of  Engineering 
Mathematics,  3,  253-284  (1969). 

[4]  Doob,  J.L.,  Stochastic  Processes,  (John  Wiley  and  Sons,  New  York,  N.  Y.,  1952). 

[5]  Feller,  W.,  An  Introduction  to  Probability  Theory  and  Its  Applications,  II,  (John  Wiley  and 
Sons,  New  York,  N.  Y.,  1966). 

[6]  Gavish,  B.,  and  P.  Schweitzer,  "The  Markovian  Queue  with  Bounded  Waiting  Time." 
Management  Science,  23,  1349-1357  (1977). 

[7]  Hokstad,  P.,  "A  Single-server  Queue  with  Constant  Service  Time  and  Restricted  Accessibil- 
ity." Management  Science,  25,  205-208  (1979). 

[8]  Sneddon,  I.,  Elements  of  Partial  Differential  Equations,  (McGraw-Hill,  New  York,  N.  Y. 
1957). 


RELIABILITY  GROWTH  OF  REPAIRABLE  SYSTEMS 


Stephen  A.  Smith  and  Shmuel  S.  Oren* 

Analysis  Research  Group 

Xerox  Palo  Alto  Research  Center 

Palo  Alto,  California 

ABSTRACT 

This  paper  considers  the  problem  of  modeling  the  reliability  of  a  repairable 
system  or  device  that  is  experiencing  reliability  improvement.  Such  a  situation 
arises  when  system  failure  modes  are  gradually  being  corrected  by  a  test-fix- 
test-fix  procedure,  which  may  include  design  changes.  A  dynamic  reliability 
model  for  this  process  is  discussed  and  statistical  techniques  are  derived  for  es- 
timating the  model  parameters  and  for  testing  the  goodness-of-fit  to  observed 
data.  The  reliability  model  analyzed  was  first  proposed  as  a  graphical  technique 
known  as  Duane  plots,  but  can  also  be  viewed  as  a  nonhomogeneous  Poisson 
process  with  a  particular  mean  value  function. 


1.   INTRODUCTION 

Predicting  the  reliability  of  a  system  or  piece  of  equipment  during  its  development  process 
is  an  important  practical  problem.  Reliability  standards  are  often  a  major  issue  in  the  develop- 
ment of  transportation  facilities,  military  systems,  and  communication  networks.  For  commer- 
cial products  that  are  to  be  leased  and  maintained  in  a  competitive  marketplace,  system  reliabil- 
ity estimates  strongly  influence  predicted  profitability  and  customer  acceptance.  When  consider- 
ing a  system  that  is  modified  in  response  to  observed  failures,  most  classical  statistical  estima- 
tion techniques  are  not  applicable.  This  is  because  the  system  reliability  is  improving  with  time, 
while  most  statistical  techniques  require  repeated  samples  under  identical  conditions. 

A  frequently  used  graphical  model  of  reliability  growth  of  repairable  systems  is  known  as 
"Duane  Plots,"  proposed  by  J.  T.  Duane  [9].  This  model  is  based  on  the  empirical  observation 
that,  for  many  large  systems  undergoing  a  reliability  improvement  program,  a  plot  of  cumula- 
tive failure  rate  versus  cumulative  test  time  closely  follows  a  straight  line  on  log-log  paper. 
Several  recent  papers  present  applications  of  Duane  plots,  e.g.,  [4],  [9]  and  [10].  Estimating 
the  parameters  of  the  Duane  model,  i.e.,  the  slope  and  intercept  of  the  straight  line  fit,  is  some- 
what difficult  to  do  directly  on  the  graph  [5].  Weighted  least  squares  and  regression  techniques 
are  sometimes  used  ([9],  [10])  to  obtain  parameter  values. 

An  underlying  probabilistic  failure  model  that  is  consistent  with  the  Duane  reliability 
model  is  the  nonhomogeneous  Poisson  process  (NHPP)  whose  intensity  is  total  test  time  raised 
to  some  power.  (See  [7]  and  [8]).  Assuming  the  sample  data  consists  of  all  the  individual 
failure  times,  Crow  [7]  derived  maximum  likelihood  estimates  for  the  Duane  model  parameters 
and  a  goodness-of-fit  test  based  on  the  Cramer-von  Mises  statistic  (Parzen  [12,  p.  143]).  A 
more   general   NHPP   model   was   proposed   by   Ascher  and   Feingold    [1],   which   also   used 

•Now  with  Dept.  of  Engineering-Economic  Systems,  Stanford  University,  Stanford,  CA. 


540  S  A.  SMITH  AND  S.S.  OREN 

the  Cramer-von  Miies  statistic  for  goodness-of-fit  testing.  Critical  values  of  this  statistic,  how- 
ever, must  be  obtained  by  Monte  Carlo  simulation  for  each  sample  size.  Crow  [7,  p.  403]  cal- 
culated and  tabulated  values  for  sample  sizes  up  to  sixty.  These  parameter  estimates  and 
goodness-of-fit  test  deal  effectively  with  Duane  model  applications  having  small  sample  sizes. 
The  facts  that  all  failure  times  must  be  stored  and  the  goodness-of-fit  measure  must  be 
evaluated  by  simulation  make  this  approach  difficult  for  larger  sample  sizes.  A  recent  paper  by 
Singpurwalla  [13]  proposes  a  time  series  model  for  reliability  dynamics.  This  model  can,  of 
course,  be  applied  to  any  type  of  reliability  trend  data,  but  requires  data  tabulation  at  a  larger 
number  of  time  stages  and  does  not  have  the  intuitive  appeal  of  the  Poisson  process  for  model- 
ing failure  occurrences  in  certain  systems. 

Our  paper  develops  statistical  estimators  for  the  Duane  model  parameters  based  on  tabu- 
lating the  number  of  failures  between  fixed  points  in  time.  This  approach  has  the  advantage  of 
using  "sufficient  statistics"  for  the  data  collection,  i.e.,  the  dimension  of  the  data  does  not 
increase  with  sample  size.  Parameter  estimates  are  obtained  by  maximum  likelihood  and  a 
goodness-of-fit  test  based  on  the  Fisher  chi-square  statistic  is  derived.  This  test  has  the  advan- 
tage that  chi-square  tables  are  readily  available  for  all  sample  sizes  and  significance  levels.  The 
accuracy  of  the  chi-square  test  decreases,  however,  as  the  sample  size  gets  small.  Sample  sizes 
for  which  the  techniques  of  this  paper  apply  are  found  in  developmental  systems  that  experi- 
ence frequent,  minor  failures  such  as  paper  jams  in  photo  copy  machines,  voltage  fluctuations 
in  power  supply  systems,  faults  in  semiconductor  manufacturing  processes,  etc.  The  last  sec- 
tion of  this  paper  illustrates  the  application  of  the  estimation  and  goodness-of-fit  techniques  to  a 
representative  set  of  simulated  failure  data. 

Regardless  of  how  the  parameters  of  the  Duane  model  are  obtained,  considerable  caution 
is  required  when  extrapolating  reliability  trends  beyond  the  observed  data  to  future  time  points. 
Major  breakthroughs  or  setbacks  in  the  reliability  improvement  program  may  cause  significant 
deviations  from  the  straight  line  projections.  Some  users  recommend  reinitializing  the  model 
and  shifting  to  a  new  straight  line  fit  when  major  changes  in  the  program  occur.  Even  if  one  is 
uneasy  about  extrapolating  the  reliability  growth  model  to  estimate  future  reliability,  it  remains 
a  valuable  tool  for  obtaining  a  "smoothed"  estimate  of  current  system  reliability.  While  reliabil- 
ity is  changing,  sample  sizes  at  any  point  in  time  are  not  sufficient  for  conventional  statistical 
estimation  techniques.  With  a  dynamic  reliability  model,  past  and  current  failure  data  can  be 
combined  to  obtain  estimates  of  current  reliability  based  on  fitting  all  observed  data. 

2.    THE  DUANE  MODEL 

The  Duane  model  states  that  cumulative  failure  rate  versus  cumulative  test  time,  when 
plotted  on  log-log  paper,  follows  approximately  a  straight  line.  More  precisely,  if  we  let  N(0,t) 
represent  the  total  number  of  failures  observed  up  to  time  /,  we  have  that 

(2.1)  log[N(0,t)/t]~-  b\ogt  +  a, 

where  the  fitted  parameters  are  a,  b  >  0.  The  relationship  is  meaningless  at  t  =  0  but,  as  most 
users  point  out  ([5], [9]),  a  certain  amount  of  early  data  is  generally  excluded  from  the  fit 
because  it  is  influenced  by  factors  such  as  training  of  personnel,  changes  in  test  procedures,  etc. 
Equation  (2.1)  therefore  implies  that 

N(0,t)/t  =  at~b,  where  a  =  log  a, 

for  t  beyond  a  certain  point.  It  should  be  emphasized  that,  in  all  applications,  time  t 
corresponds  to  cumulative  operating  time  or  test  time.  For  the  results  of  this  paper  it  is  most 
convenient  to  write  the  Duane  model  as: 

(2.2)  N(0,t)  =  at13,  where /3  =  1  -  b. 


RELIABILITY  GROWTH  OF  REPAIRABLE  SYSTEMS  541 

For  a  fairly  diverse  set  of  observed  systems,  Codier  [5,  p.  460]  has  found  b  to  be  generally 
between  0.3  and  0.5,  corresponding  to  /3  between  0.5  and  0.7. 

3.   AN  UNDERLYING  STATISTICAL  MODEL 

In  this  section  we  describe  a  statistical  model  for  the  failure  process  that  is  consistent  with 
assuming  that  the  observed  failure  data  fits  the  Duane  model.  Suppose  the  probability  that  the 
system  fails  at  time  t  (strictly  speaking  in  a  small  interval  [t,t  +  dt)),  regardless  of  the  past,  is 
determined  by  a  hazard  function  hit).   That  is, 

hit)dt  ~  P  {the  system  fails  in  the  interval^,/  +  dt)), 

independent  of  its  previous  failure  and  repair  history.    The  expected  number  of  failures  in  any 
time  interval  [tx,t2)  of  operating  time  is  then  given  by  the  mean  value  function 


(3.1)  M(th  t2)  =  j'2  h(t)dt. 


Furthermore,  it  can  be  shown  (See  Parzen  [12,  Sect.  4.2])  that  N(tu  t2),  the  number  of 
failures  observed  in  some  future  time  interval  [t\,  /2),  has  probability  distribution 

(3.2)  p{N{tx,t2)  =  k)=  ([A/(/,,  t2)]k/k\)  exp[-M(tht2)}      £  =  0,1,2,...     . 

In  addition  to  its  mathematical  convenience,  this  model  has  considerable  intuitive  appeal.  The 
simple  Poisson  process  has  been  used  successfully  to  model  the  failure  occurrences  of  many 
devices,  or  collections  of  devices  operating  in  series.  One  may  think  of  a  system  having  a 
nonhomogeneous  Poisson  failure  process  as  a  large  collection  of  simpler  devices  in  series,  with 
individual  device  failure  modes  being  gradually  removed  with  time. 

The  mean  value  function 

(3.3)  MU\,  t2)  =  aUf  -  tf ),    where  a,  £  >  0; 

corresponding  to  hit)  =  afifi~\  is  of  particular  interest.  Crow  [7,  p.  405]  pointed  out  that  the 
number  of  failures  from  a  process  with  this  mean  value  function  will  approximate  the  Duane 
Model  by  observing  that 

\o%[Mit)/t]  =  log  a  +  (/3  -  1)  log  t,     where    Mit)  --=  Af  (0,  t). 

This  means  that  system  failure  data  from  a  NHPP  with  mean  value  function  Mit)  will  approach 
the  Duane  model  with  probability  one.  Conversely,  this  process  with  mean  value  function 
Mit)  is  the  only  model  with  independent  increments  that  approximates  the  Duane  model  in  a 
probabilistic  sense  for  sufficiently  large  sample  sizes.  We  will  not  give  a  proof  of  these  state- 
ments but  refer  the  reader  to  Parzen  [12,  ch.  4]  or  Donelson  [8]  for  a  complete  discussion. 

4.   SELECTING  A  STARTING  TIME 

The  Duane  reliability  model  and  the  expected  number  of  failures  in  Equation  (3.3)  are 
both  nonlinearly  dependent  on  the  choice  of  the  time  origin.  That  is,  if  we  begin  observing 
failures  at  time  t  =  t0  >  0  and  ignore  the  first  NiO,  t0)  failures  and  the  time  interval  [0,  /0),  we 
do  not  obtain  the  same  parameters  a  and  /3  by  fitting  the  subsequent  data.  Since  the  logarithm 
is  a  strictly  concave  function,  there  is  only  one  choice  of  t0  that  can  give  a  straight  line  fit  to  the 
data  on  log-log  paper.  Specifying  the  operating  time  t0  that  is  assumed  to  have  elapsed  before 
the  beginning  of  the  modeling  process  is  therefore  an  important  step. 

Some  users  of  the  Duane  Model  ([5], [10])  suggest  reducing  the  cumulative  failures  and 
observation  time  by  removing  early  data  to  obtain  a  straight  line  fit.   This  is  done  graphically  by 


542  S.A.  SMITH  AND  S.S.  OREN 

successively  shifting  ihe  origin  to  the  right  and  replotting  the  data  until  a  straight  line  fit  is 
obtained.  With  each  shift  to  the  right,  the  shape  of  the  graph  of  cumulative  failures  versus 
cumulative  observation  time  becomes  more  downward  bending  (concave),  so  it  is  not  hard  to 
tell  when  the  best  point  has  been  located. 

Sometimes  a  terminal  straight  line  trend  on  log-log  paper  is  observed  before  the  noisy 
early  data  is  dropped.  If  the  origin  is  shifted  further  to  the  right,  the  straight  line  shape  will 
become  concave.  Therefore,  the  most  that  can  be  said  in  this  case  is  that,  for  /  greater  than 
some  t\,  the  data  fits  the  Duane  model.  The  statistical  model  (3.3)  can  still  be  applied,  how- 
ever, by  testing  to  see  if  the  number  of  failures  NU\,t)  after  the  first  N(0,t\)  fit  the  NHPP 
with  mean  value  function  M(t)  for  /  >  t\. 

5.    ESTIMATING  THE  MODEL  PARAMETERS 

If  the  Duane  model  is  applied  graphically,  the  user  can  attempt  to  estimate  the  parameters 
a  and  /3  by  drawing  the  best  straight  line  through  the  plotted  points.  This  is  somewhat  tricky 
because,  with  cumulative  failure  data,  the  later  points  should  be  weighted  more  heavily  in 
determining  the  fit.  This  section  describes  a  statistical  estimation  procedure  based  on  the 
NHPP  model  of  the  failure  process.  We  consider  two  possibilities  for  collecting  and  recording 
system  failure  times.  The  first  is  to  record  the  occurrence  time  of  each  failure,  which  yields  a 
sequence  of  observed  times  T\,  T2,  . . .  ,  TN.  This  case  has  been  analyzed  by  Crow  in  [7]  and 
the  maximum  likelihood  estimates  are  given  by 

(5.1)  a*=N/Tft' 
and 

(5.2)  /3*  =  -A7£log(r,/7V). 

A  goodness-of-fit  test  corresponding  to  these  estimators  is  derived  in  [7]  and  critical  values  of 
the  error  statistic  are  tabulated  for  sample  sizes  2-60. 

If  large  numbers  of  failures  are  observed,  it  is  often  convenient  to  record  only  the  aggre- 
gate number  of  failures  between  each  pair  in  a  sequence  of  fixed  time  points  t0,  t\,  . . .  ,  t„.  In 
this  case  the  data  is  in  the  form  N\,  N2,  ■■■  ,N„,  where  N,  =  number  of  failures  observed  in 
the  interval  [^_lf  ?,).  Maximum  likelihood  estimates  and  a  goodness-of-fit  criterion  for  obser- 
vations in  this  form  are  developed  in  the  next  few  paragraphs. 

Maximum  Likelihood  Estimates  for  the  Aggregated  Case 

We  first  calculate  the  likelihood  function  for  the  data  TV,,  N2,  ■■■  ,  N„,  given  the  time 
points  t0,  t\,  . . .  ,  t„  and  the  assumed  form  of  the  mean  value  function  in  Equation  (3.3).  The 
probability  of  N,-  system  failures  in  the  interval  [t,-i,  t,)  is  obtained  from  Equation  (3.2).  Since 
the  underlying  model  assumes  that  each  of  the  time  segments  is  independent,  the  likelihood 
function  can  be  written  as  a  product  of  these  probabilities, 

(5.3)  Ka.jS)  =  fl  P[NU,-U  /,)  =  N,}  =  exp{-M(?0,O}  f\  UMU,-u  f,)l *'/#/!). 


To  simplify  the  calculation  of  the  estimators,  we  take  the  log  of  L  (a,  /3),  noting  that  max- 
imizing the  log  will  yield  the  same  maximum  likelihood  estimates.   From  (5.3)  we  have 

(5.4)  logL(a,/S)  =  -a(t?  -  t§)  +  £  N,[\oga  +  log  (^  -  *£,)]  -  5>gfy!. 


RELIABILITY  GROWTH  OF  REPAIRABLE  SYSTEMS  543 

Taking  the  partial  derivatives  (dlog  L)/da  =  0  and  (Qlog  D/9/3  =  0,  we  obtain  the  equations 
for  the  maximum  likelihood  estimates, 

(5.5)  a*  =  N/Ur~  t§'),     where  N  =  £  Nt 

[  log  ti  -  Pi  log  f,_,         log  t„  -  p0log  /0  1 


(5.6)  0=  £#, 


1  -  Pi  1  -  Po 


p,=  (W/)"*.  '=  1.  2,  ...  ,  n,    and  Po  =  (f0/O**- 
Equation  (5.6)  is  an  implicit  function  of /3  *,  but  can  be  solved  iteratively  by  a  computer  algo- 
rithm or  programmable  calculator,  because  the  right  hand  side  is  strictly  decreasing  in  j8  *.    To 
verify  this  fact,  consider  any  two  times  t,  t'  and  compute  the  derivative 

(5.7)       '      (9/9/3)  [log  /  -  7*'logf']/[l'-  7*]  =  - 7*  (log  n2/U  -  T^)2,     where  T=  tit'. 

This  derivative  is  negative  and  decreasing  in  7Tor  0  <  T,  /3  <  1.  The  derivative  of  the  sum  in 
(5.6)  is  a  sum  of  terms  involving  the  difference  of  the  derivative  (5.7)  evaluated  at  T  =  f,_|//, 
and  t0/tn.  The  fact  that  (5.6)  is  decreasing  in  ^3*  follows  from  the  fact  that  (5.7)  is  decreasing 
in  T,  i.e.,  its  largest  or  least  negative  value  occurs  at  T  =  t0/t„.  Therefore,  (5.6)  has  a  unique 
solution. 

6.   GOODNESS-OF-FIT  CRITERION 

This  section  describes  a  procedure  for  testing  the  goodness-of-fit  of  the  observed  failure 
data  to  the  NHPP.  We  assume  that  the  parameters  a  *  and  /3  *  are  obtained  from  the  maximum 
likelihood  estimates  (5.5)  and  (5.6).  From  the  form  of  (5.5),  it  is  clear  that  the  estimate  a  *  is 
defined  in  such  a  way  that  the  total  number  of  observed  failures  N  always  equals  the  expected 
number  of  failures  for  the  time  period  [/0.  '„)•   That  is,  a  *  is  defined  so  that 

W=  E{N\a*.  (3*)  =  a*(tr-  >D- 

Therefore,  there  is  no  difference  between  the  observed  versus  predicted  total  number  of 
failures.  The  goodness-of-fit  measure  must  therefore  be  based  on  the  differences  between  the 
observed  incremental  failures  N\,  N2,  . . .  ,  N„,  and  the  predicted  values 

(6.1)  E{Ni\a*,p*}  =  a*(tr-  t?-]),    i=\,2,...,n. 

Assuming  the  estimate  (5.5)  is  used  for  a  *,  the  likelihood  function  for  a  goodness-of-fit 
statistic  will  be  expressed  only  in  terms  of  /3*.  Since  the  NHPP  has  independent  increments, 
the  probability  that  a  given  failure  occurs  in  the  interval  [tt-.\,  f,)  is  the  expected  number  of 
failures  for  that  interval,  divided  by  the  total  number  of  failures,   This  is  written  as 

(6.2)  Pt  =  Pi(0*)=  [a*Ur-t?-\)\l[a*{f$~-4m)\,  /'-  1,2.  ...,  n. 

where  the  a  *  parameter  obviously  cancels  out.  The  likelihood  function  for  a  set  of  observed 
failures  N\,  N2,  ....  N„,  given  N,  is  therefore  the  multinomial 


TV 
N2,  .. 


■  H„  =  N, 


which  depends  only  on  /3  *.   The  parameter  a  *  can  be  regarded  as  a  scale  parameter  that  guaran- 
tees the  model  will  fit  the  total  number  observed  of  failures  N. 


544  S.A.SMITH  AND  S.SOREN 

We  now  show  how  the  goodness-of-fit  of  the  incremental  failure  data  can  be  measured  by 
the  Fisher  chi-square  statistic 

(6.4)  x2=  £  (N,~  NPl)2/NPi. 

The  use  of  this  statistic  as  a  goodness-of-fit  measure  is  based  on  the  following  theorem,  which 
has  been  restated  in  the  context  of  this  discussion. 

THEOREM  6.1:  Let  the  parameters  p]t  p2,  . . .  pfn  with  I/?,  =  1,  be  functions  of  a  param- 
eter /3  and  let  a  particular  value  /3'  be  determined  from 

(6.5)  0=  Y   (ty//>,)(3fl-/dj8)| 

~f  1/3  -  j3' 

Then  the  statistic  (6.4)  with  /?,  =  Pi(fi'),  i  =  1,2,  . . .  ,  «,  has  approximately  a  chi-square  distri- 
bution with  n  —  1  degrees  of  freedom  (x2(«  -  1))  for  large  N.  The  proof  of  this  result  is  quite 
lengthy  and  can  be  found  in  [6,  pp.  424-434]. 

To  apply  this  result  to  our  particular  problem,  we  must  show  that  (3'  equals  the  estimator 
/3*  defined  by  Equation  (5.6).  Using  /?,()3)  as  defined  in  Equation  (6.2),  and  differentiating 
with  respect  to  /3,  one  can  verify  that  Equation  (6.5)  reduces  to  Equation  (5.6).  Thus,  /3'  =  /3* 
and,  since  (5.6)  has  only  one  solution,  the  value  is  unique. 

The  chi-square  error  statistic  (6.4)  has  an  additional  intuitive  interpretation  for  this  appli- 
cation. Suppose  a  and  /3  are  the  "true"  parameters  of  the  underlying  nohomogeneous  Poisson 
process,  i.e.,  the  values  to  which  the  estimators  a  *  and  /3  *  must  eventually  converge  for  very 
large  sample  sizes.  Then  the  "true"  variance  of  the  number  of  observed  failures  in  [/,_,,  /,), 
i.e.,  the  limiting  value  for  the  sample  variance  of  a  large  number  of  observations,  is  given  by 

Var{fy|a,  0}  =  a(f£j  ~  tf)       i  =  \,  2,  ...  ,  n. 
Consider  W(a*,fS*)  =  £  (N,  -  E{N,\a  *,  j8  *})2/Var{/V,|a,  0}, 

which  is  the  sum  of  square  errors  between  the  observed  and  estimated  failures,  weighted  by  the 
true  variance  for  each  of  the  time  intervals.  Suppose  we  minimize  this  with  respect  to  a  *  and 
0*  by  solving  (3  W/ba*)  =  0  and  (d  W/Sp*)  =  0.  If  we  then  substitute  our  "best  estimates", 
a  *  for  a  and  /3  *  for  /3,  these  two  equations  reduce  to  the  maximum  likelihood  equations,  (5.5) 
and  (5.6),  respectively.  Birnbaum  [2,  p.  251-2]  also  points  that  if  we  minimize  the  chi-square 
statistic  (6.4)  with  respect  to  /3,  the  estimate  obtained  must  approach  the  estimate  0'  that 
satisfies  (6.5)  as  the  sample  size  approaches  infinity. 

This  goodness-of-fit  criterion  measures,  in  effect,  how  well  the  observed  data  fits  a  NHPP 
with  mean  value  function  M(t),  where  0  *  is  the  "best"  growth  parameter  for  the  observed  data. 
If  the  x2(«  _  1)  statistic  (6.4)  exceeds  the  critical  value  at  a  reasonable  significance  level,  such 
as  0.05  or  0.1,  the  model  should  be  rejected.  Since  Theorem  6.1  gives  only  an  asymptotic 
result,  it  is  important  to  discuss  the  sample  size  requirements  for  applying  it.  Given  the  popu- 
larity of  this  test,  there  has  been  considerable  experience  with  various  types  of  data.  A  com- 
mon criterion  is  that  TV  and,  in  this  case  the  time  points  r0,  t\,  ...  ,  tn,  must  be  such  that 
Npi  >  10  for  all  /.    (See  Birnbaum  [2,  p.  248]). 


RELIABILITY  GROWTH  OF  REPAIRABLE  SYSTEMS  545 

7.    APPLICATION  EXAMPLE 

As  an  illustration,  we  will  determine  the  estimators  a  *  and  /3  *  and  apply  the  goodness-of- 
fit  test  to  the  sample  data  in  Table  1.  We  assume  that  the  failures  of  the  system  were  only 
monitored  at  fixed  points  of  time  so  that  the  observed  data  consists  of  the  first  two  columns  of 
the  table.  These  data  points  were  generated  by  computer  simulation  with  failures  sampled  from 
a  NHPP  with  mean  value  function  M(t),  having  parameters  a  =  10.0,  /3  =  0.5.    Failure  times 


TV  T2, 

(7.1) 


from  this  distribution  can  be  generated  sequentially  from  a  set  of  random  samples 
.  from  the  uniform  distribution  by  means  of  the  transformation 


TM-  W- 


(l/a)log  £/,+,]1//3,    T0 
TABLE  1 


■  0,  1,  2, 


Time  Interval 

Observed 
Failures 

Predicted 
Failures 

Standard 
Deviation 

Normalized 
Error 

1 

400  -  800 

63 

78 

8.8 

2.88 

2 

800  -  1200 

63 

61 

7.8 

0.07 

3 

1200  -  1600 

54 

51 

7.1 

0.18 

4 

1600  -  2000 

51 

46 

6.8 

0.54 

5 

2000  -  2500 

68 

51 

7.1 

5.67 

6 

2500  -  3000 

49 

46 

6.8 

0.20 

7 

3000  -  3500 

34 

43 

6.6 

1.88 

8 

3500  -  4000 

39 

40 

6.3 

0.03 

9 

4000  -  4500 

39 

38 

6.2 

0.02 

10 

4500  -  5000 

43 

36 

6.0 

1.36 

11 

5000  -  5500 

39 

34 

5.8 

0.74 

12 

5500  -  6000 

36 

33 

5.7 

0.27 

13 

6000  -  6500 

28 

31 

5.6 

0.29 

14 

6500  -  7000 

22 

30 

5.5 

2.13 

15 

7000  -  7500 

35 

29 

5.4 

1.24 

16 

7500  -  8000 

32 

28 

5.3 

0.57 

17 

8000  -  8500 

22 

27 

5.2 

0.93 

18 

8500  -  9000 

19 

27 

5.2 

2.37 

19 

9000  -  9500 

19 

26 

5.1 

1.88 
23.25 

The  data  in  Table  1  was  used  to  obtain  maximum  likelihood  estimates  a*  and  /3  *  from 
Equations  (5.5)  and  (5.6).  This  was  done  by  calculating  various  values  of  the  right  hand  side 
of  (5.6)  as  a  function  of /3  until  the  minimizing  value  /3  *  was  determined  to  two  decimal  places. 
This  gave  /3  *  =  0.52  and  a  *  =  7.97,  where  a  *  was  determined  from  (5.5)  with  {$  *  =  0.52. 


The  accuracy  of  /3  *  is  reasonably  close  to  the  correct  value  /3  =  0.5,  but  the  estimate  of  a  * 
is  off  by  more  than  20%.  Other  calculations  with  different  sets  of  random  numbers  produced 
errors  in  both  directions  but  generally  resulted  in  an  a  *  error  several  times  larger  than  the  /3  * 
error,  on  a  percentage  basis.  This  seems  to  indicate  that  one  is  more  likely  to  estimate  slopes 
of  the  Duane  Plot  lines  accurately  than  to  estimate  the  intercepts  accurately  with  the  maximum 
likelihood  estimates.  Naturally,  as  the  number  of  observation  points  in  Table  1  is  increased, 
the  estimates  become  more  accurate.  Accuracy  was  not  improved  much  by  increasing  the 
number  of  time  points  from  20,  as  shown  in  the  table,  to  100  and  the  sign  of  the  error  for  a 
given  example  generally  did  not  change  as  the  number  of  observation  points  was  increased, 


546  S  A.  SMITH  AND  S.S   OREN 

while  holding  the  underlying  failure  points  fixed.    Bringing  the  estimate  a  *  to  within  5%  of  the 
correct  value  typically  required  300  to  500  observation  time  points  for  the  computed  examples. 

To  illustrate  the  use  of  the  goodness-of-fit  test  we  calculate  the  chi-square  statistic  (6.4) 
for  this  table.   The  "Predicted  Failures"  between  the  various  time  points  are  given  by 

Np,  =  a*(t,i\  -  t,H,       r  =  1,2,  ....  19. 

The  normalized  error  terms  as  in  (6.4)  are  given  by 

(N,-  Np,)2/{Np,). 

The  sum  of  these  errors,  when  compared  with  a  x2(18)  error  table,  is  less  than  the  critical 
values  25.99  and  28.87,  associated  with  significance  levels  0.1  and  0.05,  respectively. 

For  many  applications  of  the  model  it  is  more  important  to  predict  the  number  of  failures 
that  will  occur  in  the  next  time  period  than  to  obtain  accurate  estimates  for  a  and  /3.  In  such 
cases  the  estimators  obtained  from  10-20  time  points  appear  to  be  sufficiently  accurate.  This  is 
because  there  is  a  range  of  a,  /3  pairs  that  provide  almost  as  good  a  fit  to  the  observed  data  as 
the  optimal  ones  and  any  parameters  in  this  range  provide  a  satisfactory  predictive  model. 

To  illustrate  the  prediction  accuracy  of  the  estimates  /3  *  =  0.52,  a  *  =  7.97  obtained  from 
Table  1,  we  generated  simulated  failures  out  to  40,000  time  units.  The  number  of  failures 
predicted  by  extrapolating  with  the  estimated  parameters  and  with  the  true  parameters  are  com- 
pared in  Table  2.  The  errors  in  predicting  failures  caused  by  inaccuracy  in  estimating  the 
parameters  is  much  less  than  the  random  errors  that  occur  due  to  stochastic  variations  of  the 
failure  process.   This  was  found  to  be  the  case  in  several  similar  experiments. 

TABLE  2 


Time  Interval 

Simulated 

Estimated 

True 

Standard 

Failures 

Extrapolation 

Extrapolation 

Deviation 

9500  -  10,000 

24 

25 

25 

5.0 

9500-  15,000 

235 

251 

250 

15.8 

9500  -  20,000 

412 

443 

439 

21.0 

9500  -  30,000 

715 

766 

757 

27.5 

9500  -  40,000 

999 

1041 

1025 

32.0 

8.   CONCLUSION 


Choosing  the  fixed  time  points  between  which  to  tabulate  failures  is  mainly  a  question  of 
engineering  judgement.  The  time  points  might  be  selected,  for  example,  to  correspond  to  mile- 
stones in  the  reliability  development  program.  The  parameter  estimates  and  goodness-of-fit 
tests  obtained  in  this  paper  and  those  obtained  by  Crow  are  essentially  complementary  with 
respect  to  various  applications  of  the  Duane  model.  It  is  not  possible  to  determine  the  precise 
sample  size  at  which  one  approach  becomes  more  advantageous  than  the  other.  Based  on 
experience,  the  chi-square  goodness-of-fit  test  tends  to  reject  most  sample  data,  including  data 
that  fits  the  model,  when  sample  sizes  are  too  small.  Therefore,  rejection  of  the  model  by  the 
chi-square  test,  based  on  data  with  a  questionable  total  number  of  samples,  might  be  viewed  as 
inconclusive  and  the  more  accurate  test  developed  by  Crow  could  then  be  applied.  For  large 
sample  sizes  that  have  at  least  10  failures  between  time  points,  the  chi-square  test  should  be 
accurate  and  is  computationally  easier.    Data  that  fails  to  fit  the  NHPP  model  with  mean  value 


RELIABILITY  GROWTH  OF  REPAIRABLE  SYSTEMS  547 

function  M{t)  based  on  these  tests  requires  a  more  general  approach.  A  NHPP  model  with  a 
different  intensity  such  as  discussed  in  [1],  or  a  less  constrained  model  such  as  [13]  might  then 
be  tested. 

BIBLIOGRAPHY 

[1]  Ascher,  H.  and  H.  Feingold,  '"Bad  as  Old'  Analysis  of  System  Failure  Data,"  Proceedings  of 
the  Eighth  Reliability  and  Maintainability  Conference  (July  1969). 

[2]  Birnbaum,  Z.W.,  Introduction  to  Probability  and  Mathematical  Statistics  (Harper  and  Broth- 
ers, New  York,  N.  Y.,  1962). 

[3]  Barlow,  R.E.  and  F.  Proschan,  Statistical  Theory  of  Reliability  and  Life  Testing  (Holt, 
Rinehart  and  Winston,  Inc.,  1975). 

[4]  Chapman,  W.A.  and  D.E.  Beachler,  "Reliability  Proving  for  Commercial  Products," 
Proceedings  1977  Annual  Reliability  and  Maintainability  Symposium. 

[5]  Codier,  E.O.,  "Reliability  Growth  in  Real  Life,"  Proceedings  Annual  Symposium  on  Reliability 
(1968). 

[6]  Cramer,  H.  Mathematical  Methods  of  Statistics,  (Princeton  University  Press,  Princeton,  New 
Jersey,  1946). 

[7]  Crow,  L.H.,  "Reliability  Analysis  for  Complex  Repairable  Systems,"  Reliability  and 
Biometry:  Statistical  Analysis  of  Lifelength,  Society  for  Industrial  and  Applied  Mathemat- 
ics (SIAM),  Philadelphia,  Pennsylvania  (1974). 

[8]  Donelson,  J.,  "Duane's  Reliability  Growth  Model  as  a  Nonhomogeneous  Poisson  Process," 
Institute  for  Defense  Analysis  paper  P-1162  (April  1975). 

[9]  Duane,  J.T.,  "Learning  Curve  Approach  to  Reliability  Monitoring,"  IEEE  Transactions  on 
Aerospace,  2  (1964). 
[10]   Hovis,  J.B.,  "Effectiveness  of  Reliability  System  Testing  on  Quality  and  Reliability," 

Proceedings  1977  Annual  Reliability  and  Maintainability  Symposium. 
[11]  Mead,  P.H.,  "Duane  Growth  Model:  Estimation  of  Final  MTBF  with  Confidence  Limits 
Using  a  Hand  Calculator,"  Proceedings  1977  Annual  Reliability  and  Maintainability  Sympo- 
sium. 
[12]  Parzen,  E.,  Stochastic  Processes  (Holden  Day,  1962). 

[13]  Singpurwalla,  N.D.,  "Estimating  Reliability  Growth  (or  Deterioration)  Using  Time  Series 
Analysis,"  Naval  Research  Logistics  Quarterly,  25,  1-14  (1978). 


ON  THE  DISTRIBUTION  OF  THE 

OPTIMAL  VALUE  FOR  A  CLASS 

OF  STOCHASTIC  GEOMETRIC  PROGRAMS* 

Paul  M.  Ellnert  and  Robert  M.  Stark 

Department  of  Mathematical  Sciences 

University  of  Delaware 

Newark,  Delaware 

ABSTRACT 

An  approach  is  presented  for  obtaining  the  moments  and  distribution  of  the 
optimal  value  for  a  class  of  prototype  stochastic  geometric  programs  with  log- 
normally  distributed  cost  coefficients.  It  is  assumed  for  each  set  of  values 
taken  on  by  the  cost  coefficients  that  the  resulting  deterministic  primal  program 
is  superconsistent  and  soluble.  It  is  also  required  that  the  corresponding  dual 
program  has  a  unique  optimal  point  with  all  positive  components.  It  is  indicat- 
ed how  one  can  apply  the  results  obtained  under  the  above  assumptions  to  sto- 
chastic programs  whose  corresponding  deterministic  dual  programs  need  not 
satisfy  the  above-mentioned  uniqueness  and  positivity  requirements. 


1.   INTRODUCTION 

This  paper  is  concerned  with  deriving  the  distribution  and/or  moments  of  the  optimal 
value  for  a  class  of  stochastic  prototype  geometric  programs  in  which  a  subset  of  the  cost 
coefficients  are  lognormally  distributed.  The  programs  are  assumed  to  be  superconsistent  and 
soluble  for  all  positive  values  that  can  be  taken  on  by  the  cost  coefficients.  It  is  also  required 
that  the  dual  of  a  program  has  a  unique  optimal  point,  8f ,  with  all  positive  components,  for  all 
possible  values  that  can  be  taken  on  by  the  components  of  the  cost  vector  c.  Such  programs 
include  soluble  programs  with  no  forced  constraints.  Also  included  are  superconsistent  soluble 
programs  whose  forced  constraints  are  nonredundant  (and  hence  active  at  optimality)  and 
whose  forced  constraint  gradients  are  linearly  independent  at  optimality,  for  each  positive- 
valued  cost  vector  c. 

The  class  of  problems  specified  above,  while  of  interest  in  themselves,  can  be  used  to 
obtain  the  distribution  and/or  moments  of  the  optimal  value  for  more  general  classes  of  sto- 
chastic prototype  geometric  programs.   This  will  be  indicated  in  Section  6. 

The  distribution  and/or  moments  of  the  optimal  value  of  a  stochastic  program  will  be 
expressed  in  terms  of  the  density  function  of  a  vector  L  A  (log  K0,  log  Ku  ...  ,  log  Kd)\ 
where  log  denotes  the  natural  logarithm  function,  d  is  the  degree  of  difficulty  of  the  program, 
and 


*This  research  was  supported  in  part  by  the  Office  of  Naval  Research  Contract  N00014-75-C-0254. 
tNow  with  U.S.  Army  Materiel  Systems  Analysis  Activity,  Aberdeen  Proving  Ground,  Maryland. 


550  P    ELLNER  AND  R    STARK 

log  Kj  =  £  bj{J)  log  c,  for  y  €  {0,1,  ...  ,.«/}. 

In  the  above  (c^  ...  ,  c„)'  is  the  vector  of  cost  coefficients  (where  "'"  denotes  transpose)  and 
the  bf1 )  are  constants  that  are  independent  of  the  c,-. 

One  advantage  to  deriving  the  distribution  and/or  moments  of  the  optimal  value  in  terms 
of  the  density  function  of  L  is  that  the  vector  L  is  normally  distributed  when  the  stochastic  c, 
are  jointly  lognormally  distributed.  Furthermore,  under  certain  conditions,  it  is  reasonable  to 
expect  that  L  behaves  approximately  as  if  the  vector  of  stochastic  cost  coefficients  were  lognor- 
mally distributed  even  when  it  is  not.  More  precisely,  if  the  stochastic  cost  coefficients, 
[cj\i  €  /},  are  positive-valued  and  the  variates  {log  c,\i  €  /}  are  independent  with  finite  means, 
variances,  and  third  order  absolute  central  moments,  then  one  can  apply  a  central  limit  theorem 

for  random  vectors  to  the  relation  L  =  £  Z(/),  where  Z(,)  A  (6,(0)  log  ch  6,u)  log  c,,  ...  ,  b,(d) 

log  c,)'  [11].  Thus,  under  the  above  conditions,  one  might  expect  that  L  tends  to  be  normally 
distributed  provided:  the  stochastic  c,  are  positive-valued,  strictly  unimodal,  continuous  vari- 
ates; the  number  of  indices  in  /  is  "large"  in  comparison  to  d  +  1;  and  no  partial  sum  of  d  +  1 
of  the  Z(l)  is  "excessively"  dominant  in  the  sum  for  L. 

The  results  of  this  paper  should  be  of  interest  in  instances  where  the  operating  or  con- 
struction costs  associated  with  a  contemplated  project  or  engineering  system  can  be  adequately 
approximated  as  the  optimal  value  of  a  stochastic  prototype  geometric  program  with  lognormally 
distributed  cost  coefficients.  In  such  cases  a  knowledge  of  the  distribution  function  and/or 
moments  would  be  useful  as  a  predictive  tool  in  financial  planning.  For  instance,  if  the  distri- 
bution function  of  the  optimal  value  were  known  one  would  be  able  to  predict  with  a  given  pro- 
bability that  a  proposed  system's  operating  or  construction  costs  incurred  over  a  given  period 
would  lie  within  a  specified  set  of  limits. 

To  reflect  the  uncertainty  as  to  the  future  costs,  c,,  that  will  be  encountered  during  the 
construction  or  operating  period  of  interest  a  cost  analyst  often  subjectively  chooses  a  distribu- 
tion for  each  cost  c,.  Cost  analysts  have  frequently  found  families  of  positive-valued  random 
variables  that  are  continuous  and  strictly  unimodal  useful  for  this  purpose  [9].  The  lognormal 
random  variables  form  a  two  parameter  family  that  meets  these  specifications.  Recall  a  random 
variable  X  is  said  to  be  lognormal  iff  log  X  is  normally  distributed.  Properties  of  lognormal  ran- 
dom variables  can  be  found  in  [2]. 

Cost  analysts  are  most  often  concerned  with  the  distribution  of  values  of  c,  about  a  central 
value  and  not  with  tail  values.  Thus,  an  analyst  who  wishes  to  utilize  the  present  results  might 
proceed  to  express  his  uncertainty  about  the  future  value  of  cost  coefficient  c,  as  follows: 

1.  Assume  c,  is  lognormally  distributed  and  subjectively  choose  the  median  value  of  ch 
denoted  by  £,. 

2.  Specify  an  interval  of  interest  about  £,  of  the  form  (0,-1f /*  0,£)  where  0,  €  (1,  °°). 

3.  Subjectively  choose  8,  €  (0, 1)  such  that  1  -  8,  reflects  one's  belief  that  c,  €  (0,_1£,, 
0,£,);  i.e.,  the  more  confident  one  is  that  c,  €  {OfHi,  0,-f ,■)  the  closer  1  -  8,  should  be  chosen 
to  1. 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  551 

4.  Calculate  the  unique  value  of  the  standard  deviation  of  c,  that  is  consistent  with  (1) 
and  the  equation  /V(0,_1£,  <  c,  <  0 ,•£,•)  =  1-8,  where  Pr  denotes  the  probability  function 
associated  with  c,. 

Results  of  the  paper  do  not  require  that  the  stochastic  c,  be  independently  distributed. 
Thus,  for  every  pair  of  stochastic  cost  coefficients  ch  Cj  (/  ^  j)  the  analyst  may  subjectively 
choose  a  number  between  —1  and  1,  the  correlation  coefficient  p,7  of  log  c,  and  log  c,,  to  reflect 
his  opinion  as  to  the  interdependency  of  c,  and  Cj.  The  theory  allows  for  the  possibility  that 
pij  =  ±1  (i.e.,  with  probability  1  c,  =  acf  for  some  constants  a  €  (0,  °°)  and  /3  €  (— »,  oo)). 

In  Section  2  the  notation  used  in  connection  with  the  deterministic  and  stochastic 
geometric  programming  problem  and  its  dual  and  transformed  dual  is  presented.  Also  the  spe- 
cial role  of  the  transformed  dual  program  in  obtaining  the  distribution  and/or  moments  of  the 
optimal  value  of  the  primal  program  is  indicated. 

Section  3  presents  and  discusses  the  assumptions  placed  upon  the  primal  program 
throughout  Sections  3  through  5  and  the  appendices.  Additionally,  useful  properties  of  the 
density  functions  of  L  and  L  A  (log  Ku  . . .  ,  log  KdV  are  stated. 

In  Section  4  we  use  the  density  functions  of  L  and  Z,  together  with  the  maximizing  equa- 
tions for  an  unconstrained  transformed  dual  program,  to  obtain  the  density  functions  of  r  and 
(r,v(Pc)).  Here  r  denotes  the  random  vector  of  the  optimal  point  of  the  unconstrained  sto- 
chastic transformed  dual  program  and  \(PC)  denotes  the  optimal  value  of  the  stochastic  primal 
program.   We  then  obtain  the  density  function  of  \(PC)  as  a  marginal  density  of  (r,  v(Pc)). 

In  Section  5  we  use  the  density  function  of  r  to  derive  a  formula  that  expresses  each 
moment  of  v(Pc)  as  the  integral  of  an  explicitly  given  integrand  over  an  explicitly  specified  con- 
vex polyhedral  subset  of  Rd,  where  d  is  the  degree  of  difficulty  of  the  stochastic  primal  pro- 
gram. 

Section  6  briefly  indicates  how  the  preceding  results  can  be  used  to  calculate  the  distribu- 
tion and/or  moments  of  v(Pc)  when  Pc  need  not  satisfy  all  the  assumptions  of  Section  3. 

Appendix  A  contains  the  statement  and  proof  of  a  lemma  from  which  important  proper- 
ties of  L  and  L  immediately  follow.   These  properties  are  stated  in  Theorem  1  of  Section  3. 

Finally,  in  Appendix  B  we  establish  that  boundedness  of  the  dual  feasible  set  is  a 
sufficient  condition  for  the  existence  of  all  the  moments  of  v(Pc),  under  the  assumptions  of 
Section  3. 

2.   NOTATION  AND  PRELIMINARIES 

We  shall  now  review  the  definitions  and  notation  used  in  connection  with  prototype 
geometric  programming  that  will  be  utilized  in  the  paper.  In  the  following,  for  every  positive 
integer  v,  <v>  A  {1,  . . .  ,  v)  and  <v>  A  {0, 1,  . . .  ,  v).  Also,  for  every  matrix  P,  P'  denotes 
the  transpose  of  P.  All  elements  of  Euclidean  «-space,  R",  will  be  viewed  as  column  vectors  of 
n  real  numbers  and  the  zero  vector  will  be  denoted  by  0. 

Recall  a  prototype  primal  geometric  program  has  the  following  form  [4]:  inf  go(t)  subject 
to  gKU)  <1VkC  <p>  and  t,  >  0  V  /  €  <w>  where  t  =  (tu  ...  ,  tm)'  and  gKU)  A    £  c, 


552  P   ELLNER  AND  R.  STARK 

Y\  tj"  for  k  €  </?>.    In  the  above  A  =  (a,7)  is  an  n  by  m  matrix  with  real  entries  called  the 

exponent  matrix  and  c  =  (c1(  ....  c„)'  is  a  vector  of  positive  numbers  called  the  vector  of  cost 
coefficients.  Also,  JK  A  [mK,mK+x,  ...  ,  nK}  where  m0  =  1,  mK  =  a?k_!  +  1  for  k  €  <p>,  and 
«p  =  «.  The  constraints  gK(t)  <  1  are  called  forced  constraints  and  we  allow  the  possibility  that 
a  primal  program  has  no  forced  constraints. 

In  this  paper  we  shall  be  concerned  with  problems  of  the  above  form  when  some  or  all  of 
the  cost  coefficients  are  stochastic  variables  that  are  lognormally  distributed.  Thus,  we  shall 
assume  there  exists  /  C  <  n  >  such  that  /  ^  <t>  and  /  €  /  iff  c,  is  stochastic.  Let 
Q  A  (c/ ,  ....  Cj)'  where  i\  <  ...  <  /„  and  I  =  {iv\  v  €  <o>>}.  Thus,  c,  is  a  random  vector 
formed  from  the  stochastic  cost  coefficients.  Values  taken  on  by  c,  will  be  denoted  by  ch  Also 
c  will  denote  the  value  taken  on  by  cost  coefficient  vector  c  when  q  takes  on  the  value  c,.  We 
shall  let  Pc  and  P-  denote  the  corresponding  stochastic  and  deterministic  prototype  primal 
geometric  programs.  Furthermore,  v(/>?)  will  denote  the  optimal  value  of  P-  and  v(Pc)  will 
denote  the  stochastic  variable  that  takes  on  the  value  \(P?)  when  c,  takes  on  the  value  C/. 

The  stochastic  program  Pc  is  not  convenient  to  work  with  due  to  possible  randomness  in 
coefficients  of  the  forced  constraints.  To  find  computationally  tractable  bounds  on  the  solution 
of  a  two  stage  geometric  program  with  stochastic  cost  coefficients,  Avriel  and  Wilde  [3]  con- 
sidered the  stochastic  problem  Dc  in  place  of  Pc  where,  for  every  c  >  0,  D-  is  the  dual  of  P-  as 
given  in  [4].  The  stochastic  program  Dc  has  the  attractive  feature  that  all  its  randomness  is 
confined  to  the  objective  function.    To  see  this  recall  D-  is  the  following  program:  sup  JJ 

8        p  A    (8)  "° 

(c//8,-)  '  jj  \K(8)  "      subject  to  the  normality  condition  £  8,  =  1,  the  orthogonality  conditions 

£  Ay  8, •  =  0  for  j  €  <w>,  and  the  positivity  conditions  8,-  >  0  for  /'  €  <«>.  In  the  above, 
for  every  k  €  </?>,  \K(8)  A  £  8,  for  8  6  R".  Also,  in  evaluating  the  dual  objective  function 
one  uses  the  convention  that  xx  =  x~x  =1  for  x  =  0.    When  P-  has  no  forced  constraints  we 

P  ^    (g) 

set  p  =  0  and  define  the  expression  JJ  \K(8)  K      to  be  1. 

K=1 

Under  rather  general  conditions  one  has  \(D-)  =  v(P-)  for  c  €  /?>  [4,  Ch.  6]  (where  /?> 
denotes  the  positive  orthant  of  R"  and  v(D-)  denotes  the  optimal  value  of  D-).  This  is  true, 
e.g.,  if  P-  is  superconsistent  and  soluble  [4,  Ch.  4].  Thus,  frequently  the  distribution  function 
of  v(Dc)  will  be  the  same  as  that  for  v(Pc),  where  v(Dc)  denotes  the  stochastic  variable  that 
takes  on  the  value  v(D7)  when  c,  takes  on  the  value  c;.  Obtaining  the  distribution  function 
and/or  moments  of  v(Dc)  is  facilitated  by  the  fact  that  the  constraint  region  for  Dc  is  a 
polyhedral  convex  set  that  depends  only  on  the  nohstochastic  exponent  Matrix  A. 

Instead  of  working  directly  with  D-  we  shall  use  the  transformed  dual  program,  D-,  con- 
sidered in  [4,  Ch.  3].  Recall  D-  is  obtained  from  D-  by  solving  the  normality  and  orthogonality 
constraints  of  D-. 

In  what  follows  we  shall  assume  without  loss  of  generality  that  the  rank  of  A  is  m  and  that 
q  €  R"  is  not  in  the  column  space  of  A,  where  q,  =  1  if  i  <  n0  and  q,  =  0  if  /  >  n0  (see  [4, 
Ch.  3]).  As  in  [4]  we  define  d  to  be  the  dimension  of  the  solution  space  of  the  system  of  equa- 
tions A'b  =  0,  q'b  =  0.    (Recall  d  is  called  the  degree  of  difficulty  of  Pc  and,  under  the  above 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  553 

assumptions,  equals  n  -  m  —  1.)  Throughout  the  paper  we  assume  d  >  0.  (The  distribution 
problem  for  v(Pc)  when  d  =  0  has  been  studied  by  R.  Stark  in  [13]J  In  accordance  with  the 
terminology  in  [4],  we  define  N  A  {b(J)  |  j  €  <d>}  to  be  a  nullity  set  for  Pc  iff  Nis  a  basis  for 
the  solution  space  of  the  above  homogeneous  system  of  equations.  Also  6(0)  6  R"  is  called  a 
normality  vector  for  Pc  iff  A'b(0)  =  0  and  q'bm  =  1. 

Let  A7  A  {6(;)  |  j  €  <d>)  be  any  nullity  set  and  Z>(0)  be  any  normality  vector  for  Pc. 
Note  8  €  7?"  satisfies  the  orthogonality  and  normality  conditions  for  D-  iff*  8  =  6<0)  +  £  r,6(y) 

where  the  r,  €  i?1  are  uniquely  determined  by  8.    Thus,  by  replacing  8,  in  D-  by  6,(0)  +   £ 

j=  i 
tyi/     we  obtain  the  equivalent  transformed  dual  problem  D-: 

sup  K(c,bw)  YlKfobU*)  n8,-(r)~8'(r)  flXK(r)Kir) 

r  j=\  f=l  K=l 

subject  to  the  positivity  constraint  Br  ">  —  6(°}  where  r  A  (/-l7  ...  ,  r^)'  (the  vector  of  basic 
variables).    In  the  above   {K{c,b(J))\  j  €  <d>)   is  called  a  set  of  basic  constants  for  P- 

(corresponding  to  JVand  b(0))  where  K(c,b(l))  A  f\  c, '     for  y  €  <d>.    Also,  5  is  the  nby  d 

matrix   whose  yth   column   is    b(j)   for  j  €  <d>.     Finally,   for   /  €  <«>    and  k  6  <p>, 

d 

8,(r)  A  6,<0)  +  *£  r/6,0)  and  XK(r)  A    "£  8,(r).    When  P-  has  no  forced  constraints  we  define 


Note  that  the  parameters  in  D^  depend  on  the  choice  of  nullity  set  N  and  normality  vector 
6(0).  However,  as  v(ZXr)  =  v(Zt)  for  c  6  /?>  (where  v(DF)  denotes  the  optimal  value  of  ZXJ, 
the  optimal  value  of  £L  is  independent  of  the  choice  of  N  and  b{0}.  Thus,  for  any  nullity  set  N 
and  normality  vector  bm  for  /><.,  the  distribution  function  of  v(Dc)  is  identical  to  the  distribu- 
tion function  of  v(Dc),  where  v(Dc)  is  the  stochastic  variable  that  takes  on  the  value  v(D^) 
when  C/  takes  on  the  value  Z7. 

To  obtain  the  distribution  function  and/or  moments  of  v(Dc)  we  shall  first  obtain  the 
density  function  of  the  random  vector  L  A  (L0,L\,  ...  ,  Ld)'  and  L  A  (Lx,  ...  ,  Ld)'  where, 
for  j  €  <d> ,  Lj  is  the  random  variable  that  takes  on  the  value  log  K{~c,b(J))  when  ct  takes  on 
the  value  ct. 

3.   On  the  Density  Functions  of  L  and  L 

Unless  otherwise  stated,  throughout  the  remainder  of  the  paper  we  shall  assume  the  fol- 
lowing: 

(1)  [ev\v  6  <«>}  is  a  set  of  positive-valued  random  variables  such  that,  for  every 
/'  €  <n>,  Cj  =  ati  Y\  eviv  for  some  a,  €  (0,  °o)  and  fiiv  €  (— °°,°°),  v  €  <u>.  Further- 
more, it  is  assumed  that  (log  ex,  ...  ,  log  eu)'  is  a  nondegenerate  normal  random  vector  with 
mean  vector  /j,  =  (/i,,  . ..  ,  fxu)'  and  dispersion  matrix  A; 

(2)  There  exists  a  nullity  set  [b(j)\j  €  <rf>}  for  Pc  such  that  (20)|y  €  <d>]  is  linearly 
independent  where  s  '    A  fi'b(J)  for  y  €  <</>  and  /3  is  the  n  x  u  matrix  whose  (/,  j)  entry  is 


554  P.  ELLNER  AND  R.  STARK 

(3)  For  every  value  c7  that  c,  takes  on  the  program  P-  is  superconsistent  and  soluble; 

(4)  For  every  value  ct  that  c{  takes  on  the  program  D-  has  a  unique  optimal  point  8f  and 
8-  >  0. 

Many  of  the  results  obtained  under  the  above  restrictions  form  the  basis  to  approaches  for 
calculating  the  distribution  function  and/or  moments  of  \(PC)  under  less  restrictive  assump- 
tions.  This  will  be  briefly  indicated  in  Section  6. 

Assumption  (1)  allows  for  the  possibility  that  a  cost  coefficient  c,  is  constant  (J3iv  =  0  for 
all  v  €  <u>).  Also,  (1)  permits  one  to  conveniently  work  with  a  vector  of  stochastic  cost 
coefficients  ct  =  (c,y  . . .  ,  qj'  for  which  (log  c,|(  . . .  ,  log  ct)'  is  a  degenerate  normal  random 
vector.  Degeneracy  would  occur,  e.g.,  if  C\  and  c2  where  components  of  c,  such  that  c2  =  acf 
for  some  a  €  (0,  °°)  and  )8  €  Rl. 

To  evaluate  the  mean  /n,  and  variance  cr 2  of  log  e,  a  cost  analyst  could  apply  steps  (1) 
through  (4)  of  Section  1  to  et  in  place  of  c,.  After  choosing  £,-,  the  median  of  eh  and  the  vari- 
ance of  e,  by  these  steps  the  values  of  n,  and  cr,2  can  easily  be  calculated  [2]. 

Note  Assumption  (2)  is  satisfied  if  u  =  n  and  c,  =  e,  for  every  i  €  <w>.  Also,  if  there 
exists  a  nullity  set  of  Pc  that  satisfies  (2)  then  every  nullity  set  of  Pc  satisfies  (2)  (Proposition 
1). 

Recall,  for  c  €  /?>,  P-  is  called  superconsistent  iff  there  exists  t  €  Rm  such  that  t  >  0 
and    £   c,  W  t/'  <  1  for  every  k  €  <p>.    Also,  P-  is  called  soluble  iff  P-  has  an  optimal 

(€/„  7=1 

point.  It  can  easily  be  shown  that  P-  is  superconsistent  for  all  c  €  /?>  iff  there  exists  a  linear 
combination  of  the  columns  of  A,  say  x,  such  that  x,  <  0  for  all  i  6  7K,  k  €  <p>  [1,  p.  329]. 
Alternately,  one  can  show  that  P-  is  superconsistent  for  all  c  e  i?>  iff  the  set 
{8  €  /?"|  8,-  ^  0  V  /'  €  <aj>,  ^'8  =  0,  and  q'8  =  1}  is  bounded  [1,  p.  329].  Moreover,  if  the 
above  set  is  bounded  and  contains  a  point  8  >  0  then  P^  will  be  superconsistent  and  soluble  for 
every  c  €  /?">  (by  [4,  p.  120,  Th.  2]  and  [1,  p.  329]). 

Assumption  (3)  implies  that  \(P-)  =  v(D-)  for  every  value  c,  taken  on  by  c,  ([4,  p.  117, 
Th.  1]). 

Assumption  (4)  holds  for  c  €  i?>  if  P-  is  soluble  and  has  no  forced  constraints.  More 
generally,  one  can  show  (4)  holds  at  c  €  /?>  if  P^  is  a  superconsistent  soluble  program  whose 
forced  constraints  are  nonredundant  and  whose  forced  constraint  gradients  are  linearly  indepen- 
dent at  optimality.  By  nonredundant  we  mean  that  the  optimal  value  of  P?  is  greater  than  the 
optimal  value  of  PK  -  for  every  k  €  <p>,  where  PK  -  denotes  the  program  obtained  from  P7 
by  deleting  forced  constraint  k. 

If  the  components  of  L  form  a  set  of  independent  random  variables  then  we  obtain  a 

d 

simpler  formula  for  the  density  function  of  L  since,  in  this  case,  g{lx,  ...  ,  ld)  =  JJ  £/(//) 

j-\ 
where  g  is  the  density  function  for  L  and  gj  is  the  density  function  for  £,.   If,  in  addition,  L0  is 
independent  of  the  components  of  L  then  the  calculation  of  moments  of  v(Pc)  is  simplified. 
This  follows  from  the  fact  that  one  can  express  v(Pc)  as  the  product  e  °a)(r)  where  o>  is  a 
known  function  of  a  ^-dimensional  random  vector  r  whose  density  function  can  be  calculated 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  555 

from  that  of  L.  Thus,  when  L0  is  independent  of  L  we  have  F'CvC^))  =  [Fie10)]  [£"(w(r))] 
where  £"(0)  denotes  the  i>th  moment  of  random  variable  Q  (whenever  this  moment  exists). 
If  L0  is  a  linear  function  of  the  components  of  L  one  can  obtain  a  function  a>  of  r  such  that 
v(Pc)  -  <w(r)  from  which  one  can  calculate  £"(v(Pc)).  It  will  be  shown,  under  the  previously 
listed  assumptions,  that  one  can  always  find  a  nullity  set  [biJ)\j  €<d>}  and  normality  vector 
b{0)  for  Pc  such  that  L0  is  independent  of  L  if  s(0)  £  span  {s^l./  €<</>}  and  L0  is  a  linear 
function  of  the  components  of  L  if  s(0)  €  span  [sU)\j  €  <d>)  where  s(0)  A  /3'£(0)  and  bi0)  is 
any  normality  vector  of  Pc. 

Theorem  1  indicates  how  to  obtain  a  nullity  set  for  Pc,  [biJ)\j  €  <d>},  such  that  the 
components  of  the  corresponding  random  vector  L  are  independent  normal  variates  whose 
means  and  variances  are  known.  Also,  using  the  above  nullity  set,  it  is  shown  how  to  obtain  a 
normality  vector  for  Pc,  bm,  such  that  if  s(0)  #  span  {s(j)\j  £  <d>)  then  the  components  of 
the  corresponding  random  vector  L  are  independent  normal  variates  whose  means  and  vari- 
ances are  known.  The  proof  of  Theorem  1  follows  immediately  from  Lemma  A  which  is  stated 
and  derived  in  Appendix  A.  The  proof  of  Lemma  A  uses  the  eigenvectors  of  the  dispersion 
matrix  A.  Fortunately,  however,  the  calculation  of  the  above-mentioned  nullity  set  and  nor- 
mality vector  and  the  calculation  of  the  means  and  variances  of  the  corresponding  variates  £,, 
)  €  <d>,  do  not  require  any  eigenvector  or  eigenvalue  calculations. 

THEOREM  1:  (a)  Define  {b{J)\j  €  <d>)  inductively  by  6(1)  A  £(1)  and,  for 
I  <  j  ^  d,  b{J)  AbiJ) -^  «j87>(/),  P'b(n  >  A)~x  «P'b(J\  p'bU)>x)b{'\  where 
<x,y>\  A  x'Ay  for  (x,y)  €'/?"  x  R".    Then  [bij)\j  €  <d>}  is  a  well-defined  nullity  set  of 


(b)  Define  6(0)  A  bi0)  -  ]T  «/3'6(/),  p'ba)>A)-x  «j8'£0),  p'bU)>x)bU)  if  s{0)  €  span 
[s{j)\j  €<d>],  bm  A  b(0)  otherwise.  Then  bi0)  is  a  well-defined  normality  vector  of  Pc. 

(c)  For  every  j  €  <d>  let  Lj  denote  the  random  variable  that  takes  on  the  value  log 
Kj(c,b(j))  when  c7  takes  on  the  value  c,.  Also,  define  LA  (Lu  ....  Ld)'  and 
L  A  (Lq,L\,  ...  ,  Ld)'.  Then  I  is  a  normal  random  vector  with  independent  components. 
Additionally,  L  is  a  normal  random  vector  with  independent  components  if  S(0)  £  span 
{s{J)\j  €  <d>}. 

(d)  For  every  j  €  <d>  let  gj  denote  the  density  function  of  Lj.  Then  gj(0= 
(T)7V27r)_1  exp  (—  (2t) j)~l(l  —  i>j)2)  for  every  I  €  Rl,  where  vj  is  the  expected  value  of  L,  and 

rij  is  the  variance  of  L}.    Furthermore,  v} .=  <ix,p'bij)>  -   JT  a,6,0)  and  i\)  =  </8'60), 

-  ,= ' 

P'bij)>A  for  every  y  €  <d>,  where  <  \   >  denotes  the  usual  inner  product  on  R". 

Throughout  the  remainder  of  the  paper  we  define  b^j)  and  L,  for  j  €  <d> ,  L,  and  Z  as 
in  Theorem  1.   We  also  denote  Kj{c,b(j))  by  Kj(c)  for  j  £  <d>. 

We  shall  now  show  that  if  there  exists  a  nullity  set  of  Pc  that  satisfies  Assumption  (2) 
then  every  nullity  set  of  Pc  must  satisfy  (2). 

PROPOSITION  1:  If  Assumption  (2)  holds  then  for  any  nullity  set  [bQ)\j  €  <d>]  of 
Pc  the  set  [s{J)\j  €  <d>)  is  linearly  independent  where  slJ)  A  fi'b(J)  for  every  j  €  <rf>. 


556  P.  ELLNER  AND  R.  STARK 

PROOF:  Let  [t>  J>\j  €  <d>)  be  a  nullity  set  for  Pc  such  that  [s^lj  6  <d>)  is  linearly 
independent  where  sU)  A  /3'£0)  for  every  j  €  <d>.  Let  5  A  span  {60)|y  €  <</>}  and  S  A 
span  {S0)|y  €  <d>).  Define  T  to  be  the  unique  linear  transformation  for  B  to  S  such  that 
r(£0))  -  50)  for  every  j  €  <rf>.  Thus,  since  B'b(j)  -  s0)  for  every  j  €  <rf>,  one  has 
T(b)  =  B'bforaW  b  e  B. 

Since  {S0)|y  €  <d>}  is  a  nullity  set  for  Pc  one  has  span  {£0)|y  €  <</>}«  A  Thus, 
for  every  y  €  <tf>,  T(bU))  =  /3'S0)  -  50).  Note  T  is  an  isomorphism  from  B  onto  5  and 
{50)|y  €  <</>}  is  linearly  independent.   Hence,  {50)|y  €  <</>}  is  linearly  independent. 

Next  we  consider  the  assumption  s(0)  #  span  {S(7)|y  €  <</>}. 

PROPOSITION  2:  Assume  5(0)  $  span  {s0)|y  €  <</>}.  Then  for  any  nullity  set 
{bij)\j  6  <rf>}  and  normality  vector  6(0)  for  Pc  One  has  5(0)  #  span  {50')ly  €  <</>}  where 
50)  A  /3'60)  for  every  j  £  <d>. 

PROOF:  Define  B  A  span  (£0)|y  €<</>}  and  S  A  span  {30)|y  6  <</>}.  Let  Tbe 
the  unique  linear  transformation  from  B  to  S  for  which  T(b(j))  =  5(7)  for  every  y  €  <d>. 
Since  j8'60)  =  s0)  for  every  y  €  <</>  one  has  f(b)  =  B'b  for  all  b  €  £. 

Since  {£0)|y^€  <</>}  is  a  nullity  set  of  Pc  one  has  span  (50)|y  €<</>}  C  B.  Also, 
since  6(0)  and  b{0)  are  normality  vectors  of  Pc  it  follows  that  £(0)  -  b{0)  €  span 
(S^l/  €  <«/>},  i.e.,  6(0)  €  B.  Thus,  for  every  y  €  <d>,  b{jJ  €  5  and  hence,  7-(£0))  = 
B'b(J)  =  s^K  Finally,  observe  f  is  an  isomorphism  from  B  onto  5  since  s(0)  C  span 
{s(7)|y  €  <<i>}  and  {s(;)|y  G  <</>}  is  linearly  independent  by  Assumption  (2).  Moreover, 
[b^^lj  €  <d>)  is  linearly  independent.   Thus  {S^ly  €  <d>)  is  linearly  independent. 

As  mentioned  earlier,  when  u  =  n  and  c,  =  e,  for  every  i  €  <  «  >  then  Assumption  (2) 
holds.  In  addition  one  has  s(0)  #  span  {S^ly  €  <d>)  and  hence  by  Theorem  1  the  com- 
ponents of  L  are  independent.  We  next  consider  the  case  where  Assumption  (2)  holds  but 
S(0)  €  span  {S0)|y  6  <d>). 

PROPOSITION  3:  Assume  s(0)  €  span  {S0)|y  6  <d>).  Then  there  exist  v,  €  Rx  for 
j  €  <d>  such  that  s(0)  =   Z  yJs{J)  where  s0)  A/8'60)  for  every  y  €  <tf>.    Furthermore, 

d  n    [       —        d  ] 

^o  -  Z  -^7  +  ^  where  ^  is  the  constant  £    ft/0)  -  Z  J>*r     lo8  *i- 

7=1  /-i  I  7-i  J 

PROOF:    Since  s(0)  €   span  {s0)|y  €  <</>},  by  Proposition  2  there  exist  v.  €  /J1  for 

</  d  J 

j  <e  <</>  such  that  5(0)  =  z  yjs  •  Thus  5/0)  ■  Z  >y**    for  every '  €  <M> 


By  Lemma  A,  Part  (Hi),  L0  A  log  AT0(c)  =  log     J]  a,'      (exp(L(e,5(0)))}    =  Z  bi     lo6 
a,+  £     5/0)     l08     e-=Z     */0)     l08    a-+Z       Z^/0)      l08    e'=Z    */0)    l08    «/+Z 

U  Z*/0)l°8e' -    Z   */(0)   lQg  «/+Z  ^U50))=    z   */(0)   lQg  «<+Z  yj 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  557 

\Lj  -  £  b,0)  log  a,  =  x  yjLJ '+  D  where  D  k  Z  *<(0)  loe  <*• >-  Z  y\  Z  6«0)  log «-  =  Z 

I  /-i  J        y-l  ,-i  y-i       [,-i  J        ,=i 

\btm-  ZM0)    log«,. 

We  next  consider  the  rf-dimensional  random  vector  r  A  (ru  ....  r^)'  mentioned  earlier 
whose  density  function  can  be  used  to  obtain  moments  of  \(PC).  To  define  r  assume  C/  takes 
on  the  value  c,.  Then,  by  Assumption  (4),  D-  has  a  unique  optimal  point  8?.  Since 
{b^^j  €  <d>)  is  a  nullity  set  and  bi0)  is  a  normality  vector  of  Pc  there  exists  a  unique  point 
r-c  A  (/^(c),  ...  ,  ^(c))'  €  Rd  fox  which  8?  =  ft(0)  +  £  (/-,(c))ft0>.   We  define  r to  be  the  ran- 

dom  vector  that  takes  on  the  value  r-  when  c{  takes  on  the  value  c{.  In  the  next  section  we 
shall  obtain  the  density  function  of  r  from  that  of  L.  Also,  when  s(0)  £  span  {s0)|./  €  <d>}, 
we  shall  obtain  the  density  function  of  v(Pc)  as  a  marginal  density  of  (v(Pc),r).  The  density 
function  of  (v(Pc),r)  is  obtained  from  that  of  L. 

4.   THE  DENSITY  FUNCTIONS  OF  rand  v(/»c) 

Assume  C/  takes  on  the  value  c,.  Since  8-  is  an  optimal  point  for  D-  with  all  positive 
components  (Assumption  (4))  it  follows  that  8F  satisfies  the  maximizing  equations  for  D-  [4,  p. 
88,  Th.  3].  Expressing  8-  in  terms  of  the  components  of  r-,  the  maximizing  equations  can  be 
written  in  the  form  log  Kj(c)  =  hj(r-)  for  every  j  €  <d>  where,  for  j  €  <d>,  hj  is  the 
function  defined  in  Theorem  2.  The  above  equations  will  be  used  to  obtain  the  density  func- 
tion of  r  from  that  of  L.  From  the  above  maximizing  equations  one  can  easily  show  that  the 
optimal  value  of  P?  satisfies  the  equation  log  K0(c)  =  log  (\(P7))  +  h0(r7)  where  h0  is  defined 
as  in  Theorem  2  [4,  p.  88,  Th.  3].  This  equation,  together  with  the  maximizing  equations,  will 
be  used  to  obtain  the  density  function  of  (v(Pc),r)  from  that  of  I  when  s(0>  $  span 
\s(j)\j  €  <d>). 

To  obtain  the  density  functions  of  rand  (\(Pc),r)  we  shall  first  define  the  functions  hj  for 
j  €  <d>  and  establish  several  of  their  properties. 

THEOREM  2:  Let  H  A  [r  €  Rd\Br  >  -_b(0)}  where  B  is  the  n  x  d  matrix  whose  /th 
column  is  b{J)  for  j  €  <d> .   For  every  j  €  <d>  define  hy.  H  — -  Rl  such  that,  for  r  €  H, 

fij(r)  A  £  bt(J)  log  8,(r)  -  £  \K0)  log  \K(r) 

P  d 

(where  £  x*0)  1o8  M')  A  0  if  />  =  0).  In  the  above,  for  every  r  €  Rd,  8,(r)  A  ft/0'  -I-  £ 
r/6,C/)  for  ;  €  <n>  and  XK(r)  A  £  8,(r)  for  k  €  <p>.  Also,  for  every  y  €  <d>  and 
k  €  <p>,XK0)  A   £  ft,0). 

For  every  j  €  <rf>  define  hy.  (0,«>)  x  //  —  Rd+i  such  that/for  (z,r)  €  (0,«>)  x  7/, 
ft,(z,r)  A  hj(r)  if  j  £  <d>  and  //oUr)  A  log  z  +  h0(r). 

Finally,  define  //:  H  ^  Rd  and  fe  (0,°o)  x  H  —  flrf+1  such  that,  for  every 
fe,r)  €  (O.oo)  x  H,  h(r)  A  (A^r),  ....  /*»)'  and  /J(z,r)  A  (A0(z,r),A1(r,r),  ....  Arf(z,r))'. 
Then 


558  r.  ELLNER  AND  R.  STARK 

(a)  h  and  h  aid  vontinuously  differentiable  in  //and  (O.oo)  x  //respectively; 

(b)  A  is  1-1  in  //and  A  is  1-1  in  (0,«>)  x  //; 

(c)  h  is  onto  /?rf; 

(d)  If  S(0)  tfspan  [sU)\j  €  <</>}  then  Ais  onto /Jrf+1. 

PROOF^  (a)  Clearly,  for  every  y  €  <rf>,  r/7  is  continuously  differentiable  in  //and,  for 
every  j  €  <d>,  hj  is  continuously  differentiable  in  (0,  oo)  x  H.  Thus,  /i  is  continuously 
differentiable  in  //and  h  is  continuously  differentiable  in  (0,<»)  x  H. 

(b)  Let  r  and  5  be  elements  of  //such  that  Mr)  =  h(s).  Note  8(r)  >  0  and  8(s)  >  0. 
Also,  since  (log  Kx(c),  ....  log  A^(c))'  is  a  nondegenerate  rf-dimensional  normal  random  vec- 
tor, c,  takes  on  a  value,  say  q,  for  which  log  Kj(c)  =  hj(r)  =  hj(s)  for  every  j  €  <d>. 
Thus,  by  definition  of  A!,-  and  r/7  for  j  €  <d>,  8(r)  and  8(s)  satisfy  the  maximizing  equations 
for  D-.  Also,  8(r)  and  8(s)  are  feasible  points  of  D-.  Thus,  by  [4,  p.  88,  Th.  3],  8(r)  and 
8(s)  are  optimal  points  of  D-.  However,  by  Assumption  (4),  D-  has  only  one  optimal  point 
and  hence  8(r)  =  8(s).  This  implies  r  =  s  since  the  nullity  set  {b{j)\j  €  <d>)  is  linearly 
independent. 

Next,  let  (z1(r)  and  (z2,s)  be  elements  of  (0,°o)  x  //such  that  h(zur)  =  h(z2,s).  Then, 
Mr)  =  Ms),  and  hence,  r  =  5.   Also,  ^oU^r)  =  h0(z2,s).   Hence,  by  the  definition  of  h~0, 

log  z,  =  h0(z{,r)  -  £  6,(0)  log  8,(r)  +  £  AK(0)  log  Mr) 
=  ft0(z2,s)  -  X  V0)  log  8,(5)  +  £  AJ0'  log  A«(s) 


and  thus  z,  =  z2. 

(c)  Let  w  €  Rd.  Since  (log  A^tc),  ...  ,  log  Kd(c))'  is  a  nondegenerate  rf-dimensional 
normal  random  vector,  c,  takes  on  a  value,  say  C/,  for  which  log  A}(c)  =  u}  for  every 
j  €  <d>.    By  Assumption  (4),  D-  has  an  optimal  point  8  such  that  8  >  0.    Let  r  be  the 

d 

unique  element  of  tf^for  which  8  =  6(0)  +  £  r,*"'  and  denote  8  by  8(r). 
y-i 

Since  8(r)  >  0  one  has  r  £  H.  Furthermore,  since  8(r)  is  an  optimal  point  of  D-,  by  [4, 
p.  88,  Th.  3]  8(r)  satisfies  the  maximizing  equations  for  D-.  Thus,  for  every  j  6  <</>, 
hj(r)  =  log  /^(c)  =  Uj.    Hence,  h  is  onto  Rd. 

(d)  Assume  v<0)  g  span  {50)|y  €  <</>}.  Let  u  =  («,,  ...  ,  ud)'  <E  Rd and  u0  €  /J1.  By 
Theorem  1,  (A"0(c),  #,(<:),  ....  ATrf(c))'  is  a  nondegenerate  (</  +  l)-dimensional  normal  ran- 
dom vector.  Thus,  c,  takes  on  a  value,  say  Z7,  for  which  log  Kj(c)  =  u}  for  every  j  €  <d>. 
By  Assumption  (4),  /)-  has  an  optimal  point  8  such  that  8  >  0.   Let  r  be  the  unique  element  of 

d 

RdfoT  which  8  =  b(0)  -I-  £  rfb<J)  and  denote  8  by  8(r).   Let  r0  A  v(D?). 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  559 

Since  8(r)  >  0  and  v(D-)  >  0  one  has  (r0,r)  €  (0,<»)  x  H.  Also,  since  8(r)  is  an 
optimal  point  of  D^,  by  [4,  p.  88,  Th.  3]  8(r)  satisfies  the  maximizing  equations  for  D-.  Thus, 
for  every  j  €  <d>, 

(1)  hj(r0,r)  =  hj(r)  =  log  Kj(c)  =  Uj. 

Also,  by  [4,  p.  88,  Th.  3],  since  8  0)  satisfies  the  maximizing  equations  for  D?  one  has 

r0  =  v(Z)?)  =  K0(c)  n  8,(r)"*'(0)  f\  Kir)^ '. 
Thus,  log  r0  =  log  tf0(c)  -  £  6,(0)  log  8,(r)  +  £  XK(0)  log  (XK(r)),  i.e. 


(2)  h0(r0,r)  =  log  K0(c)  =  «0. 

By  (1)  and  (2)  his  onto  Rd+X. 

Note  by  Theorem  2,  for  every  /  €  Rd  there  exists  a  unique  point  r,  €  H  such  that 
/=  h(r,).  Thus,  we  can  define  h~l:  Rd -^  H  by  h~x[l)  A  r,  for  /  €  /?<*.  Also,  if  s(0)  tf  span 
{s^l  j  €  <</>}  then  by  Theorem  2,  for  every  /  6  Rd+X  there  exists  a  unique  point 
(z7,r7)  €  (O.oo)  x  //such  that  7=  /t(z7,r7).  Thus,  when  s(0)  <7  span  {S0)|j  €  <</>},  we  can 
define  h~l:  Rd+X  —  (0,°°)  x  //by  A_1(7)  A  (z7,r7)  for  7  €  /^+1. 

PROPOSITION  4:    (a)  r  =  h~xU); 

(b)    (v(Pc),r)  =  A_1(L)  if  S(0)  g  span  {s0)|y  €  <</>}. 

PROOF:    (a)  Let  c7  take  on  the  value  c7.   Then  L  takes  on  the  value  /  A  (log  Kx(c), 
log  #</(?))'.    By  Assumption  (4)  and  [4,  p.  88,  Th.  3]  one  has  r-  €  //and  log  Kj(c)  =  fcy(r?) 
for  every  j  €  <d>.    Thus,  h~x(l)  =  r-.    Hence,  h~x(L)  takes  on  the  value  r?  when  C/  takes 
on  the  value  c7,  i.e.,  h~x(L)  =  r. 

(b)  Assume  s(0)  g  span  {s0)|y_€  <</>}.  Then  h'x:_Rd+x  —  (0,«>)  x  //is  well-defined. 
Let  C/  take  on  the  value  c7.  Then  L  takes  on  the  value  /A  (log  K0(c),  log  AT^c),  ...  ,  log 
Kd(c))'.  Note  v(Pc)  takes  on  the  value  v(/>f)  >  0.  Also,  by  Assumption  (4)  and  [4,  p.  88, 
Th.  3]  one  has  r-  €  //and  log  Kj(c)  =  hj  (v(P-),  r-)  for  every  /  <E  <</>.  Thus,  A-1  (7)  = 
(v(P?),  r?).  Hence,  h~x(L)  takes  on  the  value  (v(/>?),  r?)  when  c7  takes  on  the  value  c7,  i.e., 
A-HZ)  =  (v(Pc),r). 

We  can  now  obtain  the  density  function  of  r. 

THEOREM  3:  Let  t/»  denote  the  density  function  of  r  and  g  denote  the  density  function 
of  L.   Then 

JO  if  r  g  H, 

x,,ir)  =  \{g(h(r)))(\detDh(r)\)     if  r  €  H, 

where  Dh(r)  denotes  the  derivative  of  h  at  r. 


560  P    ELLNER  AND  R.  STARK 

PROOF:   Let  ct  take  on  the  value  c,.   Then  r  takes  on  the  value  r-  A  (r,,  . . .  ,  rd)'  where 

d  — 

h-=  b{0)  +  £  r^'7'  is  the  unique  optimal  point  of  D-.    By  Assumption  (4),  8?  >  0.    Thus, 

M 
Br-  >  —b(0\  i.e.,  r-  €  H.   Hence,  rcan  only  take  on  values  in  H.   Thus,  <//(r)  =  0  if  r  &  H. 

Let  B  be  an  open  Borel  subset  of  H.   Note,  by  Proposition  4, 

Pr(r  6  £)  =  Pr(h~xU)  £  B)  =  Pr(L  €  /?(£)). 

By  Theorem  2,  A  is  1-1  and  continuously  differentiable  in  B.    Also,  g  is  integrable  on  /;(/?) 
since  g  is  the  density  function  of  L.   Thus, 

ML  €  h(B))'JHB)g  =  fB  (g  °h)\det  Dh  | 

[12,  Ths.  3-13  and  3-14],  where  g  °h  denotes  the  composition  of  g  and  h.   Hence,  Pr(r  €  B)  = 
Jg  (#  °/?)|det  D/?|.   This  implies  ^/(r)  =  {g(h(r))}(\dQt  Dh(r)\)  for  r  €  //. 

Next  we  obtain  the  density  function  of  (v(Pc),r). 

THEOREM  4:    Let  </)  denote  the  density  function  of  (v(Pc),r)  and  g  denote  the  density 
function  of  L.   Assume  s<0)  £  span  [s^'\j  €  <d>).   Then 

|  0  if  (z,r)  <?  (0,°o)  x  H, 

|  {g(Mz,r))}(|det  LVKz,r)|)     if  (z.iO  €  (0,°°)  x  //, 
where  Dh(z,r)  denotes  the  derivative  of  h  at  (z,r). 


Hz,r)  =  \ 


PROOF:  By  the  proof  of  Theorem  3  it  follows  that  (v(Pc),r)  can  only  take  on  values  in 
(0,°o)  x  H.    Hence,  i/)(z,r)  =  0  if  (z,r)  Q  (0,~)  x  H. 

Let  B  be  an  open  Borel  subset  of  (0,°o)  x  //and  define  z  A  v(Pc).  Note  by  Proposition 
4,  Pr((z,r)  6  £)  =  Pr(h~x{L)  €  B)  =  Pr(L  €  £(£))._  By  Theorem  2,  /?  is  1-1  and  continu- 
ously differentiable  in  5.  Also,  £  is  integrable  on  h(B)  since  g  is  the  density  function  of  L. 
Thus,  ML  6  h(B))  =   f-  -    £  =   f-   (g  °A)|det  M|   [12,  Ths.  3-13  and  3-14],  where  £  °h 

J  h(B)  J  B„  p  _ 

denotes  the  composition  of  g  and   h.    Hence,   Pr((z,r)  €  B)  =  J .    (g  °h)    \del  Dh\.    This 
implies  4>(z,r)  =  {g  (h(z,r))}  (\det  Dh(z,r)\)  for  (z,r)  €  (0,«»)  x  H. 

When  \(0)  #  span  {.?(/)|y  €  <d>)  the  above  theorem  immediately  yields  the  density 
function  of  v(/*f). 

COROLLARY  4.1:  Let  /denote  the  density  function  of  v(Pc)  and  assume  s(0)  ?  span 
{s(n\j  €  <d>}.   Then 


/(*)- 


0  if  z  ?  (0,oo), 

/rew{£(A(z,r))}(|detZM(z,r)|)dr  if  z  €  (°-00)- 


Observe  that  to  evaluate  /at  z  €  (0,«>)  by  Corollary  4.1  one  must  integrate  a  specified 
function  over  the  convex  polyhedral  set  H  =  [r  6  flrf|5r  >  -b{0)}.  When  the  degree  of 
difficulty  d  equals  1  then  //will  be  an  interval  in  Rl  whose  end  points  can  easily  be  obtained. 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  561 

Thus,  when  d  =  1,  one  can  accurately  approximate  f(z)  by  applying  a  quadrature  formula  to 
evaluate  the  integral  expression  for  f(z).  However,  the  quadrature  rule  must  be  modified  as  in 
[7,  Ch.  7,  Sec.  6.2]  to  allow  for  the  fact  that  the  integrand  is  not  defined  at  the  end  points  of 
the  interval  of  integration. 

When  d  >  1  the  effort  and  expense  of  devising  and  applying  a  quadrature  scheme  to 
approximate  the  integral  expression  for  f(z)  to  a  high  degree  of  accuracy  may  not  be  justified 
since  frequently  the  distributions  chosen  for  the  stochastic  c,  will  be  subjectively  determined. 
In  such  cases  a  numerical  Monte  Carlo  method  could  be  an  attractive  alternative  for  approxi- 
mating the  multiple  integral  used  to  express  f{z)  [6,  14,  15]. 

Finally,  under  the  assumption  of  Corollary  4.1  note  the  distribution  function  of  v(Pc), 
denoted  by  F,  is  given  by 


F(y)- 


0  if^O, 

j;^€(0l>)^{^^))f(Jdeti>AUr)|)^Z  .fy  >  Q 


Thus,  if  great  precision  is  not  required  a  numerical  Monte  Carlo  technique  could  be  attractive 
for  approximating  F(y)  as  well  as  f(z). 

5.   THE  MOMENTS  OF  v(Pc) 

In  the  following,  for  each  random  variable  Q,  recall  E(v){Q)  denotes  the  moment  of  order 
v  o{  Q  whenever  it  exists,  where  v  6  N  (the  set  of  positive  integers).  Also,  let 
E{Q)  A  E{l)(Q). 

Throughout  Section  5  we  assume  EM(\(PC))  exists  for  every  v  €  N.  Proposition  B  in 
Appendix  B  establishes  that  boundedness  of  the  dual  feasible  set  FA  (8  6  ^"|^'8  =  0, 
q'8  =  1,  8,  >  0  v  i  €  <n>)  is  a  sufficient  condition  for  the  above  moments  to  exist.  Furth- 
ermore, one  can  show  P-  is  superconsistent  for  every  c  €  /?>  iff  F  is  bounded  (see  p.  554). 

To  calculate  the  moments  of  v(Pc)  it  is  advantageous  to  use  the  density  function  of  r 
instead  of  that  for  v(Pf).   To  obtain  the  moments  of  v{Pc)  in  terms  of  the  density  function  of  r 


8  ,-(/■)     '     if  p  =  Oando)(r)  =  f[8,(r)     '      f[  XK(r)  K    if  p  >  0,  for  r  €  H. 

PROPOSITION  5:   \(PC)  =  eL°co(r). 

PROOF:   Let  c,  take  on  the  value  c,.   Then  e  °o)(r)  takes  on  the  value  K0(c)(o(r-)  where 

d 

r-=  (/-,,  ....  rd)'  is  the  point  in  //for  which  8-,  =  bi0)  +  £  rJb(J)  is  the  unique  optimal  point 
of  D-.   Since  8-  >  0  one  has  K0(c)a>(r-)  =  v(P?)  [4,  p.  88,  Th.  3].   Thus,  v(Pc)  =  eL°a>(r). 

We  shall  now  obtain  the  moments  of  v(Pc)  when  v<0)  €  span  [s  j)\j  6  <d>). 

THEOREM     5:      Assume     s<0)  €     span     {s(J)\j  €  <d>).      Then,     for    every    v  e  TV, 


F 


W(v(J»))  =  §   ^  {£ (r) }>(/•) dr where,  for  r  €  //, 


P.  ELLNER  AND  R    STARK 


h(r)  A 


eDl\8l(r)u'll\Jry^     ifp>0. 


In   the  above   D  A  £    U/0)  -  £.vA(/)     log  a„    w,  A  £    v^/7 ' -6,(0)   for   /  €  <«>,   and 
vK  A  £  ^/AJ'1  -X.K(0)  for  k  €  </>>  where  (y, v^)'  is  the  unique  element  of  Rd for  which 

5'0)=  £  _y/5(/>. 

7=1 

PROOF:    We  shall  assume  p  >  0.     (The  modification  needed  in  the  proof  for  p  =  0 
should  be  clear.)      By  Assumption  (2)  and  Proposition  1  the  set  [sU)\j  6  <d>)  is  linearly 

independent.    Hence,  by  Proposition  3  there  exists  a  unique  element  (yu  ...  ,  yd)'  of  Rd  for 

d  d 

which   5(0)  =    £  yjS(J).     Also,   by   Proposition  3,   log  K0(c)  =    £  >>,-  log  K^c)  +  D  where 

I/=i  /-I 

d  \ 

b,(0)  -  L  yM'\  logo,-.    By  Proposition  5  v(Pc)  =  K0(c)co(r).   Thus, 

(1)        log  (v(Pc))  =  log  KQ(c)  +  log  (<»(/■)) 

-  £  jey.lpg  «/(<:>.-+  Z>  ;*-  £  XK(0)  log  M')  -  £  J»((0)  log8,(/-). 

/-I  K=l  ,=  l 

Let  c,  take  on  the  value  c,.    Then  /■  takes  on  the  value  r-  =  (r^c),  ...  ,  rd(c))'.    Since 

d 

8-  A  6(0)  +  £  (rji(c))6(/)  is  an  optimal  point  of  D-  and  8-  >  0  (by  Assumption  (4))  one  has 
(by  [4,  p.  88,  Th.  3]) 

log  Kj(c)  =  h,(r-)  =  £  6,(/)  log  8,(r-)  -   £  *#?  log  \K(r-) 
for  every  y  €  <d>.   Thus,  by  (1), 

log  (v(/>?))  =  £  v,  £  fcO>  log  8,-(rr)  -  £  \K(/)  log  \K(rF) 
M      (,=  i  .=  i 

+  D  +  £  \j0)  log  \K(r?)  -  £  6,(0)  log  8,(/-r) 

-  ° +  £  z  m(/)  -  bil0)\  !°g  8<(rP 

-  £  [£MK(/)-^(0,|log\K(rr). 
Hence, 

v(Pc)  =  eD  f[  8,(r)"'n  *»"""  =  w(r). 
It  follows  that  £M(v(/>,))  =  £({v(/>(.)}")  =  Je//  |«(r))'*(r)*  [8,  p.  18,  Th.  1.4.3]. 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  563 

We  next  obtain  the  moments  of  v(Pc)  in  terms  of  the  density  function  of  r  when  s(0)  $ 
span  [s(i)\j  €  <d>). 

THEOREM  6:   Assume  5(0)  tf  span  [s(i)\j  €  <d>)  and  let  v  €  N.   Then, 

(a)  e  °  is  lognormal  and  independent  of  R  A  w(r); 

(b)  £(,,)(v(Pc))  =  ^(e^^CR); 

(c)  EM  (eL°)  =  expl|t  M(VW  "  t  *i*A  +  T"  fr®'  As<0))] 

I  l'-i  '=i  I        2  J 

where  s<0)  =  /37>(0); 

(d)  £(">(i?)  =  J        {a>(r)}^(r)dr. 

PROOF:  (a)  By  Theorem  1  L0  is  normal  and  hence  e  °  is  lognormal. 

Note  by  Proposition  4  w(r)  =  a>0~'(L))  where  I  =  (Ilf  ....  Ld)'.  By  Theorem  1  L0 
and  L  are  independent.   Thus,  e  "and  i?  are  independent  [8,  p.  15,  (III)]. 

(b)  To  show  that  E(R")  exists  let  X  A  /L°  and  K  A  R".  Clearly  (A',  K)  is  a  continuous 
random  vector.  Thus,  ret  wbe  the  joint  density  function  of  (X,Y).  Also,  let  wj  and  w2  denote 
the  marginal  densities  of  X  and  Y  respectively.  Then  w(x,y)  =  wx(x)w2(y)  for  all 
(x,y)  €  /?'  x  R[  since  Zand  Kare  independent  by  (a). 

By    Proposition    5    {v(Pc)}''  =  XY.     Thus,    by    assumption,    E(XY)    exists    and    hence, 
xyw  (x,y) dxdy  is  convergent.    Thus,  by  Fubini's  Theorem  [10,  p.  207,  Th.  2.8.7] 

U.yKfl'x/?1  ^  oo  oo 

xvw(x,v)d!x4y  =  J         H(y)dy    where     //(v)  A  J         xyw(x,y)dx  =  yw2(y)     J 
Uy)ZRl*Rl  °  rco  co 

x»vi(x)rfx  =  yw2(y)E(X).   This  implies  E (XY)  =  E (X)   I      ^(v)^' and  hence   I      yw2(y)dy 
is  convergent,  i.e.,  E(R  )  exists. 

Since  the  expected  values  of  e  °  and  /?"  exist,  the  independence  of  e*  °  and  R"  implies 
E(e"L°)E(Rv)  =  E(e"L*Rv)  [5,  p.  82,  Th.  3.6.2].  Thus,  by  Proposition  5, 
EM(y{Pc))  =  E({eL°RY)  =  E(e"L(i)E{Rv)  =  EM  (/°)£("  >(/?). 

(c)  Recall  by  Theorem  1  E(L0)  =  £  /x,s,(0)  -  £  a,ft/0)  and   K(I0)  =  s(0)As(0)  where 
=  /3'6(0)  and  K(L0)  denotes  the  variance  of  L0-    By  [2]  one  has  ^^'(e^0)  =  exp  [vE(L0)  + 


2 

(d)  Using  the  density  function  </»  of  r  we  obtain  £(|,)(/?)  =  E(R")  =  J  {w(r)}Xr)<fr 
[8,  p.  18,  Th.  1.4.3].  r€W 

Note  that  to  evaluate  Eiv)(v(Pc))  by  Theorem  5  or  6  one  must  integrate  a  specified  func- 
tion over  the  convex  polyhedral  set  H  =  [r  €  /?^|5r  >  -  bm).  Hence,  the  comments  made 
in  Section  4  concerning  the  evaluation  of  f(z)  also  apply  to  the  evaluation  of  E{v)(v(Pc)).    In 


564  P.  ELLNER  AND  R.STARK 

particular,  note  that  for  a  given  precision  the  amount  of  work  required  to  calculate  EM(v(Pc)) 
by  Theorem  5  or  6  should  be  about  the  same  as  the  amount  required  to  calculate  f(z)  by 
Corollary  4.1.  Thus,  in  calculating  £(t,)(v(Pf)),  it  is  advantageous  to  express  E(v)(v\pc))  in 
terms  of  the  density  function  of  r  as  in  Theorems  5  and  6  rather  than  to  express  E(v)(\(Pc))  as 

Jo"  «"/«*• 

6.   EXTENSIONS 

In  this  section  we  shall  indicate  how  the  preceding  results  can  be  used  to  obtain  the  distri- 
bution and/or  moments  of  \{PC)  when  Pc  need  not  satisfy  all  the  assumptions  of  Section  3. 
However,  no  formal  statements  or  proofs  will  be  presented. 


I 
which  are 


n  the  following,  we  shall  refer  to  strengthened  versions  of  Assumptions  (2),  (3),  and  (4) 
are  stated  below  for  a  stochastic  geometric  program  Pc  that  satisfies  Assumption  (1). 

We  say  Pc  satisfies  Assumption  (2')  iff  Pc  satisfies  Assumption  (2)  and  s(0)  G  span 
{slJ)\j  €  <d>\  where  s(J)  A  p'blJ)  for  every  j  6  <d>.  Here  [b(j)\j  €  <d>)  is  any  nullity 
set  for  Pc  and  b{0)  is  any  normality  vector  for  Pc.  Also,  the  matrix  /3  is  defined  as  in  Assump- 
tion (1). 

Pc  is  said  to  satisfy  Assumption  (3')  iff  P-  is  superconsistent  and  soluble  for  every 
c  €  R'±. 

Finally,  Pc  is  said  to  satisfy  Assumption  (4')  iff  D-  has  a  unique  optimal  point  8-  and 
8-  >  0  for  every  c  €  R>. 

Now  consider  a  family  of  random  cost  vectors  { c (e )  |  e  €  (0,°o)},  where  cU)  A 
(c^e),  ...,  c„(e))'fore  €  (0,°°),  that  satisfies  the  following: 

(i)  (log  C\(e),  . . .  ,  log  c„(e))'  is  a  nondegenerate  normal  random  vector; 

(ii)  £(log  c,(€))  =  £(log  Cj)  for  every  /  €  <n>\ 

(iii)  lim  Cov(log  c.(e),  log  c.(e))  =  Cov(log  c.,  log  c.)  for  every  (i,j)  €   <n>  x  <«>, 

€10 

where  Cov  denotes  covariance. 

Such  a  family  of  cost  vectors  can  easily  be  constructed  if  Pc  satisfies  Assumption  (1). 
When  Pc  also  satisfies  Assumptions  (3')  and  (4')  one  can  show  that  Pc{e)  will  satisfy  Assump- 
tions (1),  (2'),  (3'),  and  (4'),  where  Pc^  is  obtained  from  Pc  by  replacing  c  with  the  cost  vec- 
tor c(e).  Thus,  the  results  of  the  preceding  sections  can  be  used  to  calculate  the  moments,  dis- 
tribution function,  and  density  function  of  Pcie)  for  e  €  (0,°o).  Additionally,  one  can  establish 
that  the  moments,  distribution  function,  and  density  function  of  Pc(()  converge  to  the 
corresponding  moments,  distribution  function,  and  density  function  of  Pc  as  €  tends  to  zero. 

Next,  consider  the  family  of  stochastic  geometric  programs  [Pc(y)\y  €  (0,<»)}  where,  for 
y  €  (0,oo),  Pc{y)  denotes  the  following  stochastic  program: 

inf  I  c,  ft  t?  n  ^ 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  563 

We  next  obtain  the  moments  of  v(Pc)  in  terms  of  the  density  function  of  r  when  sm  $ 
span  [s(J)\j  e  <d>). 

THEOREM  6:   Assume  s(0)  tf  span  [s(J)\j  £  <d>)  and  let  v  €  N.   Then, 

(a)  e  °  is  lognormal  and  independent  of  R  A  w(r); 

(b)  £w(v(/'c))  =  ^(/o^o,)  (/?); 

(c)  E^(eL°)  =  expl\±nisi{0)-  £  a,6/d)|  +  ^  ^m' Asm)\ 

I   l=i  '=i  J        2  I 

where  s(0)  =  /37>(0); 

(d)£(^(/?)  =  Xg//   (<o(r)}>(r)rfr. 

PROOF:  (a)  By  Theorem  1  L0  is  normal  and  hence  e  °  is  lognormal. 

Note  by  Proposition  4  a>(r)  =  a(h~l(L))  where  L  =  (Lt,  ...  ,  Ld)'.  By  Theorem  1  L0 
and  L  are  independent.   Thus,  e  °  and  /?  are  independent  [8,  p.  15,  (III)]. 

(b)  To  show  that  £(/?")  exists  let  IA  e"L°  and  Y  A  /r.  Clearly  (Z,K)  is  a  continuous 
random  vector.  Thus,  let  wbe  the  joint  density  function  of  (X,  Y).  Also,  let  wx  and  w2  denote 
the  marginal  densities  of  X  and  Y  respectively.  Then  w(x,y)  =  w\(x)w2(y)  for  all 
(x,y)  €  R[  x  i?1  since  Zand  Fare  independent  by  (a). 

By    Proposition    5    {v(Pc)Y  =  XY.     Thus,    by   assumption,    E(XY)    exists    and    hence, 
xy»v(x,v)d!x^v  is  convergent.    Thus,  by  Fubini's  Theorem   [10,  p.  207,  Th.  2.8.7] 

UV)€J(IXXI  ^  CO  oo 

J  J  xywOc,y)<£«/v  =  JQ       H(y)dy    where     //(v)  A  Jq       xvw(x,v)d!x  =  vh>2(v)     JQ 

(x.y)€RlxRl  roo  oo 

xw!(x)rfx  =  yw2(y)E(X).   This  implies  £(AT)  =  E(X)  J      vw2(v)d[y and  hence  J      ytv2(y)</v 
is  convergent,  i.e.,  E(R")  exists. 

Since  the  expected  values  of  e  °  and  R"  exist,  the  independence  of  e  °  and  /?"  implies 
E(evL°)  E(RV)  =  E{eL»R")  [5,  p.  82,  Th.  3.6.2].  Thus,  by  Proposition  5, 
E(v)(y{Pc))  =  E({eL°R}")  =  E(evL°)E(R")  =  £(v)  (/°)£MU). 

(c)  Recall  by  Theorem  1  E(L0)  =  £  /a,S/(0)  "'E  a,V0)  and  K(I0)  -  5(0)As(0>  where 
5(0)  =  ^^(0)  and  K(Lo)  denotes  the  variance  of  L0.  By  [2]  one  has  £M(/°)  =  exp  [i/£(L0)  + 
—  v2V(L0)]  since  L0  is  normal. 

(d)  Using  the  density  function  <//  of  r  we  obtain  EM{R)  =  £(i?")  =  J  {a>(r)}>(r)</r 
[8,  p.  18,  Th.  1.4.3].  '*" 

Note  that  to  evaluate  E{v)(\(Pc))  by  Theorem  5  or  6  one  must  integrate  a  specified  func- 
tion over  the  convex  polyhedral  set  //={/•€  Rd\Br  >  -  b(0)).  Hence,  the  comments  made 
in  Section  4  concerning  the  evaluation  of  /(z)  also  apply  to  the  evaluation  of  Elv)(y(Pc)).    In 


564  P.  ELLNER  AND  R.STARK 

particular,  note  that  for  a  given  precision  the  amount  of  work  required  to  calculate  EM(v(Pc)) 
by  Theorem  5  or  6  should  be  about  the  same  as  the  amount  required  to  calculate  f(z)  by 
Corollary  4.1.  Thus,  in  calculating  ^"'(vC/^)),  it  is  advantageous  to  express  EM(v(Pc))  in 
terms  of  the  density  function  of  r  as  in  Theorems  5  and  6  rather  than  to  express  EM{y(Pc))  as 
/."  **/(*)<&. 

6.   EXTENSIONS 

In  this  section  we  shall  indicate  how  the  preceding  results  can  be  used  to  obtain  the  distri- 
bution and/or  moments  of  \{PC)  when  Pc  need  not  satisfy  all  the  assumptions  of  Section  3. 
However,  no  formal  statements  or  proofs  will  be  presented. 

In  the  following,  we  shall  refer  to  strengthened  versions  of  Assumptions  (2),  (3),  and  (4) 
which  are  stated  below  for  a  stochastic  geometric  program  Pc  that  satisfies  Assumption  (1). 

We  say  Pc  satisfies  Assumption  (2')  iff  Pc  satisfies  Assumption  (2)  and  s(0)  &  span 
{sin\j  €  <d>\  where  s</>A/3'6</,  for  every  j  €  <d>.  Here  [b(n\j  6  <d>)  is  any  nullity 
set  for  Pc  and  b(Q)  is  any  normality  vector  for  Pc.  Also,  the  matrix  /3  is  defined  as  in  Assump- 
tion (1). 

Pc  is  said  to  satisfy  Assumption  (3')  iff  P-  is  superconsistent  and  soluble  for  every 
c  €  R'i. 

Finally,  Pc  is  said  to  satisfy  Assumption  (4')  iff  D-  has  a  unique  optimal  point  8F  and 
8-  >  0  for  every  c  €  R">. 

Now  consider  a  family  of  random  cost  vectors  (c(€)|e  €  (0,°°)},  where  c(t)  A 
(c\(e) c„(€))'fore  €  (0,°°),  that  satisfies  the  following: 

(i)  (log  C](e) log  c„(e))'  is  a  nondegenerate  normal  random  vector; 

(ii)  £(log  c,(e))  =  £"(log  c,)  for  every  i  €  <n>, 

(iii)  lim  CovOog  c,(e),  log  c,(e))  =  Cov(log  c.,  log  c.)  for  every  (ij)  €   <n>  x  <«>, 

610 

where  Cov  denotes  covariance. 

Such  a  family  of  cost  vectors  can  easily  be  constructed  if  Pc  satisfies  Assumption  (1). 
When  Pc  also  satisfies  Assumptions  (30  and  (4')  one  can  show  that  Pc{e)  will  satisfy  Assump- 
tions (1),  (2'),  (30,  and  (40,  where  Pcif)  is  obtained  from  Pc  by  replacing  c  with  the  cost  vec- 
tor c(e).  Thus,  the  results  of  the  preceding  sections  can  be  used  to  calculate  the  moments,  dis- 
tribution function,  and  density  function  of  Pc(t)  for  €'€  (0,°o).  Additionally,  one  can  establish 
that  the  moments,  distribution  function,  and  density  function  of  Pc(()  converge  to  the 
corresponding  moments,  distribution  function,  and  density  function  of  Pc  as  €  tends  to  zero. 

Next,  consider  the  family  of  stochastic  geometric  programs  {/V?)ly  €  (0,°°)}  where,  for 
y  €  (0,°o),  P}y)  denotes  the  following  stochastic  program: 

y  I  c,  ft  t?  n  *r 

'•~     i€7„         /=1  k=1 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  565 

subject     to      £  c,  Yl     tj1'  +  zk  ^  1     for    every     k  €  <p>,     t  >  0,     and     z  >  0     where 

f  =  (flf  . . .  ,  tm)'  and  z  =  (z),  ....  z^,)'.  One  can  show  Pc{y)  satisfies  each  assumption  that  Pc 
satisfies.  In  addition,  if  Pc  satisfies  (3')  then  Pc{y)  satisfies  (3')  and  (4')  (even  when  Pc  does  not 
satisfy  (4)).  Thus,  one  can  apply  the  results  of  the  preceding  sections  to  calculate  the  density 
function,  distribution  function,  and  moments  of  Pc(y)  for  y  €  (0,°°)  when  Pc  satisfies  (1),  (2'), 
and  (3').  Furthermore,  one  can  establish  that  the  moments,  distribution  function,  and  density 
function  of  Pc{y)  converge  to  the  corresponding  moments,  distribution  function,  and  density 
function  of  Pc  as  y  approaches  zero. 

Finally,  for  y  €  (0,°°)  and  e  (0, °°),  let  P}?f\  denote  the  stochastic  program  obtained 
from  Pc{y)  by  replacing  cost  vector  c  by  c(e)  in  Pciy).  The  family  of  cost  vectors 
{c(c)|  e  €  (0,  °o)}  is  assumed  to  satisfy  the  properties  (i),  (ii),  and  (iii)  previously  listed.  One 
can  show,  for  (y,e)  €  (0,°o)  x  (0,«>),  program  Pc(e)iy)  satisfies  (1),  (2'),  (30,  and  (4')  if  Pc 
satisfies  (1)  and  (3').  Thus,  in  this  instance,  one  can  apply  the  results  of  the  preceding  sections 
to  Pc%\.  This  suggests  that  the  family  of  programs  [Pc{$  \y  €  (0,°o)  and  e  €  (0,°°)}  may  be 
useful  in  obtaining  the  moments,  distribution  function,  and  density  function  of  Pc  when  Pc 
need  only  satisfy  Assumptions  (1)  and  (3'). 

APPENDIX  A. 

Theorem  1  in  Section  3  is  an  immediate  consequence  of  the  following  lemma. 

LEMMA    A:     Define    L(z,s)  A  £    5,  log  z,    for   every    positive-valued    random    vector 

z  A  (z\,  ...  ,  zu)'  and  5  €  R".  Also,  define  the  inner  product  <-,->A  on  R"  by 
<x,v>A  A  x'Ay  for  (x,y)  €  R"  x  R".  (Note  <•,  >A  is  an  inner  product  since  A  is  a  disper- 
sion matrix  of  a  nondegenerate  normal  random  vector  and  hence  is  positive  definite.)   Then 

(i)  (L(e,stu),  ...  ,  L(e,s(d)))'  is  a  normal  random  vector  with  independent  components 
where  e  A  (eu  ....  eu)\  s(1)  =  s(1),  and 

,(/)  =  JO)  _    £   «5(/),5(/)>A)-1  «S(i),SU)>x)s{l) 

for  1  <  j  <  d, 

(ii)    (L  (e,si0)),L(e,siu),  . . .  ,L(e,s{d)))'   is  a   normal   random   vector  with   independent 

d 
components     if     si0)  g     span     [siJ)\j  €  <d>)     where     s(0)  =  s(0)  -     £     (<s(/),s(/)>Ar' 

(<s(0\s(/)>A)s(/)  when  s(0)  <?  span  {sin\j  €  <</>}. 

(iii)  For  every  j  €  <d>,  s0)  =  p'bU)  and  f\  c-'     =    ]]«,''''     {exp  (L(e,sU)))}  where 

lw  ) 

bw  =  bw  and,  for  1<  j  <  rf,  Z>0)  -*</>_£  «fi'b{l),fi'bU)>  x)~{  «p'bU),p'b(l)> A)6(/). 

Also   6(0)  =  6(0)   if  5<0)€    span   {S0)L/  €  <rf>}   and   6(0)  =  £(0)  -    £    «j8'6(,)^'6(/)>A)-1 

(<P'b{0),p'bU)>A)bU)  if  5(0)  <?  span  {s0)|y  €  <d>).  Furthermore,  [bU)\j  €  <d>)  is  a  nul- 
lity set  and  b(0)  is  a  normality  vector  of  Pc\ 


566  P.  ELLNER  AND  R   STARK 

(iv)  For  j  e  <d>,  the  density  function  <£,  of  L(e?,s0))  is  given  by  <£,(/)  A  j=  exp 

-  ojyv27r 

r —    for  every  /  €  Rx  where  p,  is  the  expected  value  of  L(e,s{J))  and  cof  is  the  vari- 

l  2<u/     j 

ance  of  L(e,s0)).   Furthermore,  pj ■.  =  *£  nisi(J)  and  ajj  =  <s0),s0)>A. 
/=i 

PROOF:  Since  A  is  real  symmetric,  A  has  an  orthonormal  set  of  u  eigenvectors 
[p\>  ■  •  •  <Pu\-  Let  P  be  the  u  x  u  matrix  whose  yth  column  is  p}.  Then  P  is  orthogonal  (i.e., 
P~x  =  P')  and  A  A  P-1  A/Ms  diagonal. 

For  every  i  e  <u>  let  y,  A  log  e,  and  y  A  (yi,  ...  ,  -yu)'.  Let  -y  A  P'y.  Then  y  is  a  m- 
variate  normal  vector  with  dispersion  matrix  P'AP  =  A  ([8,  Th.  2.1.1]).  Since  A  is  diagonal, 
the  components  of  y  are  independent. 

Let  (s,w)  (E  Ru  x  R"  such  that  w  =  P's.  We  shall  show  L(e,s)  =  L(e,w)  where  e  A 
(?!,  ....  eu)'  and  e,  A  ey'  for  /  €  <«>.  Note  y  =  />y.  Thus,  for  /  €  <w>,  y,  =  ]£  AkTV 
Hence, 

(1)  L(e,s)=  £s, \£PiKyK   -  t    2>k4?« 

=  H  wKyK  =  £(e,w). 

For  every  y  €  <d>  define  w^' A  P's(J).  By  assumption  {s"  L/  €  <</>}  is  linearly 
independent.  Thus  {w^}\j  €  <d>]  is  linearly  independent  since  P'  is  nonsingular.  Thus,  one 
can  apply  the  Gram-Schmidt  orthogonalization  process  to  {w^ly  6  <</>}  to  obtain  the 
orthogonal  set  [w^^lj  €  <d>}  with  respect  to  <v>a  where  w(1)  A  h>(1)  and,  for  1  <  j  <  4 

(2)  w°')  A  w0)-  £  (<w(/),w(/)>A)-1(<w'0')(iv(/)>A)w(/). 

(Note  <•,  •>£  is  the  inner  product  on  Ru  defined  by  <x,y>A  A  x'Ay  for  (x,y)  €  R"  x  Ru. 
<•, ->jj  is  an  inner  product  on  /?"  since  A  is  the  dispersion  matrix  of  the  nondegenerate  nor- 
mal random  vector  y  and  hence  is  positive  definite.)     Also  define 

(3)  w(0)  A  wi0)-  j^«WU),wU)>^«w(0\WU)>^win 

/=i 

if  si0)  g  span  (S°H/  €  <</>};  otherwise  define  w<0)  A  u>(0).  Observe  if  .?(0)  <?  span 
{sQ)\j  6  <</>}  then  w(0)  £  span  {w^l/  €  <d>).  Thus,lw0)U'  6  <</>}  is  an  orthogonal 
set  in  i?"  with  respect  to  <•,  >K  when  s(0)  <?  span  {s{n\j  €  <d>}. 

Define  s0)  A  Pw0)  for  every  j  €  <d>.  Then,  for  j  €  <^>,  w(J)  =  P's0).  Thus,  by 
(1), 

(4)  Lie.s^^  =  L(e,w0))    for  every  J  e  <d>. 

We  shall  next  show  (L(e,w(x)),  ...  ,  L(e,w(d)))'  is  a  normal  random  vector  with  independent 
components.  Also,  whenever  s(0)  $  span  [S^J)\j  €  <d>),  we  shall  show  that 
{L(e,wm),L{e,w(x)),  ...,  L(e,wid)))'  is  a  normal  random  vector  with  independent  com- 
ponents. 


OPTIMAL  VALUE  OF  STOCHASTIC  GFOMFTRIC  PROGRAMS  567 

For  every  /  €  <u>  let  9}  and  t,  be  the  variance  and  mean,  respectively,  of  log  eh  For 
i  €  <u>  define  i//,  A  0,-1  (log  q  -  t;).    Note,  for  every  7  €  <^>,  L^w0*)  =  £  w,0)  log 

*/ -  E    0,w,O)    <//,+  £    tiwi<j)-      Let    r,t  £  <d>     such    that    r  ^  r.      Recall    -y  A     (log 

fi,  ...  ,  log  eu)'  is  a  normal  vector  with  independent  components.  Thus,  (<//,|/  €  <w>}  is  a 
set  of  independent  unit  normal  random  variables.    Thus,  L{e,w(r))  and  L{e,w{,))  are  normal 

random  variables.  Moreover,  L(e,w(r))  and  L(e,wU))  are  independent  provided  £ 
^VV"  =  0  [8,  Th.  4.1.1,  p.  70]. 

Since,  for  every  i  €  <«>,  9}  is  the  variance  of  y,  and  A  is  the  dispersion  matrix  of  y 
one    has    JT    Ofw^w^  =  <w(r),w(f)>A.     By    construction    of    {h>0)|./'€  <J>)    one    has 

<M,('-)(M;(')>A  =  0  for  r,r  6  <d>  with  r  ^  /.  Also,  <w{r) ,wU)> A  =  0  for  r,r  €  <</>  with 
r  *  t  provided  s(0)  <2  span  {s{,)\j  6  <</>}.  Hence,  by  (4),  (L(e,s(1)),  ...  ,  L{e,s(d)))'  is  a 
normal  random  vector  with  independent  components.  Also  by  (4),  (L(e,s{0)), 
L(e,sw),  ...  ,  L(e,s(d)))'  is  a  normal  random  vector  with  independent  components  if  s(0)  £ 
span  [s{n\j  €  <d>). 

Next,  let  0t(' V)  €  Ru  x  /?"  for  /  €  {1,2}  such  that  y(,)  =  P'xU).   Then, 

(5)  <  v(l),y(2)>A  =  <P'xa>,P'xi2)>i  =  0c(1))7>A/>';c(2)  =  <x(1),x(2)>  A. 

Observe  s(1)  =  Pwn)  =  Ph>(1)  =  s(1).   Also,  by  (2)  and  (5),  for  1  <  j  <  done  has 

(6)  s0)  =  /V'>  =  Pw(J)  -  £  (<*v(/\h>,,)>a)-1(<>v0),vv(/)>a)/V/) 

=  S(J)  -    £   «S(/),5(/)>A)-1«50,,5(/)>A)S(/). 

Moreover,  by  (3)  and  (5)  one  has 

(7)  5(0)  =  PW(0)  =  5(0)  -    £   «5('),S(/)>A)-1  «5(0,,5(/)>A)S(/> 

/=1 

if  5(0)  £  span  [s{l)\j  €  <rf>}.   This  completes  the  demonstration  of  (i)  and  (ii). 

Next,  we  shall  obtain  a  nullity  set  {b{,)\j  6  <d>)  and  normality  vector  b{0)  for  Pc  such 

that  s(J)  =  fi'b(J)  and  f[  c*  '   =     j]o*'       {exp(L  (e,s{J )»}  for  every  J  €  <</>.    Let  5A 

span  {S0)|j  €  <rf>)  and  B  A  span  {£0'[/  €  <rf>}.  By  assumption  {s0)L/  €  <d>)  is 
linearly  independent  and  hence  a  basis  for  S.  Thus,  there  exists  a  unique  linear  transformation 
T.  S  —  #  such  that  r(5C/))  =  £0)  for  every  y  €  <rf>.  Since  {£(/)|./  €  <</>}  is  a  basis  for 
5,  r  is  an  isomorphism  from  5  onto  B.  Also,  since  T~l(b^  )  =  s^  =  /3'biJ)  for  every 
./€<</>  one  has  7^(6)  =  ]3'6  for  every  6  6  5. 

Recall  s(1)  =  s(,)  €  5.  Let  1  <  j  ^  rf  and  assume  sin  6  S  for  1  <  /  <  j.  Then  by  (6) 
one  has  s(J)  €  S.  Hence,  [s(J)\j  €<d>)  C  S.  Also  {s(/ } |y  €  <</>}  is  linearly  independent 
since  {w(/)|y  6  <d>)  is  orthogonal  with  respect  to  <  ,>A  and  P  is  a  nonsingular  matrix  for 


568  P.  ELLNER  AND  R.STARK 

which  5(/)  =  Pw{j)  for  every  j  6  <d>.  Thus,  [s^^lj  €  <d>}  is  a  basis  for  5.  Since  Tis  an 
isomorphism  from  S  onto  B,  [T(s(,))\j  €  <d>)  is  a  basis  for  2?.  Thus,  [b(i)\j  €  <</>}  is  a 
nullity  set  for  P,  where  bU)  A  Hs0')  for  y  €  <rf>. 

Let  5  A  span  [sU)\j  f  <rf>)andSA  s_panj6(/,Jy  €  <d>).  Suppose  sj0)  ?  5.  Then 
there  exists  a  unique  linear  transformation  t:  S  — ■  B  such  that  f(s(/))  =  b{,)  for  every 
y  €_<rf>.  Also,  by  (7),  s(0)  €  S.  Thus,  we  can  define  b{0)  A  f  (s(0)).  Note  by  the  definition 
of  Tone  has  T(s)  =  T(s)  for  every  s  €  5.   Thus,  by  (7), 

(8)  6(0)  =    f(5(0))  -    £   «5(/))5(/)>A)-1«5(0),S(/)>A)f(5(/)) 

=   6<0)  -    £   «S(/\5(/)>/V)-1«5(0)(5(/)>,)6</) 

since  f(5(/))  =  T(sV))  =  6(/)  for  /  €  <rf>.  For  y  €  <m>  let  ^  denote  column  j  of 
exponent  matrix  A.  Recall  q  €  R"  such  that  <7,  =  1  if  /  €  <no>  and  9,  =  0  if  /  >  aj0  where  w0 
is  the  number  of  elements  in  J0.  Then  by  (8),  for  every  j  6  <m>,  one  has  <b(0\  Aj>  =  0 
since  b{0)  is  a  normality  vector  and  {b(l)\l  €  <d>)  is  a  nullity  set  of  Pc,  where  <•,•>  denotes 
the  usual  inner  product  on  R".  Also,  <bi0),q>  =  <£(0\<7>  =  1  since  <bU),q>  =  0  for 
every  /  €  <d>.  Thus,  bi0)  is  a  normality  vector  for  Pe.  If  v(0)  €  S  we  define  bm  A  £(0). 
Thus,  whether  s<0)  £  Sor  5(0)  €  Sone  has  that  bm  is  a  normality  vector  for  Pc. 

To  show  (3'bV'=  5(/>  for  every  j  €  <</>  first  observe  for  j  €  <d>  one  has 
p'b{l)  =  T-Hb^)  =  T[(T(s{l))  =_s0).  Next  suppose  s(0)  (?  5.  Then  f  is  an  isomorphism 
from  S  onto  5  since  [b^  \j  €  <d>}  is  linearly  independent  and  f(5(/>)  =  b(J)  for  every 
y  €  <d>.  Also,  f~1(£0'))  =  s0)  =  p'8U)  for  every  y  €  <</>.  Hence,  f"'(6)  =  p'b  for  all 
6  6  B.  Note  6(0)  A  f(5(0))  €  5.  Thus,  /37>(0)  =  f~Hb{0))  =  f^f  (s(0)))  =  s(0).  Finally,  sup- 
pose s(0)  6  S.  Then  5'M0)_=  /3'£(0)  =  v(0)  =  Pwi0)  =  Pw(0)  =  s(0).  Thus,  from  the  above, 
pb(j )  =  s(j )  for  every  j  €  <d>. 

Next     let     j  €  <d>      and     observe      J  J     c, '     =  Y[     \ai  Yl  e"    \        =       IT01' 


0/„*i 


5>„A'' 


IT    n<V  [=    11"/'  11*7'  •    Thus'  since  fi'bU)=s('\  one  has  fjc,-'     = 

n«*'    n^,!  =  na'6'     exp  £  ^/,io&^ [}-  na'*    iexpa(e,s(y)))}. 

Note      6(,)  =  r(5(,))  =  T(Pw(l))  =  r(/>H>(1>)  =  r(s(1))  =  £(1).       Also,      by      (6),     for 
KJ<d     one     has      b(n  =  T(s(l))  =  T(s{n)  -      £      «5(/),5(/)>/V)-1      «.v0),s(/)>A) 

F(s(/))  -  blJ)  -     X    (<)8'*(/^/3'6<,)>A)-,     (<p'b(,\p'bU)>x)bU).     Recall    if    5(0)  €     span 
{s0>|/  €  <rf>)  then  6(0)  A  b(0).   If  5(0)  ?  span  {.v(/)|y  €  <d>)  then  by  (8)  one  has 

&(0)  =  Sm  -  £  «p'b(l),l3'bU)>^-H<{i'bm,(3'bu)>/i)b{l). 
i=\ 

This  completes  the  demonstration  of  (iii). 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  569 

Finally,  let  j  €  <d>  and  recall  L(e,s{j))  =  £  s,(/)  log  e,-.  Thus,  pj  —  £  S/^V/  ar>d 
<s(/),  5(/)>A  [8,  Th.  2.1.1,  p.  29]  since  (ji\,  ...  ,  ixu)'  is  the  mean  vector  and  A  is  the 
•sion   matrix   of   e.     Also,    since   L(e,sij))    is   normal,    one   has   </>,(/)  =    j=   exp 

OJ  ,  V  27T 

for  every  /€/?'. 

APPENDIX  B. 


2w 


PROPOSITION  B:  (a)  If  //is  bounded  then  EM(v(Pc))  exists  for  every  v  €  TV. 

(b)  //is  bounded  iff  the  set  F  A  (8  €  /?"|8,  ^  Ov  /  €  <  «  >,  ^'8  =  0,  and  ^'8  =  1}  is 
bounded. 

PROOF:  (a)  Assume  //is  bounded.  Then  for  every  j  €  <d>  there  exist  real  numbers  /, 
and  Uj  such  that  /,  ^  r  <  «,  for  every  r  €  //.  Let  v  £  N  and  define  zc  A  v(Pc).  Finally, 
assume  q  takes  on  the  value  c7  and  recall  r-  =  (/-,(c),  ...  ,rd(c))'  is  the  element  of  //for 

which  8  A  bi0)  +  £  (rj(c))bin  is  the  unique  optimal  point  of  Dc. 

j=  i 

Since  8  is  the  optimal  point  of  D-,  by  [4,  Ch.  3,  Sec.  3]  and  Assumption  3  one  has 
zjf  =.Ko(c)"ll  Kj(£)"jm  J!  8,{r,)~"h'{r;)  f\  M'/**^ 
where    TJ  \K(r)k"lr)  A  1    for    r  €  /Y,    the   closure   of   //,    if  p  =  0.     Define   t:    /7  —  Rx    by 

r(r)  A  TJ    8 ,•(/■)     '  '      Q    \K(r)  K  '     for    /■  €  //.     In    evaluating    t(/0    use    the    convention 

xx  =  x~x  =1  for  x  =  0.    Then  t  is  continuous  on  H.    Thus,  since  //  is  compact,  there  exists 
U  €  (0f«>)  such  that  0  <  t(/-)  <   [/for  every  r  6  //.    Hence, 

(1)  0  <  z/  <   ITKvCcY  n  ATyCc)"'^. 

I.  Assume  s(0)  6  span  {s(/,|y  €  <</>}.  Then  by  Proposition  3  there  exists  y, ;  €  Rx  for 
J  €  <</>  and  W  €  (0,°°)  such  that  A"0(c)  =  ^fl  ^(c)'7.   Thus,  by  (1), 

(2)  0  <  z/  <  (£W  rj  Kj{-cY(yi+r'(c)) . 


Let  j  e  <d>.  If  0  <  tf/c)  <  1  then  Kj{c)"yjKj{c)vrj  <  KjW'KjCc)  J.  Also,  if 
1  <  A-7(c)  then  Kj(c)vyjKj(cYr]{c)  <  Kj<£)vyj Kjffi"' .  Thus,  0  <  ^•(c)I,(v^'(c)) 
^  Z/c)  A  max  (Kj(c)v(yj+lj) ,  Kj(c)v{yj+Uj)) .   Hence,  by  (2), 

(3)  0  <  z/  <  (CW  TJ  Z,(c). 

7=1 

Moreover,  by  the  choice  of  {6(/)|./  €  <d>),  the  variates  AT,(c)  for  j  €  <d>  are  independent 
and  hence  {Zj(c)\j  €  <d>}  is  a  set  of  independent  variates. 


570  P.  ELLNER  AND  R.STARK 

By  definition  ox  Z.y(c)  one  has 

(4)  0  <  Zj(c)  <  Kj(c)v(yJ+lj)  +  Kj(c)v(yj+Uj\ 

Since  Kj(c)  is  lognormal  so  are  KjicY  J+J  and  Kj(cY  J+"J .  Thus,  the  expected  values  of 
these  two  variates  exist  and  hence  so  does  E(Kj(cY  yj+J  +  Kj{cY  yj+Uj ).  Thus,  by  (4),  the 
expected  value  of  Zj(c)  exists.  Since  the  variates  Zjic)  for  j  €  <d>  are  independent  the 
expected  value  of  {UWY  f\  Zj(c)  must  also  exist  [5,  p.  82,  Th.  3.6.2].  Hence,  by  (3),  the 
expected  value  of  zc"  exists. 

II.  Assume  s(0)  tf  span  {s(,)\j  e  <d>).  Then  by  the  choice  of{b{j)\j  6  <d>)  the  vari- 
ates Kj(c)  for  y€  <d>  are  independent.  For  j  €  <J>  define  Z/c)  A  max 
(Kj(cY\Kj(xYUi)  and  let  Z0(c)  A  LFK^Y.  Then  the  variates  Z,{c)  for  y  €  <^>  must  be 
independent.    Furthermore,  0  <  KjicY"'  c    <  Z;(c)  for  j  €  <c/>.    Hence,  by  (1), 

(5)  0  <  zF"  <  n  z/^)- 


Note  that  E(Z0(c))  exists  since  Z0(c)  is  lognormal.  Also,  for  j  €  <d>,  EiZ^c)) 
exists  since  A^7(c)  is  lognormal  and  0  <  Z^c)  <  KjicY  '  +  Kj(cY"J.  Since  the  variates  Zf(c) 
for  y  e  <d>  are  independent  it  follows  from  (5)  that  £"(z/)  exists. 

By  I  and  II  we  conclude  £(")(v(/>c))  exists  for  all  v  €  N. 

_  <y 

(b)  Observe  8  €  F  iff  there  exists  r  =  (r, rd)'  £  H  such  that  8  =  6<0)  +  £  r,60). 

_  /-i 

Also,  //is  bounded  iff  //is  bounded.   Thus,  /"is  bounded  iff  //is  bounded. 

REFERENCES 

[1]    Abrams,    R.A.,   "Consistency,    Superconsistency,   and    Dual    Degeneracy   in   Posynomial 

Geometric  Programming,"  Operations  Research,  24,  325-335  (1976). 
[2]  Aitchison,  J.  and  J.A.C.  Brown,  The  Lognormal  Distribution,  (Cambridge  University  Press, 

London,  England,  1957). 
[3]  Avriel,  M.  and  D.J.  Wilde,  "Stochastic  Geometric  Programming,"  Proceedings  Princeton 

Symposium  on  Mathematical  Programming,  Princeton,  New  Jersey  (1970). 
[4]  Duffin,  R.J.,  E.L.  Peterson  and  C.  Zener,  Geometric  Programming,  (John  Wiley  &  Sons, 

New  York,  New  York,  1967). 
[5]  Fisz,  M.,  Probability  Theory  and  Mathematical  Statistics,  (John  Wiley  &  Sons,  New  York, 

New  York,  1963). 
[6]  Hammersley,  J.M.  and  D.C.  Handscomb,  Monte  Carlo  Methods,  (John  Wiley  &  Sons,  New 

York,  New  York,  1964). 
[7]  Isaacson,  E.  and  H.B.  Keller,  Analysis  of  Numerical  Methods,  (John  Wiley  &  Sons,  New 

York,  New  York,  1966). 
[8]   Lukacs,   E.  and  R.G.  Laha,  "Applications  of  Characteristic  Functions,   (Charles  Griffin  & 

Company,  Ltd.,  London,  England,  1964). 
[9]  McNichols,  G.R.,  "On  the  Treatment  of  Uncertainty  in  Parametric  Costing,"  Ph.D.  Thesis, 

The  George  Washington  University,  Washington,  D.C.  (1976). 
[10]  Munroe,  M.E.,  Introduction  to  Measure  and  Integration,  (Addison-Wesley  Publishing  Com- 
pany, Reading,  Massachusetts,  1953). 


OPTIMAL  VALUE  OF  STOCHASTIC  GEOMETRIC  PROGRAMS  571 

[11]    Rotar,   V.I.,   "On   the   Speed   of  Convergence   in   the   Multidimensional   Central   Limit 

Theorem,"  Theory  of  Probability  and  Its  Applications,  15,  354-356  (1970). 
[12]  Spivak,  M.,  Calculus  on  Manifolds,  (W.A.  Benjamin,  New  York,  New  York,  1965). 
[13]  Stark,  R.M.,  "On  Zero-Degree  Stochastic  Geometric  Programs,"  Journal  of  Optimization 

Theory  and  Applications,  23,  167-187  (1977). 
[14]  Tsuda,  T.,  "Numerical  Integration  of  Functions  of  Very  Many  Variables,"  Numerische 

Mathematik,  20,  377-391  (1973). 
[15]  Turchin,  V.F.,  "On  the  Computation  of  Multidimensional  Integrals  by  the  Monte-Carlo 

Method,"  Theory  of  Probability  and  Its  Applications,  16,  720-724  (1971). 


A  CLASS  OF  CONTINUOUS  NONLINEAR  PROGRAMMING  PROBLEMS 
WITH  TIME-DELAYED  CONSTRAINTS 

Thomas  W.  Reiland 

North  Carolina  State  University 
Raleigh,  North  Carolina 

Morgan  A.  Hanson 

Florida  State  University 
Tallahassee,  Florida 

ABSTRACT 

A  general  class  of  continuous  time  nonlinear  problems  is  considered. 
Necessary  and  sufficient  conditions  for  the  existence  of  solutions  are  esta- 
blished and  optimal  solutions  are  characterized  in  terms  of  a  duality  theorem. 
The  theory  is  illustrated  by  means  of  an  example. 


1.  INTRODUCTION 

Recently  Fair  and  Hanson  [1]  proved  existence  theorems,  duality  theorems,  and  continu- 
ous time  analogues  of  the  Kuhn-Tucker  Theorem  for  a  class  of  continuous  time  programming 
problems  in  which  nonlinearity  appears  both  in  the  objective  function  and  in  the  constraints. 
More  recently  this  class  was  extended  in  Farr  and  Hanson  [2]  to  include  problems  with 
prescribed  time  lags  in  the  constraints.  In  this  paper  we  generalize  these  results  by  considering 
a  more  general  form  of  the  constraints  and  by  assuming  a  less  stringent  constraint  qualification. 
This  constraint  qualification  is  analogous  to  that  of  Kuhn  and  Tucker  [5]  and  provides  further 
unification  between  the  areas  of  finite-dimensional  and  continuous  time  programming.  An 
example  is  presented  wherein  these  results  are  applied  to  a  version  of  Koopmans1  [4]  water 
storage  problem  which  has  been  modified  to  address  the  economic  ramifications  of  an  energy 
crisis. 

2.  THE  PRIMAL  PROBLEM 

The  problem  under  consideration  (Primal  Problem  A)  is: 
Maximize 


V(z)  =  foTcf>(z(t),t)dt 


subject  to  the  constraints 

(1)  z(t)  >  0,    0  <  t  <  T, 

(2)  f(z(t),t)  <  h(y(z,t),t),    0<  /  <  T, 


574  T.W.  REILAND  AND  M.A.  HANSON 

and 

(3)  z(t)  =  0,    t  <  0, 

where  z  6L,^  [0,  T],  i.e.,  z  is  a  bounded  and  measurable  w-dimensional  function;  y  is  a  map- 
ping from  L~  [0,  71  x  [0,  71  into  Ep  defined  by 


(4)  v(z,f)  =  X  J0  £,-(z(s  -  a7),  5  -  ctj)ds\ 


f(z(t),t),  h{y(z,t),t)  €  £m;  $(z(s  -  a,-),  s  -  ay)  6  #\  j  -  0,  . . .  ,  r.  The  set  0  =  a0  < 
a]  <  . . .  <  ar  is  a  finite  collection  of  nonnegative  numbers;  and  </>  is  a  scalar  function,  concave 
and  continuously  differentiate  in  its  first  argument  throughout  [0,  T]. 

It  is  further  assumed  that  each  component  of  —  /,  g,,  and  h  is  a  scalar  function,  concave 
and  differentiable  in  its  first  argument  throughout  [0,71,  that  each  component  of  the  composite 
function  h(y(-,t),t):L™[0,T]  — ►  Em  is  concave  in  z,  there  exists 8  >  0  such  that 

(5)  either  V*//^,/)  -  0  or  V*/ X-q.t)  >h, 

and  for  each  t  and  k  there  exists  ik  =  ik(t)  such  that 

(6)  Vkfik  (.7i,t)  >  8, 
where 

v*/,(i»,/)  -  e/ydj./va^*, 

/  =  1 ,  . . .  ,  m,    k  —  1,  . . .  ,  n, 
forrj  e  E\  f)  >  0,  and  t  €  [0,7]; 

(7)  «/Cz(r),r)-  0,  t  <  0, 

7  =  0,  ...,  r; 

(8)  VMv.O^dhib.tydvt  >  0, 
for  v  £  Ep  and  f  €  [0, 71 ;  and 
(9a) 


(9b)  sup   g.AOj)  <  oo,     sup     Vkgja(0,t)  <  <*>,  j  =  0 


(9c) 
(9d) 


sup  /;,(0,n  <  oo, 

sup    VJ/, 

(0,t) 

< 

oo: 

Ik-  1,  ...  ,  «, 

SUp     £;,((),/)   <   oo 

sup 

v, 

:ft(0, 

7) 

< 

<?  =  1,  ...  ,  p,  k  = 

1,  ...  , 

n, 

inf    f,(0,t)  >  -oo 

,  /=  1, 

,  m, 

sup   V^<£  (rj.r)  < 

oo,   7,   € 

E\ 

TJ    > 

0, 

k 

A  function  z  6  L~  [0, 71  is  termed  feasible  for  Primal  Problem  A  if  it  satisfies  the  con- 
straints (1),  (2),  and  (3).   The  primal  problem  is  itself  said  to  be  feasible  if  a  feasible  z  exists. 

It  should  be  noted  that  Primal  Problem  A  is  identical  to  that  considered  in    [2]  if  p  =  m 
and 

h{yiz,t),t)  =  Imyiz,t) 


NONLINEAR  PROGRAMMING  WITH  TIME-DELAYED  CONSTRAINTS  575 

where  Im  is  an  m-dimensional  identity  matrix. 

3.   EXISTENCE  THEOREM 

THEOREM  1:    If  Primal  Problem  A  is  feasible,  then  it  has  an  optimal  solution,  that  is, 
there  exists  a  feasible  z  for  which 

Viz)  =  sup  Viz), 

where  the  supremum  is  taken  over  all  feasible  z. 

We  preface  the  proof  of  this  theorem  with  a  brief  discussion  of  weak  convergence  and  two 
lemmas. 

Let  J  be  a  normed  linear  space  and  denote  by  X*  the  collection  of  all  bounded  linear 
functionals  on  X.   If  we  define  the  norm  of  an  element  /  €  J*  by 

11/11=  ^  l/(x)l 

and  define  addition  and  scalar  multiplication  of  linear  functionals  in  the  obvious  manner,  then 
X*  is  a  Banach  space  and  is  commonly  referred  to  as  the  dual  space  of  X.  A  sequence  {x„}  in  X 
is  said  to  converge  weakly  to  x  €  X  if  fixn)  — -  fix)  as  n  — -  °°,  for  every  /  6  X*. 

LEMMA  1:    Let  the  uniformly  bounded  sequence  of  scalar  measurable  functions  [q^it)}, 
d  =  1,2,  ...  ,  converge  weakly  on  [0,71  to  q0it).  Then  except  on  a  set  of  measure  zero 

<7o(')  <   Hm  sup  qdit). 

PROOF:  See  Levinson  [6] 

LEMMA  2:    If  q  is  a  nonnegative  integrable  function  for  which  there  exists  scalar  con- 
stants 9X  ^  0  and  62  >  0  such  that 

qit)  <  6>,  +02J*O'  qis)ds,  0  ^  t  <  T, 

then<?(r)  <  0,/2',    0  <  /  <  T. 

PROOF:  See  Levinson  [6]. 

PROOF  OF  THEOREM  1:   Let  z  be  feasible  for  Primal  Problem  A  and  multiply  the  con- 
straint (2)  by  the  m-dimensional  vector  (1,  . . .  ,  1)  to  obtain  the  inequality 

£  /,(z(f),/)  ^  £  h,(y(z,t),t),  0  <  t  <  T. 

From  the  convexity  of  each  f,  in  its  first  argument,  if  follows  from  [8,  p.  242]  that 
£  fiizit),t)  >  £  /,«)./)  +  £  akit)  zkit), 

;=1  r-1  k=\ 

where 

OfcW  -  £  V*/,<0ff). 


3/0  1    W.   KL1LAINU   AINU   MA.   HArOUIN 

Set  0O  —  max  JO,    sup   I  -  JT  /(0,/)U  by  (9c)  and  observe  that  by  assumption  (6) 
A  =  inf  min  ak(t)  >  0. 

/        k 

Since  z  is  feasible  and  therefore  satisfies  constraint  (1),  it  then  follows  that 
(10)  A  £  zk(i)  ^  90  +  £  /j,(y(z, /),/),    0  <  t  <  T. 

k=  1  /=  1 

Define 

[Vs/tj.s)]  =  (V^Gm)}^,  for  t,  6  £",  s  €  [0,71.  7  =  0,  ...  ,  r, 
and 

[VM^,/)]  =  {Vjhiiy.t)}^,  for*,  €£',/€  [0,71, 


G(z,/,s)  =  y    /(s)    [V/;(y(z,/),/)k(z(5),5) 

7to  [o. '-«./] 


and 


H(z,t,s)  =  Y  /  (s)     [V/;(y(z,/),/)][Vs,(z(s),s)] 

where  /^O)  is  the  indicator  function  of  the  set  E. 

Since  h  and  gj  are  concave  in  their  first  arguments  it  follows  from  [8]  and  from  (3),  (7)  and  (8) 
that 

h(y(z,t),t)  <  MO./)  +  J"  G(0,t,s)ds  +  J"  H(0,t,s)z(s)ds. 

By  (9a)  and  (9b)  we  select  0,  >  0  and  02  >  0,  such  that 
sup   £  /;,(0,/)  +  y   (     Gi(0,t,s)ds\  ^  0, 

'   [£1  PlJo  ) 

and 

sup  max  1 52  Z/,^  (0,/,s)[  ^  02. 
'       k     l'-i  ) 

From  (10)  we  have  that  0*  =  (90  +  9\)/A  and  9$=  9?/ A  are  nonnegative  and  positive  con- 
stants, respectively,  for  which 

X  z,(/)  ^  o?  +  6$  ('£,  zk{s)ds,    o<  /  <  r. 

From  Lemma  2  we  conclude  that 

(11)  £  zyt(/)  <  0f  exp  (0?/)  <  0f  exp  (0J71,  0  <  /  <  T, 

k=\ 

and  hence  the  set  of  feasible  solutions  for  Primal  Problem  A  is  uniformly  bounded  on  [0,71. 

Since  <f>  is  concave  and  differentiable  in  its  first  argument  throughout  [0,71,  it  follows 
from  (9d),  [8]  and  the  uniform  boundedness  property  that,  for  any  feasible  solutions  z  and  z°, 


NONLINEAR  PROGRAMMING  WITH  TIME-DELAYED  CONSTRAINTS  577 

V(z)  -  V(z°)  <  T  £  sup  izkit)  -  zk°it))  sup  Wk<t>iz*it),t)  <  ~ 

k=\       '  ' 

and  hence  V  is  bounded  above  for  all  feasible  z. 

Let  V  =  lub  Viz),  where  the  least  upper  bound  Hub)  is  taken  over  all  feasible  z.    Then 
there  exists  a  sequence  (z^  of  feasible  solutions  such  that 

lim  viz*)  =  V. 

Since  {z^  is  uniformly  bounded,  it  follows  from  [10]  that  there  exists  a  z  to  which  a  subse- 
quence of  {z^  converges  weakly  in  L„2  [0, 71.  Denote  this  weakly  convergent  subsequence 
itself  by  {z^};  the  application  of  Lemma  1  to  each  component  of  zd  then  provides  uniform 
boundedness  for  z  except  possibly  on  a  set  of  measure  zero  where,  as  will  be  shown  later,  it  can 
be  assumed  to  be  zero. 

Since  each  component  of  the  composite  function  hiyi-,t),t)  is  concave  in  z,  it  follows 
from  [8],  (3)  and  the  chain  rule  for  differentiation  that 

hiyizd,t),t)  <  hiyiz,t),t)  +  J'  Hiz,t,s)izdis)  -  zis))ds,    0  ^  t  <  T. 

Since  each  entry  of  the  m  x  n  matrix  Hiz,t,s)  is  bounded  and  measurable,  it  follows  that  each 
row  Hjiz,t,s)  €  L~  [0,71  Q  L}  [0, 71  and  so,  by  weak  convergence, 

f '  Hiz,t,s)  izdis)  -  zis))ds  —  0,  as  d  —  °°. 

Thus,  by  constraint  (2) 

(12)  lim  sup  fizdit),t)  ^  hiyiz,t),t),  0  ^  t  <  T. 

Define  [Vfi-qj)]  =  ((V^(T>,f))mX„,  tj  €  E",  tj  ^  0;  by  the  convexity  of/ 

fizdit),t)  >  fizii),t)  +  [Vfizit),t)]  izdit)  -  2(f)),  0  <  f  <  r. 
Therefore,  from  (12), 

(13)  /(z(/),r)  ^  h(y(z,0,t) 

except  on  a  set  of  measure  zero,  since  by  [8],  assumption  (5)  and  Lemma  1  we  have 
lim  sup  [Vfizit),t)]  izdit)  -  zit))  >  0 

except  on  such  a  set. 

A  second  application  of  Lemma  1  to  each  component  of  zd  provides 
-2(f)  ^   lim  sup(-  zdit))  ^  0,  a.e.m  [0,T], 

d—oo 

and  consequently  z  is  nonnegative  except  on  a  set  of  measure  zero.  From  this  result  and 
expression  (13),  we  observe  that  z  can  violate  the  constraints  of  Primal  Problem  A  on,  at  most, 
a  set  of  measure  zero  in  [0, 71.  We  define  z  to  be  zero  on  this  set  of  measure  zero,  as  well  as 
for  t  <  0,  and  equal  to  2  on  the  complement  of  this  set.  The  feasibility  of  z  is  then  established 
by  noting  that 

y(z,t)  =  y&t),  0  <  t  ^  T, 

and  that 

lim  sup  fizd(t),t)  >  /(0,f),  0  <  f  <  T, 


578  T.W.  REILAND  AND  MA.  HANSON 

by  the  convexity  nstraint  (1),  and  assumption  (6). 

By  the  concavity  and  differentiability  of  $ 

fQ    0  (zd(t),t)dt  ^  fj<t>  izit),t)dt  +  jj  izdit)  -  zit))'  V<f>(zU),t)dt. 
Therefore,  by  weak  convergence 

V  =  lim   f    <t>izdit),t)dt 

^  fQ   <f>(z(t),t)dt  =  Viz). 

By  the  definition  of  V  and  the  feasibility  of  z,  Viz)  <  V,  thus  Viz)  =  K  and  z  is  an  optimal 
solution  for  Primal  Problem  A.  Q.E.D. 

4.   WEAK  DUALITY 

Before  the  dual  to  Primal  Problem  A  is  formally  stated,  a  continuous  time  Lagrangian 
function  and  its  Frechet  differential  will  be  introduced. 

For  u  €  L~  [0,71  and  w  €  L~  [0,71,  define 

(14)  Liu,w)  =  fQT  [<f>iuit),t)  +  w'it)  Fiu,t)]dt 
where 

Fiu,t)  =  hiyiu,t),t)  -  fiuit),t),  0  ^  t  <  T, 

and  let  8iL(«,w;y)  denote  the  Frechet  differential  [7]  with  respect  to  its  first  argument, 
evaluated  at  u  with  the  increment  y  €  L„°°  [0, 71.  The  differentiability  of  each  of  the  functions 
involved  in  L  insures  that  the  Frechet  differential  exists  and  allows  h\Liu,w\y)  to  be  deter- 
mined by  the  simple  differentiation 

(15)  8,Z.(w,w;<y)  =  -j-  Liu  +  ay,w)|Q=0- 

da 

The  Frechet  differential  has  two  additional  properties  that  will  be  used  in  the  ensuing  discus- 
sion, namely,  the  linearity  of  81L(«,w;y)  in  its  increment  y  and  the  continuity  of  b\Liu,w\y) 
in  y  under  the  norm 

\\y\\n  =  max  llyjl00. 
Here  ||-||°°  denotes  the  essential  supremum  [9,  p.  112]. 
If  y  (/)  =  0  for  t  <  0,  then  from  (14)  we  have 

(16)  8,Z,(«fw;y)  =  J0    [[V<f>(uU),t)]'yU) 

+  J*o'  w'it)Hiu,t,s)yis)ds  -W'it)[Vfiuit),t)]yit)}dt. 

An  application  of  Fubini's  theorem  [9]  to  interchange  the  limits  of  integration  enables  us  to 
express  (16)  as 

(17)  8,I(w,w;y)  =  bViu\y)  +  J*Q    y'it)F*iu,w,t)dt, 
where 

8^(w;y)  =  /o    [V<f>(u(t),t)]'y(t)dt 


NONLINEAR  PROGRAMMING  WITH  TIME-DELAYED  CONSTRAINTS  579 

and 

(18)  F*(u,w,t)  =  fT  H'(u,s,t)w(s)ds  -  [Vfiu  (/),/)]'  wit),  0  <  t  <  T. 
With  this  notation  the  dual  of  Primal  Problem  A  will  be  shown  to  be: 

Dual  Problem  A: 

Minimize 

(19)  Giu.w)  =  L(u,w)  -81I(m,w;w) 

subject  to  the  constraints 

(20)  uit),  wit)  >  0,  0  <  t  <  T, 

(21)  F*iu,w,t)  +  [Vcj>(u(t),t)]  <  0,  0  <  t  <  T, 

(22)  uit)  =  0,  t  <  0 
and 

(23)  w(/)  =  0,  /  >  T. 

THEOREM  2  (Weak  Duality):  If  z  and  («,w)  are  feasible  solutions  for  Primal  and  Dual 
Problems  A,  respectively,  then 

VU)  ^  G(u,w). 

PROOF:  By  the  concavity  of  <j>  and  —  /  in  their  first  arguments  and  the  concavity  of  the 
composite  function  hiy{-,t),t)  in  z  it  follows  that  L  is  concave  in  its  first  argument  and 

L(z,w)  —  L(u,w)  <  d\L(u,w;z  —  u). 

Thus, 

Viz)  -  Giu,w)  =  Liz,w)  -f    w'it)Fiz,t)dt 

-  Liu,w)  +8\Liu,w;u) 

<  bxLiu,w\z  -  u)  +  8xLiu,w\u) 

-  f    w'it)Fiz,t)dt 

T 

=  hxLiu,w;z)  -  JQ    w'it)Fiz,t)dt 
by  the  linearity  of  the  Frechet  differential  in  its  increment.   By  (17)  we  have 

8xLiu,w;z)-  fQ    w'it)Fiz,t)dt  =  fQ   z'it){[V<f>iuit),t)]  +  F*iu,w,t)}dt 
~So    »"</)F(2;/)ift 
which  is  nonpositive  by  constraints  (1),  (2),  (20)  and  (21).  Q.E.D. 

From  Theorem  2  it  is  observed  that  if  there  exist  feasible  solutions,  z  and  iu,w),  for  the 
primal  and  dual  problems  and  if  the  corresponding  primal  and  dual  objective  function  values, 
Viz)  and  Giu,w),  are  equal,  then  these  solutions  are  optimal  for  their  respective  problems. 


580  T.W.  REILAND  AND  MA.  HANSON 

5.   THE  CONSTRAINT  QUALIFICATION. 

The  constraint  qualification  introduced  here  is  motivated  by  the  form  of  the  Kuhn-Tucker 
constraint  qualification  presented  by  Zangwill  [11]  and  also  by  Property  1  given  below.  The 
basic  theory  surrounding  this  qualification  is  established  to  provide  a  framework  for  the 
theorems  of  Section  6. 

PROPERTY  1:   If 

(24)  8Viz;y)  =  fj  y'it)V7<t>izit),t)]dt  >  0 

where  z,  y  €  L„°°  [0,  T] ,  then  there  exists  a  scalar  o-  >  0  such  that 
Viz  +  ry)  >   V(z),  for  0  <  r  <  a. 

PROOF:   By  (15)  and  (24) 

lim  [Viz  +  ry)  -  Viz)]/r  =  8  Viz\y)  >  0, 

TiO 

thus  a  positive  o-  can  be  chosen  which  is  sufficiently  small  so  that 

Viz  +  ry)  >   Viz),  for  0  <  t  <  cr.  Q.E.D. 

DEFINITION  1:  For  each  z  which  is  feasible  for  Primal  Problem  A,  define  Diz)  to  be 
the  set  of  ^-vector  functions  y  for  which 

(i)     yeLn°°[0,T] 

(ii)     yit)  =  0,  for  t  <  0 

(iii)    there  exists  a  scalar  o-  >  0  such  that 
zit)  +ryit)  >  0,  0<  t  <  T, 


for 


Fiz  +  ry.t)  ^  0,  0  <  t  <  T, 


0<  - 


DEFINITION  2:  Define  Diz)  to  be  the  closure  of  Diz)  under  the  norm  ||-||~-  that  is,  if 
a  sequence  [y**]  C  Diz)  is  such  that  \\yd  —  y  1 1"  — »  0,  as  d  — -  oo,  then  y  6  Diz). 

Henceforth,  the  Frechet  differential  of  the  mapping  Fi-.t):  L™  [0,71  — -  Em  evaluated  at  z 
and  with  increment  y,  will  be  denoted  by  8Fiz\y),.  It  should  be  observed  that,  for  any 
specified  value  of  t  €  [0,T],  the  existence  of  8Fiz,y),  is  ensured  by  the  differentiability  of/, 
gj,  and  h  and  that  when  y  it)  =  0  for  r  <  0,  we  have 

(25)  8F(z;y),  =  $^  H iz.t.s)  y  is) ds  -  [Vfizit),t)]yit). 

Similarly,  the  Frechet  differential  of  a  component  Fji-,t)  of  Fi,t),  evaluated  at  z  with  incre- 
ment y,  will  be  denoted  by  8Fjiz;y)t. 


NONLINEAR  PROGRAMMING  WITH  TIME-DELAYED  CONSTRAINTS  581 

DEFINITION  3:   For  each  z  which  is  feasible  for  Primal  Problem  A  define  2>(z)  to  be  the 
set  of  ^-vector  functions  y  for  which 

(i)  y  €  L,r  [0,71, 

(ii)  y{t)  =  0,  for  /  <  0, 

(iii)  ykU)^0a.e.    in  Tlk(z),  k  -  1  ... ,  n, 

(iv)  bF,{z\y),  >  0  a.e.  in  T2,(z),  /  =  1 m, 

where 

Tu(z)=  {/  €  [0,7l:z*(f)  =  0},  *-  1,  ...,  /i 
and 

T2i(z)  =  {t  £  [0,71:  F,(zff)  =  0},  /  =  1 m. 


In  a  comparison  of  the  sets  Z)(z)  and  3(z)  with  their  finite-dimensional  counterparts 
presented  in  Zangwill  [11],  it  is  observed  that  D(z)  is  analogous  to  the  set  of  "feasible  direc- 
tions" at  z  and  2(z)  is  analogous  to  that  set  of  directions  for  which  the  directional  derivatives 
of  each  of  the  active  constraints  at  z  are  nonnegative. 

PROPERTY  2:   Z)(z)  C  9{z). 

PROOF:  Part  1.  Let  y  €  D(z).  Then  by  Definition  1,  there  exists  a  scalar  o-  >  0  such 
thatO  <  t  ^  a-  implies  z{t)  +  Ty(t)  >  0,  0  <  t  ^  T.   Thus,  if  zk(t)  =  0,  thenyfc0)  ^  0. 

Assume  that  Fjizj)  =  0.  If  8Fj(z\y),  <  0,  then  by  the  same  technique  used  in  the  proof 
of  Property  1 ,  it  follows  that  for  t  sufficiently  small, 

Fj(z  +Ty,t)  <  F,{z,t)  =  0. 

This   result   contradicts    the   assumption    that   y  €  D(z)    and    therefore    we   conclude    that 
D(z)  C  Mz). 

Part  2.   Assume  that  there  is  a  y  €  L„°°  [0,71  and  a  sequence  {y*}  C  D(z)  such  that  max 

Wyk-VkW00  —  0>  as  rf^oo.   Then  for  all  /such  that  z*(r)  =  0,  yj^(r)  ^  0,  d  =  1,  2,  . . .  .   It 
then  follows  from  convergence  in  L°°  [0,71  that  yk(t)  ^  0  a.e.  on  T]k(z),  k  =  1,  . . .  ,  «. 

Assume  there  exists  an  /  and  a  set  E  of  positive  measure  over  which  Fj(z,t)  =  0  and 
8Fj(z;y),  <  0  for  all  /  €  £.  By  the  continuity  of  87](z;-),  in  the  L°°  norm  [7],  we  can  choose 
a  J*  sufficiently  large  such  that  for  d  ^  d* 

8/;-(z;y<0,  <  0 

over  some  subset  of  £  which  has  positive  measure.   This  result  yields  _a  contradiction  to  Part  1 
since  it  was  assumed  {y^  C  D(z)  and  we  can  therefore  conclude  that  D(z)  c  2)(z).      q  E  D 

DEFINITION  4  (Constraint  Qualification):  Primal  Problem  A  will  be  said  to  satisfy  the 
Constraint  Qualification  if  the  problem  is  feasible  and  if 

D  (z)  =9<z), 


582  T.W   REILAND  AND  MA   HANSON 

where  z  is  an  optimal  solution  to  the  problem. 

In  more  general  problems  where  convexity  and  concavity  properties  are  not  assumed,  the 
purpose  of  the  Constraint  Qualification  would  be  to  eliminate  "cusps"  in  the  feasible  region. 
For  example,  the  constraints 

z,(/)  ^  0,  z2U)  ^  0,  0  <  t  <  T, 

and 

[I  -  ZlU)V  -  z2U)  >  0,  0<  t  <  T, 

do  not  satisfy  the  Constraint  Qualification  when  z(t)  =  (1,0),  0  <  /  <  T,  since 
(1/2,0)  €  #(z)  but  (1/2,0)  <t  D{z). 

In  problems  such  as  Primal  Problem  A  where  convexity  and  concavity  properties  are 
assumed,  violations  of  the  Constraint  Qualification  can  be  shown  to  arise  when  the  constraints 
take  the  form  of  equalities  on  some  set  of  positive  measure.  For  example,  consider  the  con- 
straints 

ZlU)  >  0,  z2(t)  >  0,  0  <  t  <  T, 

and 

biW  +  z2(0  -  l]2  <  l  -  IEU),  0  <  t  <  T, 

where  £  is  a  set  of  positive  measure  in  [0,T]  and  IE{)  is  its  indicator  function.  It  is  observed 
that  for  z(t)  =  (1/2,1/2),  we  have  (1,1)  6  9{z)  but  (1,1)  $  D(z),  thus  the  Constraint 
Qualification  is  not  satisfied. 

THEOREM  3:  If  z  is  optimal  for  Primal  Problem  A,  then  under  the  Constraint 
Qualification 

8  K(z;y)  <  0,  for  all  y  <E  2{z). 

PROOF:  Part  1.  Suppose  there  exists  a  y  6  D(z)  such  that  8  V(z;y)  >  0.  Then  by  Pro- 
perty 1  there  exists  a  o-  >  0  such  that  0  <  t  ^  o-  implies  V(z  +  ry)  >  V(z);  however,  since 
y  €  D(z)  we  can  choose  a  <r0  sufficiently  small  so  that  z  +  a-0y  is  feasible  for  Primal  Problem 
A.  Thus,  by  contradiction  of  the  optimality  of  z,  we  can  conclude  that  8  V(z;y)  <  0,  for  all 
y  €  D(l). 

Part  2.    Let  {y^   be  a  sequence  of  functions  in  D(z)  and  let  y°  be  such  that  max 

k 

I  \y(  ~  7  k  1 1°°  ~" "  0,  as  d  — *  oo.   It  then  follows  from  Part  1  and  the  continuity  of  8  K(z;)  that 

8  K(z>°)  =  lim  8  K(z><0  ^  0. 
Thus,  8  K(z;y)  <  0  for  all  y  <E  5  (z).  Q.E.D. 

6.   DUALITY  AND  RELATED  THEOREMS 

In  proving  strong  duality  and  its  related  theorems  two  additional  assumptions  will  be 
made.  These  are: 

(26)  H(z,t,s)  ^  0,    0  <  s  ^  t  <  T 


NONLINEAR  PROGRAMMING  WITH  TIME-DELAYED  CONSTRAINTS  583 

and 

(27)  F(z,t)  -  8F(z;z),  >  0,    0  <  t  <  T, 

where  z  is  an  optimal  solution  for  Primal  Problem  A.    It  will  be  shown  in  Corollary  1  that 
assumption  (26)  is  implied  if  z  (r)  =  0  is  feasible. 

THEOREM  4  (Strong  Duality):  Under  the  Constraint  Qualification  and  assumptions  (26) 
and  (27),  there  exists  an  optimal  solution  (u,w)  for  Dual  Problem  A  such  that  u  =  z  and 
G(z,w)=  V{z). 

Before  proving  Theorem  4  the  following  linearized  problem,  called  Primal  Problem  A',  will 
be  considered: 

Maximize 

8  V(z;z  -  z) 

subject  to  the  constraints 

(28)  z(t)  >  0,  0<  /  ^  T, 

(29)  Fat)  +  8F(z;z  -  z)t  >  0,  0  <  t  <  T, 
and 

(30)  z(r)  =  0,  for/  <  0. 

LEMMA  3:  Under  the  Constraint  Qualification,  z  is  an  optimal  solution  for  Primal  Prob- 
lem A'. 

PROOF:  If  z  is  feasible  for  Primal  Problem  A',  then 

Ht)  -  z(t)  >  0,  for  t  €  Tlk(z),  k=\,  ....  n, 

and 

8Fj(z,z  -  z),  ^  0,  for  /  €  T2i(z),  i  =  1,  . . .  ,  m, 

and  therefore  (z  —  z)  €  2(j).    It  then  follows  from  Theorem  3  that,  under  the  Constraint 
Qualification, 

8  V(z;z  -  z)  <  0 

for  all  z  satisfying  (28) ,  (29)  and  (30) .  The  optimality  of  z  follows  since  z  is  feasible  for  Primal 
Problem  A'  and  since  8  V(z;0)  =  0.  0  E  D 

PROOF  OF  THEOREM  4:  We  rewrite  Primal  Problem  A'  in  the  form 
maximize 


J0    a'(t)z(t)dt 


subject  to  the  constraints 

zit)  >  0,  0<  t  <  T, 


584  T.W.  REILAND  AND  MA.  HANSON 

and 

B(t)zit)  ^  c(t)  +  fi  KU,s)z(s)ds,  0  <  t  <  T, 

where  ait)  =  [V0(z(f),/)1,  Bit)  =  [Vfizit),t)],  cit)  =  Fiz.t)  -  8F(z;z)t,  and  Kit,s)  = 
Hiz,t,s).  From  assumptions  (5),  (6),  (26)  and  (27)  it  is  observed  that  the  matrices  ait), 
Bit),  cit)  and  K(t,s)  satisfy  the  requirements  of  Grinold's  Duality  Theorem  [3].  Therefore, 
there  exists  an  m- vector  function  w  satisfying 

(31)  wQ)  ^  0,  0  ^  t  ^  T, 

and 

(32)  B' it)  wit)  ^  a(f)  +  J    £'(«*)  w(s)4  0  <  t  <  r, 

such  that 

(33)  .C^')  c(t)dt  =  fj a'(t)z(t)dt. 

Setting  wit)  =  0  for  t  >  T,  we  observe  from  the  identities  (14),  (17),  and  (18)  that  expres- 
sions (32)  and  (33)  can  be  expressed  as 

(320  F*iz,w,t)  +  tV0(F(f),/)]  <  0,  0  <  t  <  T, 

and 

(330  L(z,w)  -  hxLiz,w;z)  =  Viz), 

respectively.  From  (31)  and  (320  and  the  fact  that  wit)  =  0  for  t  >  T,  it  then  follows  that 
iz,w)  is  feasible  for  Dual  Problem  A  and,  from  (19)  and  (330 

(34)  Giz,w)  =  Viz). 

Finally,  by  the  weak  duality  established  in  Theorem  2,  it  is  concluded  from  (34)  that  iz,w)  is 
an  optimal  solution  for  Dual  Problem  A.  Q.E.D 

In  order  to  apply  Theorem  4  in  practice,  it  is  desirable  to  be  able  to  verify  conditions  (26) 
and  (27)  without  prior  knowledge  of  the  optimal  solution  z.  The  following  corollary  provides 
this  capability. 

COROLLARY  1:   If 

VkgjM,t)lfrqk  =  dgj,(ri.t)/B7ik  >  0, 

(35)  j  =  0 r,  i  =  1,  ...  ,  p,  k=\,  ...  ,  n, 

for  tj  €  E",  ri  ^  0,  and  0  <  t  <  T, 

(36)  Fi0,t)  ^  0,    0  <  /  <  T, 

then  under  the  Constraint  Qualification  there  exists  an  optimal  solution  iu,w)  for  Dual  Problem 
A  such  that  u  =  zand  Giz.w)  =  Viz). 

PROOF:  We  have  from  (8)  and  (35)  that 

Hiz.t.s)  -  £  /     is)    [Vhiyiz,t),t))  [Vgjizis),s)]  >  0,  0  <  t  <  T, 

and  by  (36)  and  the  concavity  of  F  that 

Fiz,t)  -  8Fiz;I),  >  FiO.t)  >  0,    0  <  t  <  T. 


NONLINEAR  PROGRAMMING  WITH  TIME-DELAYED  CONSTRAINTS  585 

From  these  results  it  follows  that  the  conditions  of  Theorem  4  are  satisfied.  Q.E.D. 

THEOREM  5  (Complementary  Slackness  Principle):   If  z  and  (z,w)  are  optimal  solutions 
for  the  Primal  and  Dual  Problems  A,  then 


(37)  fQTw'(t)Fat)dt  =  0 
and 

(38)  fQTz'(t){F*(z,w,t)  +  [V<f>GU),t)])dt  =  0. 


PROOF:  Since  z(t)  ^  0  and  F*{z,w,t)  +  [V0(z(f),f)]  <  0,  0  <  t  <  T,  it  follows  from 
identity  (17)  that 

J0    z'(t){F*(z,w,t)  +  [V<f>a(t),t)]}dt  =  8xL(z,w;z)  ^  0, 
and  therefore,  by  (330 

L(z,w)  -  V(z)  =  fj  w'(t)F(z,t)dt  <  0. 
Since  w(t)  ^  0  and  F(z,t)  ^  0,  0  ^  t  <  T,  it  also  follows  that 
(39)  J*    w'U)F(z,t)dt  ^  0, 

thus  the  equality  in  (37)  is  established. 

Similarly,  (33')  and  (39)  imply  that 

8,£(z,w;z)  >  0 
and  therefore,  by  (17) 

fQTz'U){F*(z,w,t)  +  [V<t>(zU),t)])dt  >  0. 
Since  z(t)  ^  0  and  F*(z,w,t)  +  [V<f>(z(t),t)]  ^  0,  0  <  t  <  T,  we  have 

Jorr(f){F*(z,^)  +  [V0(z(/),/)]}<ft  <  0 
and  thus  the  equality  in  (38)  is  established.  Q.E.D. 

THEOREM  6  (Kuhn-Tucker  Conditions):  Assume  that  (35)  and  (36)  are  satisfied  for 
Primal  Problem  A.  Then  under  the  Constraint  Qualification  z  is  an  optimal  solution  if  and  only 
if  there  exists  an  m- vector  function  w  such  that 

(i)     F*(z,w,t)  +  [V0(z(f),f)l  <  0,  0  <  t  <  r, 

(ii)    fJz'UHFHZw.t)  +  [V<f>(z(t),t)]}dt  =  0 

(iii)   J0    w'(t)F(z,t)dt  -  0 

(iv)    »(/)  >  0,  0  <  /  <  Tand  w(f)  =  0,  /  >  T. 

PROOF: 

Necessity:  The  necessity  of  the  conditions  follows  from  Corollary  1  and  Theorem  5,  since 
the  m-vector  function  vv  of  the  optimal  solution  (z,w)  to  Dual  Problem  A  satisfies  conditions  (i) 
through  (iv). 


586  T.W    REILAND  AND  MA.  HANSON 

Sufficiency:   Let  z  be  feasible  for  Primal  Problem  A.   Then  since  Fis  concave 
Viz)  -  Viz)  <  8  V(z;z  -  z) 

=  Jj  [zit)  -Iit)]'[V<t>izit),t)]dt. 
Since  z(f)  ^  0,  0  <  /  <  T,  it  follows  from  conditions  (i)  and  (ii)  that 
Viz)  -  Viz)  ^  -   f     hit)  -  zit)]'F*iz,w,t)dt, 

♦'O 

and  by  (18),  (25)  and  Fubini's  Theorem  [9] 

fo    [zit)-zit)]'F*iz,w,t)dt  =  fo    w'it)iFiz;z-z)tdt. 
By  (i),  (iii)  and  the  concavity  of  F, 

-  J*Q    w'it)8Fiz;z  -  z),dt  <  -  /     w'it)[Fiz,t)  -  Fiz,t)]dt 

=  -  X    *'it)Fiz,t)dt 

which  is  nonpositive  since  wit)  ^  0  and  Fiz,t)  ^  0,  0  ^  t  <  T.  Thus,  Viz)  <  Viz)  and  zis 
an  optimal  solution  for  Primal  Problem  A.  Q.E.D. 

7.    EXAMPLE  -  WATER  STORAGE  PROBLEM 

In  the  water  storage  problem  posed  in  [4],  the  hydroelectric  company  incurred  a  penalty  if 
it  could  not  meet  a  prescribed  demand  for  power.  This  penalty  was  characterized  in  the  objec- 
tive function 

jJ^iDit)-  Pit))dt 

where  [0,71  represents  a  planning  period  of  specified  duration,  Dit)  is  the  demand  rate,  Pit) 
is  the  production  rate  of  hydroelectric  power,  and  «/»  is  the  penalty  function  which  was  assumed 
to  be  strictly  convex.  The  imposition  of  such  a  penalty  favors  the  consumer  or  a  middleman 
utility  company  which  retails  electric  power  to  the  consumers.  In  short,  it  characterizes  a 
"buyers  market." 

If  there  is,  in  fact,  a  pending  energy  crisis,  it  seems  appropriate  to  consider  a  "sellers 
market"  where  the  demand  for  power  exceeds  production  capacity  and  a  premium  is  paid  to  the 
hydroelectric  company  for  any  power  which  it  produces  beyond  some  prescribed  level.  In  the 
case  where  the  hydroelectric  company  is  supplying  power  directly  to  the  consumer,  these  premi- 
ums may  take  the  form  of  increasing  prices  per  unit  beyond  some  allotment  level.  When  the 
hydroelectric  company  is  supplying  a  middleman,  the  premiums  may  represent  an  incentive  pol- 
icy which  encourages  maximum  production  during  peak  demand  periods. 

The  premiums  to  the  hydroelectric  company  will  be  represented  by 

fjir  iPit)-  Ait))dt 

where  [0,71  represents  the  planning  period,  Pit)  is  the  power  production  rate,  Ait)  is  the 
prescribed  aggregate  allotment  or  incentive  level,  and  it  is  the  premium  function  which  is 
assumed  to  be  differentiable  and  concave  with  a  positive  slope  at  zero. 

For  the  dynamics  of  the  problem,  we  assume  a  confluent  system  of  rivers  supplying  water 
to  a  hydroelectric  plant  on  the  main  stream  with  r  of  its  tributaries  also  having  their  own 


NONLINEAR  PROGRAMMING  WITH  TIME-DELAYED  CONSTRAINTS  587 

hydroelectric  plants.    The  variables  and  parameters  which  relate  to  the  dam,  reservoir  and  plant 
on  the  main  stream  will  be  subscripted  by  0,  and  those  for  the  r  dammed  tributaries  by  j, 

j-  1.   ....  r. 

We  let  Cl  j  denote  the  initial  store  of  water  in  reservoir  j  and  0;  the  capacity  of  reservoir  j. 
The  rate  of  spillage  and  rate  of  discharge  through  the  turbines  of  dam  ./at  time  /are  denoted  by 
S/(t)  and  dj(t),  respectively.  The  rates  of  inflow  of  water  into  the  reservoirs  on  the  dammed 
tributaries  are  £,(/)  7=1,  ...  ,  r,  and  that  into  the  main  reservoir  from  its  undammed  tribu- 
taries is  £o(t). 

It  is  assumed  that  it  takes  aJt  j  =  1,  .. .  ,  r  units  of  time  for  the  water  released  from  dam 
j  to  reach  the  main  reservoir  and  that  there  is  no  spillage  or  discharge  through  the  dams  on  the 
tributaries  for  at  least  a  units  of  time  prior  to  the  start  of  the  planning  period,  where 
a  =  max  {a,}.   The  store  of  water  in  reservoir  j  at  time  rcan  then  be  expressed  as 

Wj(t)  =    ft;  +  J0'  (gjW)   ~  Sj(t')  -  dj(t'))dt' 

for  j  =  1 ,  . . .  ,  r,  and 

w0(t)  - 

for  the  main  reservoir. 


:  fto  +  /0'|fo('')  -  s0W)  -  d0W)  +  £  (sjU'-ccj)  +  djU'-aj))\dt' 


The  power  production  rate  for  a  given  rate  of  discharge  d  is  assumed  to  be  proportional  to 
d.  In  [4],  it  was  necessary  to  assume  the  factor  of  proportionality  to  be  unity.  Here  we  allow 
this  factor  to  be  proportional  to  the  head  of  water  in  the  reservoir,  an  assumption  which  is  con- 
sistent with  constant  turbine  efficiency.  The  head  is  the  difference  h  between  the  surface  level 
of  the  reservoir  and  the  tailwaters  below  the  dam  and  is  therefore  dependent  primarily  upon  the 
store  of  water  Win  the  reservoir. 

The  relationship  between  hj(t),  the  head  of  reservoir  j,  and  Wj(t)  will  be  represented  by 
hj(t)  =  h*  (Wj(t)),  where  h*  is  an  increasing  concave  differentiable  function.  The  functions 
h*  owe  their  concavity  to  the  shapes  of  the  reservoirs  which  are  assumed  to  yield  a  continu- 
ously disproportionate  increase  in  reservoir  surface  area  as  the  store  of  water  increases.  The 
production  rate  for  the  yth  hydroelectric  plant  is  then  expressible  as 

Pj(t)  =  dj(t)  °  h*  (WjU)) , 

in  which  case  the  production  rate  for  the  entire  system  becomes 

PU)=  £pjU). 


Assuming  the  role  of  the  hydroelectric  company,  we  want  to  select  our  water  storage  pol- 
icy (s,d)  so  as  to  maximize  the  premium  payments  over  the  planning  period.  This  problem 
takes  the  form 


ris.d)  =  f    ir (Pit)  -  A(t))dt 

•'O 


588  T.W.  REILAND  AND  MA    HANSON 

subject  to 

0  ^  SjQ)  <  pj(t) 
0  ^  dj(t)  ^  <t>, 
0  <   WjU)  ^  9j 

./  =  0,  ....  r,  and 

A(t)  <  Pit) 

for  0  ^  /  ^   T,  where  /3,-(/)  is  the  maximum  allowable  spillage  rate  through  dam  7  and  </>,  is  the 
turbine  capacity  of  plant  / 

Through  proper  association  of  the  terms  of  this  model  with  those  of  Primal  Problem  A  it 
can  be  shown  through  application  of  Theorem  1  that  feasibility  ensures  the  existence  of  an 
optimal  water  storage  policy  which  will  maximize  the  total  premium  payment. 

REFERENCES 

[1]  Farr,  W.H.  and  M.A.  Hanson,  "Continuous  Time  Programming  with  Nonlinear  Con- 
straints," Journal  of  Mathematical  Analysis  and  Applications  45,  96-115  (1974). 

[2]  Farr,  W.H.  and  M.A.  Hanson,  "Continuous  Time  Programming  with  Nonlinear  Time- 
Delayed  Constraints,"  Journal  of  Mathematical  Analysis  and  Applications  46,  41-60 
(1974). 

[3]  Grinold,  R.,  "Continuous  Programming  Part  One:  Linear  Objectives,"  Journal  of 
Mathematical  Analysis  and  Applications  28,  32-51  (1969). 

[4]  Koopmans,  T.C.,  "Water  Storage  in  a  Simplified  Hydroelectric  System,"  Proceedings  of  the 
First  International  Conference  on  Operational  Research,  M.  Davies,  R.T.  Eddison  and  T. 
Page,  Editors  (Operations  Research  Society  of  America,  Baltimore,  1957). 

[5]  Kuhn,  H.W.  and  A.W.  Tucker,  "Nonlinear  Programming,"  Proceedings  of  the  Second  Berke- 
ley Symposium  on  Mathematical  Statistics  and  Probabilities,  481-492,  J.  Neyman,  Editor 
(University  of  California  Press,  Berkeley,  1951). 

[6]  Levinson,  N.,  "A  Class  of  Continuous  Linear  Programming  Problems,"  Journal  of 
Mathematical  Analysis  and  Applications  16,  73-83  (1966). 

[7]  Luenberger,  D.G.,  Optimization  by  Vector  Space  Methods  (Wiley,  New  York,  N.Y.,  1969). 

[8]  Rockafellar,  R.T.,  Convex  Analysis  (Princeton  University,  Princeton,  New  Jersey,  1970). 

[9]  Royden,  H.L.,  Real  Analysis  (MacMillan,  New  York,  1968). 
[10]  Taylor,  A.E.,  Introduction  to  Functional  Analysis  (Wiley,  New  York,  N.Y.,  1958). 
[11]   Zangwill,   W.I.,   Nonlinear  Programming:  A    Unified  Approach  (Prentice  Hall,  Englewood 
Cliffs,  New  Jersey,  1969). 


EQUALITIES  IN  TRANSPORTATION  PROBLEMS  AND 
CHARACTERIZATIONS  OF  OPTIMAL  SOLUTIONS* 

Kenneth  O.  Kortanek 

Department  of  Mathematics, 
Carnegie-Mellon  University 
Pittsburgh,  Pennsylvania 

Maretsugu  Yamasaki 

Department  of  Mathematics 

Shimane  University 

Matsue,  Shimane,  Japan 

ABSTRACT 

This  paper  considers  the  classical  finite  linear  transportation  Problem  (I)  and 
two  relaxations,  (II)  and  (III),  of  it  based  on  papers  by  Kantorovich  and  Rubin- 
stein, and  Kretschmer.  Pseudo-metric  type  conditions  on  the  cost  matrix  are 
given  under  which  Problems  (I)  and  (II)  have  common  optimal  value,  and  a 
proper  subset  of  these  conditions  is  sufficient  for  Problems  (II)  and  (III)  to 
have  common  optimal  value.  The  relationships  between  the  three  problems 
provide  a  proof  of  Kantorovich's  original  characterization  of  optimal  solutions 
to  the  standard  transportation  problem  having  as  many  origins  as  destinations. 
The  results  are  extended  to  problems  having  cost  matrices  which  are  nonnega- 
tive  row-column  equivalent. 


1.   INTRODUCTION  WITH  PROBLEM  SETTING 

Over  25  years  ago  Kantorovich  in  his  classic  paper,  "On  the  translocation  of  masses"  [4], 
formulated  generalized  transportation  problems  which  are  continuous  analogs  of  the  well-known 
transportation  problem  in  the  theory  of  finite  linear  programming.  He  raised  the  question  of 
characterizing  optimal  solutions  to  those  problems  whose  finite  dimensional  versions  have  the 
same  number  of  origins  as  destinations.  As  is  well  known,  optimal  solutions  to  the  standard 
finite  dimensional  transportation  problem  having  "m"  origins  and  "n"  destinations  are  charac- 
terized by  means  of  a  system  of  linear  inequalities  involving  m  row  numbers  and  n  column 
numbers  which  together  comprise  a  feasible  list  of  dual  variables. 

Within  the  finite  dimensional  context  m  =  n,  Kantorovich's  goal  was  to  use  only  n 
numbers  in  a  linear  inequality  system  characterization  of  an  optimal  solution  rather  than  the 
standard  In  (row  plus  column)  numbers.  In  order  to  accomplish  this,  three  conditions  defining 
a  pseudo-metric  were  imposed  on  the  cost  coefficient  matrix.  Actually,  the  triangle  inequality 
condition  on  unit  costs  is  what  Gomory  and  Hu  later  termed  "reasonable  costs"  in  their  network 

*The  work  of  the  first  author  was  supported  in  part  by  National  Science  Foundation  Grants  ENG76-05191   and 
ENG78-25488. 


590  K..O.  KORTANEK  AND  M.  YAMASAKI 

studies  [3],  Section  2.   Violation  of  this  particular  condition  is  also  related  to  the  "more  for  less" 
paradox  in  the  transportation  model,  see  Ryan  [7]. 

The  original  application  of  the  pseudo-metric  conditions  involved  subtleties  which  were 
later  clarified  in  Kantorovich-Rubinstein  [5]  but  for  a  transformed  version  of  the  standard  tran- 
sportation problem,  which  we  state  as  Problem  III  in  the  next  section.  In  attempting  to  give  a 
proof  of  Kantorovich's  characterization,  Kretschmer  [6]  introduced  yet  another  transformation 
of  the  standard  problem,  which  we  shall  term  Problem  II  in  the  next  section. 

The  basic  purpose  of  this  paper  is  to  delineate  the  key  relationships  between  these  three 
problems:  the  standard  transportation  Problem  I,  the  Kretschmer  transformed  Problem  II,  and 
the  Kantorovich-Rubinstein  Problem  III.  The  results  we  obtain  depend  on  how  the  three 
pseudo-metric  cost  conditions,  denoted  (C.l)  through  (C.3)  in  Sections  3  and  4,  are  coupled 
together. 

Our  main  application  is  to  obtain  a  proof  of  the  originally  sought  for  characterization  of 
optimal  solutions  of  the  standard  transportation  problem  where  the  number  of  origins  equals 
the  number  of  destinations.  We  are  not  prepared  at  this  time  however  to  state  that  we  have 
industrial  or  public  sector  applications  of  the  type  II  or  type  III  transportation  models. 

2.   THE  KANTOROVICH-RUBINSTEIN  AND  KRETSCHMER  TRANSFORMS 
OF  THE  STANDARD  TRANSPORTATION  PROBLEM 

Let  Cy,  a,  and  bj(i  —  1,  ....  n\  j  =■  1,  . . .  ,  n)  be  nonnegative  real  numbers  and  assume 
that  a,  and  bj  satisfy 

(1.1)  X>/=I>,>0. 
r-i         j=\ 

The  original  transportation  problem  may  be  expressed  as  follows: 

(I)  Determine  the  minimum  value  Mof 

(1.2)  ttdjXu 

i-l  7=1 

subject  to  the  condition  that  xtj  are  nonnegative  and 

(1.3)  jtxij-a,    (/=1 n), 

7=1 

£*/,  =  bj    (/-  1,  ...  ,  n). 

i=\ 

Let  us  consider  the  following  transportation  problems  "which  were  studied  in  [5]  and  [6]: 

(II)  Determine  the  minimum  value  N  of 

(1.4)  H^M+yy) 

1=1  7=1 

subject  to  the  condition  that  xtj  and  y^  are  nonnegative  and 

(1.5)  Z  (*,,->>/,)  =  a,     0-1,  ....  n), 

7=1 

lt(xij-yij)  =  bj     0-1,  ....  n). 


EQUALITIES  IN  TRANSPORTATION  PROBLEMS  591 

(III)   Determine  the  minimum  value  V  of 

(1.6)  ttcuzu 

i-l  7-1 

subject  to  the  condition  that  zy-  are  nonnegative  and 

(1.7)  £  Zjj  -  £  zy,  =  a,  -  bt    0=1 n). 

7=1  7=1 

Program  I  of  course  is  the  classical  transportation  problem  which  may  be  solved  by  the 
well-known  row  and  column  number  method  ([1],[2])  and  other  more  modern,  large  scale  pro- 
gramming methods.  The  row  and  column  number  method  easily  extends  to  solving  Program  II. 
On  the  other  hand,  the  structural  matrix  of  Program  III  is  a  network  incidence  matrix,  and  so 
III  is  an  uncapacitated  network  problem. 

It  is  clear  that  V  <  M  and  N  ^  M  and  in  this  sense  Problems  II  and  III  are  relaxations  of 
Problem  I.  We  shall  study  when  one  of  the  equalities  V  =  N,  V  =  A/,  and  M  =  N  holds. 

3.  THE  EQUALITY  N  =  V  OF  PROBLEMS  II  AND  III 

First  we  have 

LEMMA  1:  The  inequality  V  <  N  holds  if  the  following  condition  is  fulfilled: 
(C.l)  Cy  =  Cjj  for  all  i  and  j. 

PROOF:  There  exists  an  optimal  solution  xy  and  jvy  of  Problem  (II),  i.e.,  xi}  and  y,j  are 
nonnegative  and  satisfy  (1.5)  and 

i-I  7=1 

Then 

\Lxij  +  i,yji\  ~  It,**  +  !>//   -  a>  ~  bi- 
[j-i        y-i    J     u-i        y-i    j 

Taking  zy  =  xu  +  yJh  we  see  that  zu  are  nonnegative  and  satisfy  (1.6),  so  that  by  condition 
(C.l) 

y<tlcuzu  =  N. 

i-l  7-1 

THEOREM  1:  The  equality  V—  N  holds  if  condition  (C.l)  and  the  following  condition 
are  fulfilled: 

(C.2)  c„  =  0  for  all  /'. 

PROOF:  There  exists  an  optimal  solution  zy  of  Problem  (III),  i.e.,  zti  are  nonnegative 
and  satisfy  (1.7)  and 

/-I7-1 


592  K.O.  KORTANEK  AND  M    YAMASAKI 

Then 

£  z,j  +  b, ,  -  £  z7,  +  a,  =  4  >  0. 

7=1  7=1 

Let  us  take  Xy  =  0  if  /  ^  y  and  x„  =  d,  and  put  jfy  =  Zyt.  Then  xy  and  jty  are  nonnegative  and 
satisfy  (1.5),  so  that 

*  s  £  £  cMj  +  .v-y)  -  £  £  djZji  =  y 

by  conditions  (C.l)  and  (C.2). 

We  show  by  an  example  that  the  equality  N  =  V  does  not  hold  in  general  if  we  omit  con- 
dition (C.2). 

EXAMPLE  1:   Let  n  =  2  and  take 

C\\  =  c22=  1,       c12=  c2i  =  2, 
a\  —  1,    a2  =  2,    6]  =  2,    61  =  1. 
Then  we  easily  see  that  K  =  2  and  Af  =  TV  =  4. 

4.  THE  EQUALITY  M  =  N  OF  PROBLEMS  I  AND  II 

Next  we  show  that  the  equality  M  =  N  does  not  hold  in  general  even  if  both  conditions 
(C.l)  and  (C.2)  are  fulfilled. 

EXAMPLE  2:   Let  n  =  3  and  take 

cu  -  c22  -  c33  -  0,     c12=c21  =  20, 

Cl3  =  C31  =  C23  =  C32  =   1, 

fl!  =  3/2,  a2=  1/2,  a3=  1/4, 
bx=  b2=  1,  63  =  1/4. 

optimal  solution  of  Problem  (I)  is  given  by  xn  =  1,  x22  =  1/2  x12  =  x13  =  x32  =  1/4  and 
*2i  =  *3i  -  *23  =  *33  =  0-  we  have  N  =  I.  An  optimal  solution  of  Problem  (II)  is  given  by 
*n  =  1,  *22  =  1/2,  xu  =  x32  =  1/2,  x12  =  x2,  =  x23  =  x31  =  x33  =  0,  y33  =  1/4  and  yu  =  0  if 

ay)  *  (3,3). 

Our  main  result  is  the  following  one. 

THEOREM  2:  The  equality  M  =  N  holds  if  the  following  condition  is  fulfilled: 
(C.3)  qj  <  ciq  +  cpj  +  cw  for  all  /,  j,  p,  q. 

PROOF:  There  exists  an  optimal  solution  Xjj  and  yu  of  Problem  (II).  In  case 
Zjj  =  Xjj  —  yjj  is  nonnegative  for  each  i,  /,  we  see  that  z^  is  a  feasible  solution  of  Problem  (I) ,  so 
that 


A/<II  cijzij  <££  cu(x0  +  y0)  =  N. 
'=1  7=1  '-1  7=1 


EQUALITIES  IN  TRANSPORTATION  PROBLEMS  593 

We  consider  the  case  where  some  x/y  -  ytj  are  negative.   We  may  assume  that  min(x,7,  yu)  =  0 
for  all  /  and  /   There  exist  p  and  q  such  that  0  =  xpq  <  ypq.   Then  we  have  by  (1.5) 

Z  xiq  kypq  and  £  xpj  ^  yp,,. 

/=1  7=1 

Let  us  define  Ah  Bj  and  dy  by 
Ap  =  Bq  -  0, 

A  =  *fcW  Z  x«?  U  *  p),    Bj  =  xwj>w/  X  Jf/y  0"  *  Q), 

flfc  -  AtBjly„. 
Then 

(4.1)  Z4/  =  ^  <*,<,,    td0=Bj<xPJ> 

/-i  ,=i 

(4.2)  i^-i^-jw- 

,-i  y-i 

We  define  x,,  and  ^  by 

(4.3) 


X'ij  =  Xij  +  dy 

if  i  ^  p  and  y  ^  #, 

Xpj  =  Xpj  —  Bj 

if  /  *  9, 

Xjq         Xjq         Aj 

if  /  ^  a 

y\j  =  y>j 

if  i  ^  p  or  y  ^  9, 

x'n-y'n-  0. 

Then  x\j  and  v,y  are  nonnegative  and  satisfy  (1.5)  and 

N^tt  Cijixlj+ylj) 

/-I  7=1 

=  11  CijiXij  +  yu)  +  £  £  rfylcy  -  c„  -  cOT  -  c,,] 
/=iy=i  ,-iy-i 

^ZZ  tyCxfc+j^-tf 

/=1  7=1 

by  condition  (C.3).   Repeating  the  above  procedure  (4.3)  a  finite  number  of  times,t  we  obtain 
xfj  which  are  nonnegative  and  satisfy  (1.3)  and 

^  =  Zl¥5-ii  CM  +  y^  =  n. 

/-I  7-1  '-1  /-I 

Hence,  M  =  N. 

THEOREM  3:    Let  ky,  ft  and  gj  be  nonnegative  numbers  and  assume  that  condition 
(C.3)  holds  for  ky  instead  of  Cy.  If  Cy  =  k0  +  /,  +  #,,  then  M  =  N. 


hw,' 


This  number  is  at  most  the  number  of  {>»y  >  0} . 


594  K.O   KORTANEK  AND  M    YAMASAKI 

PROOF:    Denote  by  M(k)  and  Nik)  the  values  of  Problems  (I)  and  (II)  respectively  if 
Cjj  are  replaced  by  ktj.  Then  we  have  M  =  M(k)  +  C/g  with 

cfs  =  £  Mi  +  £  gjbj. 
<=i  j-l 

Let  Xjj  and  v,y  be  nonnegative  and  satisfy  (1.5).  Then 

1-1 /-I  /=U=1 


^11  fcy(xy  +  j/y)  +  C/g  >  AT(/r)  +  CA. 

1=1  y=l 

Thus,    N  £  iV(Ar)  +  Cfg.     Since    #(*)  =  M(k)    by   Theorem   2,    we    have   N  ^    Af(A:)  + 
Cfg  =  M,  and  hence  M  =  N. 

As  is  well  known,  two  transportation  problems  of  type  (I)  with  costs  [cy]  and 
[cjj  +  fj  +  gj)  respectively,  are  equivalent  for  any  list  of  real  numbers  {/}},  {#,},  /'  =  1,  ....  m; 
j  —  1,  . . .  ,  «.  The  following  example  shows  that  the  nonnegativity  of  all  the  //  and  gj  is 
required  in  Theorem  3. 

EXAMPLE  3:  Let  n  =  2  and  take  fc„  =  0,  kn  =  1/2,  fc21  =  1/2,  k22  =  0,  a,  =  1,  a2  =  1, 
bx  =  1/2,  b2  =  3/2,  /,  =  1,  f2  =  -5/2,  ?j  =  2,  g2  =  7/2.  Then,  Jl/(fc)  =  N(k)  -  1/4  while 
AT  -  9/2  <  5  -  Af. 

5.  KANTOROVICH'S  THEOREM  FOR  PROBLEM  (I) 

The  finite  version  of  Kantorovich's  Theorem  [4]  can  be  written  as  follows: 

A  feasible  solution  xtJ  of  Problem  (I)  is  an  optimal  solution  if  and  only  if  there  exist 
numbers  w,  such  that 

(5.1)  \uj  —  Uj\  <  cu    for  each  /,  j, 

(5.2)  U,  -  Uj=  Cy      ifxy  >   0. 

We  show  that  this  theorem  is  not  valid  as  it  stands.  In  fact,  let  us  recall  Example  2  and 
let  Xjj  be  the  optimal  solution  obtained  there.  If  there  exist  numbers  w,  which  satisfy  (5.1)  and 
(5.2),  then  we  must  have 

"i  -  "2  -  en  =  20, 
u3-  u2=  c32=  1, 

"1  -  "3=  Cn=   1. 

This  is  impossible. 

In  order  to  give  another  proof  of  Kantorovich's  Theorem,  Kretschmer  considered  Prob- 
lem (II)  and  asserted  N  =  M without  any  assumption.   Notice  that  N  <  Mm  Example  2. 


EQUALITIES  IN  TRANSPORTATION  PROBLEMS  595 

Kantorovich's  Theorem  was  amended  by  Kantorovich  and  Rubinstein  [5;  Theorem  3]  in 
the  following  form: 

THEOREM  4:  Assume  that  conditions  (C.l),  (C.2)  and  (C.3)  hold.  Then  a  feasible 
solution  Xy  of  Problem  (III)  is  an  optimal  solution  if  and  only  if  there  exist  numbers  «,  which 
satisfy  (5.1)  and  (5.2). 

Under  conditions  (C.l)  and  (C.2),  the  dual  problems  of  Problems  (II)  and  (III)  coincide 
and  Theorem  4  is  an  immediate  consequence  of  the  well-known  duality  theorem  applied  to 
Problem  (II).  Thus,  condition  (C.3)  can  be  omitted  in  Theorem  4. 

Notice  that  conditions  (C.l),  (C.2)  and  (C.3)  hold  if  and  only  if  the  cost  c^  is  a  pseudo- 
metric,  i.e.,  Cjj  satisfies  conditions  (C.l)  and  (C.2)  and  the  following  condition 

(C.4)  cu  <  cik  +  ckj    for  all  /,  j,  k. 

With  the  aid  of  Theorems  2  and  4,  we  have 

THEOREM  5:  Assume  that  conditions  (C.l),  (C.2)  and  (C.3)  hold.  Then  a  feasible 
solution  Xy  of  Problem  (I)  is  an  optimal  solution  if  and  only  if  there  exist  numbers  w,  which 
satisfy  (5.1)  and  (5.2). 

ACKNOWLEDGMENT 

We  are  indebted  to  a  referee  for  helpful  comments,  weakening  the  original  assumptions  of 
Theorem  2,  in  particular. 

REFERENCES 

[1]  Charnes,  A.  and  W.W.  Cooper,  Management  Models  and  Industrial  Applications  of  Linear  Pro- 
gramming, /and  //,  (J.  Wiley  and  Sons,  New  York,  N.Y.,  1961). 

[2]  Dantzig,  G.B.,  Linear  Programming  and  Extensions,  (Princeton  University  Press,  Princeton, 
1963). 

[3]  Gomory,  R.E.  and  T.C.  Hu,  "An  Application  of  Generalized  Linear  Programming  to  Net- 
work Flows,"  Journal  of  the  Society  for  Industrial  and  Applied  Mathematics,  10,  260-283 
(1962). 

[4]  Kantorovich,  L.V.,  "On  the  Translocation  of  Masses,"  Management  Science,  5,  1-4  (1958). 
(English  translation  of  Doklady  Akademii  Nauk  USSR,  37,  199-201  (1942). 

[5]  Kantorovich,  L.V.  and  G.  Sh.  Rubinstein,  "On  a  Space  of  Completely  Additive  Functions," 
Vestnik  Leningrad  University,  13,  52-59  (1958)  (Russian). 

[6]  Kretschmer,  K.S.,  "Programmes  in  Paired  Spaces,"  Canadian  Journal  of  Mathematics,  13, 
221-238  (1961). 

[7]  Ryan,  M.J.,  "More  on  the  More  for  Less  Paradox  in  the  Distribution  Model,"  in  Extremal 
Methods  and  Systems  Analysis,  An  International  Symposium  on  the  Occasion  of  Professor  Abra- 
ham Charnes'  Sixtieth  Birthday,  A.V.  Fiacco,  K.O.  Kortanek  (Editors),  275-303,  Volume 
174  of  Lecture  Notes  in  Economics  and  Mathematical  Systems,  Managing  Editors:  M. 
Beckmann  and  H.P.  Kiinzi,  Springer- Verlag,  Berlin-Heidelberg-New  York,  1980. 


A  NETWORK  FLOW  APPROACH  FOR  CAPACITY 
EXPANSION  PROBLEMS  WITH  TWO  FACILITY  TYPES 


Bell  Laboratories 
Holmdel,  New  Jersey 


ABSTRACT 

A  deterministic  capacity  expansion  model  for  two  facility  types  with  a  finite 
number  of  discrete  time  periods  is  described.  The  model  generalizes  previous 
work  by  allowing  for  capacity  disposals,  in  addition  to  capacity  expansions  and 
conversions  from  one  facility  type  to  the  other.  Furthermore,  shortages  of 
capacity  are  allowed  and  upper  bounds  on  both  shortages  and  idle  capacities  can 
be  imposed.  The  demand  increments  for  additional  capacity  of  any  type  in  any 
time  period  can  be  negative.  All  cost  functions  are  assumed  to  be  piecewise, 
concave  and  nondecreasing  away  from  zero.  The  model  is  formulated  as  a 
shortest  path  problem  for  an  acyclic  network,  and  an  efficient  search  procedure 
is  developed  to  determine  the  costs  associated  with  the  links  of  this  network. 


INTRODUCTION 

In  a  previous  paper  [9],  we  described  a  deterministic  capacity  expansion  model  for  two 
facility  types.  The  model  has  a  finite  number  of  discrete  time  periods  with  known  demands  for 
each  of  the  two  facilities  in  any  period.  At  the  beginning  of  each  period,  facility  /'(/'  =1,2) 
may  be  expanded  either  by  new  construction  or  by  converting  idle  capacity  of  one  facility  to 
accommodate  the  demand  for  the  other  facility. 

In  this  paper,  we  extend  our  previous  work  by  allowing  for  the  reduction  of  facility  size 
through  capacity  disposals.  Furthermore,  shortages  of  capacity  are  allowed  and  upper  bounds 
on  idle  capacities  and  shortages  may  be  imposed.  These  generalizations  allow  us  to  deal  with 
more  realistic  situations.  Capacity  disposals  are  often  initiated  due  to  high  holding  cost  of  idle 
capacity  when  the  cumulative  demand  decreases  over  some  successive  periods.  Capacity  shor- 
tages may  be  attractive  when  capacity  may  be  temporarily  rented  or  imported  from  other 
sources.  Also,  in  some  applications  it  may  be  economical  to  permit  temporary  shortages  and 
pay  a  penalty  for  unsatisfied  demand,  rather  than  expanding  the  facilities  at  that  time.  Finally, 
upper  bounds  on  idle  capacity  and  shortages  are  usually  imposed  by  management. 

The  costs  incurred  include  those  for  construction  of  new  capacity,  disposal  of  existing 
capacity,  conversion,  holding  of  idle  capacity,  and  for  having  capacity  shortages.  As  in  [9], 
conversion  implies  physical  modification  so  that  the  converted  capacity  becomes  an  integral  part 
of  the  new  facility  and  is  not  reconverted  automatically  at  the  end  of  the  period.  The  capacity 
expansion  policy  consists  of  timing  and  sizing  decisions  for  new  constructions,  disposals,  and 
conversions  so  that  the  total  costs  are  minimized. 


597 


598  h. luss 

The  model  is  useful  for  communication  network  applications,  such  as  the  cable  sizing 
problems  examined  in  [9].  Suppose  the  demands  for  two  cable  types  is  known  for  the  next  T 
periods.  Furthermore,  suppose  the  more  expensive  cable  can  accommodate  both  demand 
types,  whereas  the  cheaper  cable  can  be  used  only  to  satisfy  its  associated  demand.  Since  the 
construction  cost  functions  are  often  concave,  reflecting  economies  of  scale,  it  can  become 
attractive  to  use  the  more  expensive  cable  for  future  demand  for  both  cables.  Thus,  careful 
planning  of  the  expansion  policy  is  needed.  A  similar  application  is  the  planning  of  capacity 
expansion  associated  with  communication  facilities  which  serve  digital  and  analog  demands. 
Other  areas  of  applications  include  production  problems  for  two  substitutable  products,  and 
inventory  problems  of  a  single  product  produced  and  consumed  in  two  separate  regions;  see  [9] 
for  more  details. 

Many  capacity  expansion  models  and  closely  related  inventory  models  have  been 
developed  for  the  single  facility  problem  with  a  finite  number  of  discrete  time  periods.  The 
first  such  model  was  proposed  by  Wagner  and  Whitin  [13]  who  examined  a  dynamic  version  of 
the  economic  lot  size  model.  Many  authors  extended  this  model;  for  example,  Manne  and 
Veinott  [11],  Zangwill  [16]  and  Love  [8].  Zangwill  used  a  network  flow  approach,  and  Love 
generalized  the  model  to  piecewise  concave  cost  functions  and  bounded  idle  capacities  and 
shortages. 

Several  models  and  algorithms  for  two  facility  problems  have  been  developed.  Manne 
[10],  Erlenkotter  [1,2],  Kalotay  [5],  and  Fong  and  Rao  [3]  examined  models  in  which  it  is 
assumed  that  converted  capacity  is  reconverted  automatically,  at  no  cost,  at  the  end  of  each 
period.  Kalotay  [6],  Wilson  and  Kalotay  [14],  Merhaut  [12],  and  Luss  [9]  examined  models  in 
which  converted  capacity  is  not  reconverted  automatically  at  the  end  of  each  period. 

In  Section  1  we  describe  the  generalized  model.  The  algorithm  in  [9]  is  extended  and 
used  to  solve  the  new  model  with  the  additional  features  described  before.  In  Section  2  a  shor- 
test path  formulation  is  presented,  and  in  Section  3  some  properties  of  an  optimal  solution  are 
identified.  These  properties  are  used  to  compute  the  costs  associated  with  the  links  of  the  net- 
work constructed  for  the  shortest  path  problem.  In  Section  4  the  solution  is  illustrated  by  a 
numerical  example,  and  some  final  comments  are  given  in  Section  5. 

1.   THE  MODEL 

The  model  assumes  a  finite  number  of  discrete  time  periods  in  which  the  demand  incre- 
ments, new  constructions,  capacity  disposals,  and  capacity  conversions  occur  instantaneously 
and  simultaneously  immediately  after  the  beginning  of  each  period.  We  define  the  following 
notation: 

;  —  index  for  the  two  facilities. 

/  —  index  for  time  periods  (r  =  1,2,  . . ,  ,  T)  where  Tis  the  planning  horizon. 

r„  —    the   increment   of  demand   for   additional   capacity   of  facility    i  incurred 

immediately  after  the  beginning  of  period  t.    The  /-,-,' s  may  be  negative,  and 
for  convenience  are  assumed  to  be  integers. 

Ri(tut2)   «  £  r„,  for  r,  <  t2. 


—  the  amount  of  new  construction  (x„  >  0),  or  capacity  disposal  (x„  <  0), 
associated  with  facility  i  immediately  after  the  beginning  of  period  t. 


CAPACITY  EXPANSION  WITH  TWO  FACILITY  TYPES  599 

y,  —  the  amount  of  capacity  converted  immediately  after  the  beginning  of  period 

t.  y,  >  0  (y,  <  0)  implies  that  capacity  associated  with  facility  1  (facility  2)  is 
converted  to  satisfy  the  demand  of  the  other  facility.  Once  converted,  the 
capacity  becomes  an  integral  part  of  the  new  facility. 

/,,  —  the  amount  of  idle  capacity  (/,-,  >  0),  or  capacity  shortage  (/,,  <  0),  associ- 

ated with  facility  /'  at  the  beginning  of  period  t  (or  equivalently,  at  the  end  of 
period  t  —  1,  /  =  2,3,  ...  ,  T  +  1).  We  assume  that  initially  there  is  no  idle 
capacity  or  capacity  shortage,  that  is,  Iti  =  0. 

hi  —  lower  bound  on  /,,,  that  is,  the  maximum  capacity  shortage  of  facility  i 

allowed  at  the  beginning  of  period  t ;  the  /,,'s  are  assumed  to  be  integers  and 

-   OO    <    /,.,    <    0. 

wH  —  upper  bound  on  the  idle  capacity  of  facility  i  at  the  beginning  of  period  /. 

The  w,,'s  are  assumed  to  be  integers  and  0  <  wtt  <  oo. 

Ci,(xit)      —  the  construction  and  disposal  cost  function  for  facility  /'at  time  period  t. 

Si(y,)       —  the  conversion  cost  function  at  time  period  t. 

hit(Ii,t+\)—  the  cost  function  associated  with  idle  capacity,  or  capacity  shortage,  of  facil- 
ity /'carried  from  period  t  to  period  t  +  1. 

All  cost  functions  are  assumed  to  be  concave  from  0  to  °o  and  from  0  to  —  oo,  but  not 
necessarily  concave  over  the  entire  interval  [-oo,  ©o].  Such  functions  are  called  piecewise  con- 
cave functions,  see  Zangwill  [15].  All  cost  functions  are  also  assumed  to  be  nondecreasing 
away  from  zero;  for  example,  c„(x„)  is  nondecreasing  with  xit  for  xit  >  0,  and  nondecreasing 
with  —  xit  for  xjt  <  0.   For  convenience,  we  assume  that  c,,(0)  =  £,(0)  =  //,,(0)  =  0. 

The  problem  can  be  formulated  as  follows: 

(1.1)  Minimize    £    £  cit  0c„)  +  hit  (/,  ,+1)  I  +  g,  (y,) 

(1.2)  /u+1  -  /„  +  x„  -  yt  -  n,      < 

(1.3)  /2,,+1  -  I2l  +  x2l  +  y,  -  rlt 
(1) 

(1.4)  Iu  <  In  <  "a 

(1.5)  7(1  =  0 

(1.6)  7,,7-+1  =  0 

The  objective  (1.1)  is  to  minimize  the  total  cost  incurred  over  all  periods.  Equations 
(1.2)  -  (1.3)  express  the  idle  capacity  or  capacity  shortage  7,,+i  as  a  function  of  /,,,  the  actions 
undertaken  at  period  t,  xit  and  v,,  and  the  demand  increments  r„.  Constraints  (1.4)  specify  the 
bounds  on  idle  capacities  and  capacity  shortages,  and  Equation  (1.5)  is  introduced  by  assump- 
tion. Constraint  (1.6)  7,  r+1  =  0  implies  that  idle  capacity  or  capacity  shortages  are  not  allowed 
after  period  T.   Such  a  constraint  is  not  restrictive  since  one  can  add  to  problem  (1)  a  fictitious 


/=  1,2,  ...  ,   T 
i-l,2 


600  H.  LUSS 

period  T' =  T  +  I  with  riT  =  max  R,(X,t)  -  R,(l,T)  (yielding  R,(l,T')  >  U/(U)  Vr), 
/,r  =  0,w,r  =  °°,  and  c,r(-)  =  hjr  (•)  =  ?7-(0  =  0.  (/,r  is  fixed  at  zero  since  no  shortages  are 
allowed  at  the  end  of  period  T).  This  allows  us  to  fix  /,r+i  at  zero  since  then  there  always 
exists  an  optimal  solution  with  /,,r+i  =  0.  To  simplify  notation,  we  assume  that  period  Tin 
formulation  (1)  is  the  added  fictitious  period. 

The  constraints  (1.2)  -  (1.6)  form  a  nonempty  convex  set.  Since  each  term  of  the  objec- 
tive function  is  nondecreasing  away  from  zero  with  a  finite  value  at  zero,  there  exists  a  finite 
optimal  solution.  Furthermore,  suppose  each  of  the  variables  x„,  v,,  and  /„  is  replaced  in  for- 
mulation (1)  by  the  difference  of  two  nonnegative  variables,  for  example,  x„  =  xl,  —  x  ■',,  where 
x'u  ^  0  represents  constructions  and  x"r  ^  0  stands  for  disposals.  In  that  case,  the  objective 
function  becomes  concave  on  the  entire  feasible  region;  hence,  there  exists  an  extreme  point 
optimal  solution.  From  Pages  124-127  in  Hu  [4],  the  constraints  (1.2)  -  (1.3)  are  totally  uni- 
modular.  Thus,  since  rit,  /„  and  wu  are  assumed  to  be  integers,  such  an  extreme  point  solution 
consists  of  integers.  In  the  next  sections  we  describe  an  algorithm  which  finds  an  optimal 
extreme  point  solution. 

2.   A  SHORTEST  PATH  FORMULATION 

Since  all  cost  functions  are  nondecreasing  away  from  zero,  it  can  be  shown  that  there 
exists  an  optimal  solution  in  which 

(2)  |/„|  <  max[/?,(T1,r)  +  R2(t2,T)]  =  b  Vi,  t. 

T,,T2 

However,  usually,  better  bounds  than  those  given  by  (2)  can  be  assigned.  To  simplify  the 
presentation,  we  assume  that  the  lower  and  upper  bounds  on  the  /„  variables  satisfy  wit  ^  b 
and  /„  ^  -b  for  all  values  of  /  and  t. 

Generalizing  the  concept  of  capacity  point  in  [9],  we  define  a  capacity  point  as  a  period  /  in 
which  /„  =  0,  or  /„,  or  w„  for  at  least  one  value  of  i.  Since  an  extreme  point  optimal  solution 
consists  of  integers,  the  set  of  capacity  points  is  defined  as  follows: 

(3.1)  /„  =  /21  =  0 

(3.2)  Iu  =  /if,0(wn  and  I2t  =  ht>0,w2t 
(3) 

(3.3)  /„  =  lu,  0,wh  and  I2l  =  l2t  +  1,  . . .  ,  -1, 1,  . . .  ,  w2,,. 

(3.4)  I2t  =  l2„0,w2l  and/1;  =  lu  +  1,  ...  ,  -1,1,  ...  ,  wu. 

t  =  2,3,  ...  ,  T 

(3.5)  /,  r+1  = /2  r+1  =  0. 

The  capacity  point  values  can  be  conveniently  specified  by  a  single  parameter  a,.  For 
example,  a,  =  1,2,  ....  9  can  be  used  to  specify  the  combinations  given  by  (3.2),  etc.  A 
complete  example  of  a  special  case  can  be  found  in  [9]. 

The  set  of  capacity  points  can  be  limited  to  those  satisfying 

(4)    (4.1)    Iu  +  I2,<  RiU,T)+  R2U,T) 

(4.2)    Iu  +  I2t>    -max    [/?,(t,,/-  1)  +  R2(t2,i-  1)]. 


CAPACITY  EXPANSION  WITH  TWO  FACILITY  TYPES  601 

Equation  (4.1)  states  that  the  total  idle  capacity  at  the  beginning  of  period  /does  not  exceed  the 
cumulative  demand  from  period  t  to  T.  Equation  (4.2)  restricts  the  maximum  capacity  shor- 
tages to  the  maximum  demand  increase  from  any  period  prior  to  /  -  1  up  to  period  t  —  1. 
Clearly,  there  exists  an  optimal  solution  which  satisfies  (4). 

We  now  describe  a  shortest  path  formulation  which  can  be  used  to  solve  Problem  (1). 
Let 

duv(au,av+l)  —  the  minimal  cost  during  periods  u,  u  +  \,  . . .  ,  \  associated  with  an 
extreme  point  solution  of  (1)  when  u  and  v  +  1  are  two  successive  capacity 
points  with  values  defined  by  au  and  av+1.   More  specifically: 


lii  c,,(x,,)  +  hituu+A 


(5)  duw{au,av+\)  =  minimum  ]£   £  cit(xu)  +  //„(/,, +1)    -I-  g,(y,)\ 

such  that 

(i)      Constraints  (1.2)  and  (1.3)  are  satisfied  for  t  =  u,  u  +  1,  . . .  ,  v, 
(ii)     /„  <  /„  <  wu  and  /„  ^  0  for  i  =  1,  2  and  t  =  u  +  1,  u  +  2,  ...  ,  v, 
(iii)    I\u  and  I2u  are  defined  by  aw,  and  7i,v+i  and  /2,v+i  are  defined  by  av+i> 
(iv)    xit  and  y,  for  t  =  u,  u  +  1,  . . .  ,  v  satisfy  the  necessary  conditions  (to  be  developed 
later)  for  an  extreme  point  solution  of  (1). 

Suppose  that  all  subproblem  values  duv(au,av+\)  are  known.  The  optimal  solution  can  then 
be  found  by  searching  for  the  optimal  sequence  of  capacity  points  and  their  associated  values. 
As  shown  in  Figure  1,  Problem  1  can  be  formulated  as  a  shortest  path  problem  for  an  acyclic 
network  in  which  the  nodes  represent  all  possible  values  of  capacity  points.  Each  node  is 
described  by  two  values  it.a,)  where  t  is  the  time  period  and  a,  is  the  associated  capacity  point 
value.  From  each  node  (u,au)  there  emanates  a  directed  link  to  any  node  (v  +  l,av+i)  for 
v  ^  u  with  an  associated  cost  of  duy/(au,av+i). 

Let  C,  be  the  number  of  capacity  point  values  at  period  /.  Clearly,  C\  =  CT+\  =  1,  and  C, 
for  all  other  periods  can  be  obtained  from  Equations  (3)  and  (4).  The  total  number  of  links  N 
in  the  shortest  path  problem  is 

T  I  T+l         I 

(6)  N=±  CA  £   Cj\. 

;=1  [j-i+l        } 

Since  most  of  the  computational  effort  is  spent  on  computing  the  duv(au,av+\)  values,  it  is 
important  to  reduce  N,  if  possible.  One  way,  of  course,  is  to  reduce  the  values  of  C,  through 
the  imposition  of  appropriate  bounds  /„  and  w„ . 

The  shortest  path  problem  can  be  solved  using  various  algorithms.  Since  the  network  is 
acyclic  a  simple  dynamic  programming  formulation  can  be  used.  Let  a,  be  described  by  the  set 
of  integers  1,2,  ....  C,,  where  a,  =  1  represents  I\,  =  I2t  =  0.  Furthermore,  let  f,(a,)  be  the 
cost  of  an  optimal  policy  over  periods  /,  t  +  l,  ...  ,  T,  given  that  period  t  is  a  capacity  point, 
and  that  /t,  and  I2,  are  specified  by  a,.  The  following  dynamic  programming  formulation  is 
then  obtained: 

/r+i(«r+i)  =  0,  ar+1  =  1 

(7)  fu(au)  =        min        [^uv(au,av+1)  +  /v+1(av+1)], 

K«V+)^"CV+1 

u=  T,T-\,  ...  ,  \ 
au=  1,2 Cu. 


Figure  1.  The  shortest  path  foi 


The  first  term  of  the  minimand  is  the  minimum  cost  of  the  optimal  policy  during  periods  «, 
u  +  1,  ....  v,  given  that  u  and  v  +  1  are  two  successive  capacity  points  with  values  au  and 
av+1.   The  second  term  is  the  optimal  cost  for  periods  v  +  1,  v  +  2,  ...,   T,  given  av+1. 

3.   SOLUTION  OF  THE  SUBPROBLEMS  duMu.<*r+0 

Most  of  the  computational  effort  is  spent  on  computing  the  subproblem  values.  As  shown 
in  [9],  when  r„  ^  0,  x„  ^  0,  /,,  =  0  and  wit  =  oo  for  all  /and  /,  the  subproblems  are  solved  in  a 
trivial  manner,  however,  when  the  r„'s  are  allowed  to  be  negative  the  effort  required  to  solve 
the  subproblems  increases  significantly.  The  additional  modifications  needed  to  solve  the  sub- 
problems  duv(au,ay+i),  as  defined  by  (5)  for  the  generalized  model,  require  a  more  careful 
analysis  than  needed  in  [9],  however,  the  resulting  computational  effort  appears  to  be  about  the 


To  compute  the  subproblem  values  duv(au,av+l),  it  is  convenient  to  describe  Problem  (1) 
as  a  single  commodity  network  problem.  The  network,  shown  in  Figure  2,  includes  a  single 
source  (node  0)  with  a  supply  of  /?,(1,D  +  R2(l,T).  There  are  2T  additional  nodes,  each 
denoted  by  (/,/)  where  /specifies  the  facility  and  /specifies  the  time  period.  At  each  node  (/,/) 
there  is  an  external  demand  increment  /•„,  possibly  negative.  The  nodes  are  connected  by  links, 
where  the  flows  along  these  links  represent  the  constructions,  disposals,  conversions,  idle  capa- 
cities, and  capacity  shortages.  The  flows  on  each  link  can  be  in  either  direction,  and  the  link 
direction  in  Figure  2  indicates  positive  flows.   The  nodes  are  connected  by  the  following  links: 

—       A  link  from  node  0  to  each  node  (/,/)  with  flow  x„.   xit  is  positive  if  the  flow  is  from 
node  0  to  node  (/,/),  and  negative  otherwise. 


CAPACITY  EXPANSION  WITH  TWO  FACILITY  TYPES 


603 


Rgd.T) 


Figure  2.  A  network  flow  representation  of  the  capacity  expansion  problem 

—  A  link  from  each  node  (i,t)  to  node  (i,t  +  1)  with  flow  Iu+\.    Iu+\  is  positive  if  the 
flow  is  from  (i,t)  to  (i,t  +  1)  and  negative  otherwise. 

—  A  link  from  each  node  (1,/)  to  node  (2,t)  with  flow  y,.   y,  is  positive  if  the  flow  is 
from  node  (l,f)  to  (2,/),  and  negative  otherwise. 

As  discussed  before,  we  are  interested  in  finding  an  optimal  extreme  point  solution  to  a 
modified  version  of  Problem  (1),  in  which  each  of  the  variables  xit,  yt,  /„  is  replaced  by  the 
difference  of  two  nonnegative  variables.  It  can  be  shown  that  a  feasible  flow  in  the  network 
given  in  Figure  2  corresponds  to  an  extreme  point  solution  of  Problem  (1)  modified  as 
described  above,  if  and  only  if  it  does  not  contain  any  loop  with  nonzero  flows  in  which  all  /„ 
flows  satisfy  /„  <  Iit  <  wit  and  Iit  ^  0. 

Concentrating  upon  a  single  subproblem,  as  shown  in  Figure  3,  one  may  observe  that  a 
feasible  flow  does  not  contain  such  loops  if  and  only  if  the  following  properties  are  satisfied: 


(8.1)  xitixil2=0Ui*  t2), 
(8)    (8.2)  V,2=0(^r2), 
(8.3)  xlrix2,2^3=0. 


i  =  l,2 

u  <  fi,  t2,  h  ^  v 


For  example,  suppose  (8.3)  is  violated  and  tx  <  t2  <  t3,  then  xlri,  /i,,1+i,  . . .  ,  /i,3,  ytj,  /2/3, 
h,t3-\>  ■••  •  h,t2+\>  xit  form  a  loop  with  nonzero  flows  and  all  relevant  /„  values  satisfy 
/„  <  Iit  <  wu  and  4  J*  0. 

Equation  (8)  implies  that  in  the  optimal  solution  of  duv(au,aw+i)  there  is  at  most  one  new 
construction  or  disposal  for  each  facility  (8.1),  and  at  most  one  conversion  (8.2).  Furthermore, 
if  two  constructions  or  disposals  (one  per  facility)  are  being  considered,  conversion  is  then  not 
allowed  (8.3). 


Figure  3.  A  network  flow  representation  of  a  subproblem 

Let  Dj  be  the  capacity  change  of  facility  /'  during  periods  u,u  +  1,  . ..  ,  v,  that  is: 

(9)  A  =  /,,  v+1  +  /?, (u,  v)  -  /,„,  i  -  1,2 
or,  equivalents 

(10)  A  =  ±  xu  -  y, 


D2  =  ±x2t+yt. 

t—u 

Let  t\  and  t2  be  two  time  periods  u  <  U\,t2)  <  v.  From  the  optimal  properties  (8)  shown 
above,  the  possible  policies  associated  with  an  optimal  solution  to  any  subproblem  duv(au,av+{) 
can  be  restricted  to  three  different  policies.  These  policies  are  summarized  in  Table  1  below. 

To  illustrate  the  table,  let  us  concentrate  on  the  column  A  ^  0  and  D2  ^  0.  Policy  (a) 
indicates  a  single  disposal  of  A  capacity  units  of  facility  1,  and  a  single  construction  of  D2  units 
of  facility  2.  Policy  (b)  implies  a  single  construction  of  A  +  D2  of  facility  1  if  Z>!  +  Z)2  ^  0,  a 
single  disposal  of  A  +  -A  of  facility  1  if  Dx  +  D2  ^  0,  and  a  single  conversion  of  D2  units 
from  facility  1  to  facility  2.  Obviously,  if  D^  +  D2=  0,  no  constructions  or  disposals  take 
place,  and  if  D2  =  0,  no  capacity  is  converted.  Finally,  policy  (c)  consists  of  a  single  construc- 
tion of  A  +  Di  capacity  units  of  facility  2  if  Z>i  +  Z)2  >  0,  a  single  disposal  of  A  +  A  units 
of  facility  2  if  A  +  &i  ^  0>  and  a  single  conversion  of  -A  from  facility  1  to  facility  2. 

The  optimal  solution  of  a  subproblem  duy{auiay+x)  is  therefore  obtained  by  the  following 
procedure: 

(1)  For  each  of  the  policies  (a),  (b),  and  (c)  in  Table  1,  find  the  optimal  values  of  t\  and 
t2y  which  minimize  duv(au,a^.\)  as  given  by  Equation  (5),  while  satisfying  condi- 
tions (i)  -  (iv)  given  below  Equation  (5).  If  no  feasible  values  of  t^  and  t2  exist,  set 
the  value  of  the  corresponding  policy  to  «>. 


CAPACITY  EXPANSION  WITH  TWO  FACILITY  TYPES 

TABLE  1.    Possible  Policies  for  Optimal  Subproblem  Solutions 


605 


— ^_01.02 

D,  ^  0 

Z>,  <  o 

Z>,  >  0 

Z>,  <  0 

Policy 

d2  ^  o 

Z)2  >  0 

Z)2  <  0 

£>2  <  0 

x,M  =  Du  xu  -  0  t  7*  f, 

construction 

disposal 

construction 

disposal 

(a) 

*2,2  =  D2,  x2l  =  0  t  *  t2 
y,  =  0  Vr 

construction 

construction 

disposal 

disposal 

xUj  =  Z),  +  D2,  xu  -  0  f  ?*  fj 

construction 

construction 
or  disposal 

construction 
or  disposal 

disposal 

(b) 

y,;  =o2,j,  =  0/^/2 

conversion 

conversion 

conversion 

conversion 

from  1  to  2 

from  1  to  2 

from  2  to  1 

from  2  to  1 

x2,  =  0V/ 

*2,,   -  X>1  +   ^2-    X2,  =   0  ^    r. 

construction 

construction 
or  disposal 

construction 
or  disposal 

disposal 

(c) 

jte-4/)!,  y,  =  o^/2 

conversion 

conversion 

conversion 

conversion 

from  2  to  1 

from  1  to  2 

from  2  to  1 

from  1  to  2 

x„  =  OVf 

(2)     Choose  as  the  optimal  policy  the  best  of  those  found  in  Step  (1).    If  none  of  the  policies 
is  feasible,  duv(au,av+])  =  «>. 

The  procedure  above  may  involve  spending  a  significant  amount  of  computation  on 
finding  all  feasible  policies  and  comparing  the  costs  associated  with  these  policies. 

4.    A  NUMERICAL  EXAMPLE 

As  an  illustration,  we  solve  the  capacity  expansion  problem  shown  in  Figure  4. 
^14  =  ^24  =  0  by  assumption,  thus,  a  fictitious  period  is  not  added.  The  cost  functions  are  given 
in  Table  2  below. 

The  shortest  path  formulation  is  shown  in  Figure  5.  The  capacity  point  values  are  given 
inside  the  nodes  in  terms  of  lu  and  I2,  rather  than  a,.  Using  Equation  (4.1),  several  capacity 
point  values  are  omitted  in  periods  2  and  3.  Furthermore,  all  links  from  period  t  =  1  to 
periods  t  =  3  and  4  are  omitted  since  there  is  no  feasible  solution  to  the  associated  subprob- 
lems  with  lu  <  hi  <  wu  and  hi  ^  0.  The  number  associated  with  each  link  is  the  optimal 
solution  of  the  corresponding  subproblem.   The  shortest  path  is  marked  by  stars. 

Consider  the  subproblem  dn{a\,  a2)  where  a\  represents  the  capacity  point  value 
hi  =  hi  =  0,  and  a2  represents  I2\  =  hi  =  0-  By  Equation  (9),  Dx  =  1  and  D2  =  1.  Using 
the  results  of  Table  1,  policy  (a)  yields  xn  =  x2i  =  1  with  a  total  cost  of  68,  policy  (b)  yields 
xn  =  2  and  y\=  I  with  a  cost  of  46,  and  policy  (c)  yields  x2i  =  2  and  y\  =  —  1  with  a  cost  of 
45.  Hence  policy  (c)  is  the  optimal  one. 

To  illustrate  further,  consider  d23(a2,a4),  where  a2  stands  for  712  =  -1  and  722  =  0,  and 
a4  stands  for  714  =  724  =  0,  so  that  D^  =  1  and  D2  =  0.  From  Table  1,  policy  (a)  implies  that 
either  xu  =  1  or  xJ3  =  1.  However,  if  x13  =  1  (and  x12  =  0)  then  713  =  0  so  that  d23(-)  =  «>. 
Hence,  policy  (a)  implies  x12  =  1  with  construction  and  holding  cost  of  43.2.  Policy  (b)  yields 
the  same  solution  as  policy  (a),  and  policy  (c)  results  in  x22  =  1  and  y2  =  -I  with  a  total  cost 
of  40.5;  hence,  policy  (c)  is  optimal  for  that  subproblem. 


606 


Figure  4.  A  network  flow  representation  of  the  example 
TABLE  2  -  The  Cost  Functions 


^\function 
argument^\ 

c\,(xh) 

c2l(x2t) 

MW 

/=  1,2 

g,(yt) 

positive 

zero 

negative 

(30  +  8  xu)0.9'-1 

0 

6  •  0.9'"1 

(20  +  10x2,)0.9'"1 

0 

5  •  0.9'" ' 

5Iu+l0.9'-1 

0 

-l0Iiit+l0.9'-{ 

0 

0 

-5^,0.9'-' 

CAPACITY  EXPANSION  WITH  TWO  FACILITY  TYPES 


607 


Figure  5.  The  shortest  path  problem  for  the  example 

Finally,  consider  tf23(a2,a4)  with  a2  standing  for  712  =  /22  =  0,  and  a4  standing  for 
A  4  =  ^24  =  0-  From  Table  1,  since  Dx  =  D2  =  0,  all  decision  variables  are  zero  in  all  three  pol- 
icies and  the  total  costs  incurred  are  equal  to  9. 

After  solving  the  subproblems  for  all  the  links  of  Figure  5,  the  shortest  path  can  be  found 
using  the  dynamic  programming  formulation  (7)  or  any  other  shortest  path  algorithm.  The 
shortest  path  in  this  example  is  54  and  consists  of  two  links.  The  first  link  connects  node 
hi  =  hi  =  0  to  node  hi  =  hi=  0>  and  the  second  link  connects  node  l\i  =  /22  =  0  to  node 
A4  =  -^24=  0-  The  optimal  policy  of  the  entire  problem  is  x2i  —  2,  y\  =*  — 1,  with  all  other 
decision  variables  xit  and  y,  being  equal  to  zero. 

5.   FINAL  COMMENTS 

This  paper  generalizes  our  previous  work  [9]  by  allowing  for  capacity  disposals  and  capa- 
city shortages.  Furthermore,  bounds  on  idle  capacities  and  capacity  shortages  can  be  imposed. 
The  model  is  formulated  as  a  shortest  path  problem  in  which  most  of  the  computational  effort 
is  spent  on  computing  the  link  costs.  Using  a  network  flow  approach,  properties  of  extreme 
point  solutions  are  identified.  These  properties  are  used  to  develop  an  efficient  search  for  the 
link  costs. 


Further  generalizations  may  include  bounds  on  new  constructions  and  capacity  disposals, 
and  operating  costs  which  depend  on  the  facility  type  and  time  period.  As  shown  by  several 
authors,  for  example  Lambrecht  and  Vander  Eecken  [7],  bounded  constructions  or  disposals 
complicate  considerably  even  the  single  facility  problem.  Introducing  operating  costs  may 
require  major  changes  in  the  algorithm  since  the  amount  of  each  capacity  type  used  to  satisfy 
the  demand  in  each  period  affects  the  total  cost. 


608  H   LUSS 

Finally,  negative  costs  for  disposals  (credit  for  salvage  value)  can  be  incorporated  for  cer- 
tain cost  functions  c„(x„)  for  which  the  optimal  solution  would  be  finite.  For  example,  cost 
functions  in  which  the  credit  per  unit  of  disposed  capacity  is  always  smaller  than  the  construc- 
tion cost  per  unit  of  capacity.  In  general,  however,  cost  functions  c„(x„)  that  are  negative  for 
xit  <  0  may  result  in  an  unbounded  solution. 

REFERENCES 

[1]  Erlenkotter,  D.,  "Two  Producing  Areas— Dynamic  Programming  Solutions,"  Investments  for 
Capacity  Expansion:  Size,  Location,  and  Time  Phasing,  210-227,  A.  S.  Manne,  Editor, 
(MIT  Press,  Cambridge,  Massachusetts,  1967). 

[2]  Erlenkotter,  D.,  "A  Dynamic  Programming  Approach  to  Capacity  Expansion  with  Speciali- 
zation," Management  Science,  21,  360-362  (1974). 

[3]  Fong,  CO.,  and  M.R.  Rao,  "Capacity  Expansion  with  Two  Producing  Regions  and  Con- 
cave Costs,"  Management  Science,  22,  331-339  (1975). 

[4]  Hu,  T.C.,  Integer  Programming  and  Network  Flows,  124-127,  (Addison  Wesley,  Reading, 
Massachusetts,  1969). 

[5]  Kalotay,  A.J.,  "Capacity  Expansion  and  Specialization,"  Management  Science,  20,  56-64 
(1973). 

[6]  Kalotay,  A.J.,  "Joint  Capacity  Expansion  without  Rearrangement,"  Operational  Research 
Quarterly,  26,  649-658  (1975). 

[7]  Lambrecht,  M.  and  J.  Vander  Eecken,  "Capacity  Constrained  Single  Facility  Lot  Size  Prob- 
lem," European  Journal  of  Operational  Research,  2,  132-136  (1978). 

[8]  Love,  S.F.,  "Bounded  Production  and  Inventory  Models  with  Piecewise  Concave  Costs," 
Management  Science,  20,  313-318  (1973). 

19]  Luss,  H.,  "A  Capacity-Expansion  Model  for  Two  Facilities,"  Naval  Research  Logistics 

Quarterly,  26,  291-303  (1979). 
[10]  Manne,  A.S.,  "Two  Producing  Areas— Constant  Cycle  Time  Policies,"  Investments  for  Capa- 
city Expansion:  Size,  Location,  and  Time  Phasing,  193-209,  A.S.  Manne,  Editor,  (MIT 
Press,  Cambridge,  Massachusetts,  1967). 
[11]  Manne,  A.S.  and  A.F.  Veinott,  Jr.,  "Optimal  Plant  Size  with  Arbitrary  Increasing  Time 
Paths  of  Demand,"  Investments  for  Capacity  Expansion:  Size,  Location,  and  Time  Phasing, 
178-190,  A.S.  Manne,  Editor,  (MIT  Press,  Cambridge  Massachusetts,  1967). 
[12]  Merhaut,  J.M.,  "A  Dynamic  Programming  Approach  to  Joint  Capacity  Expansion  without 
Rearrangement,"  M.  Sc.  Thesis,  Graduate  School  of  Management,  University  of  Cali- 
fornia, Los  Angeles,  California  (1975). 
[13]  Wagner,  H.M.  and  T.M.  Whitin,  "Dynamic  Version  of  the  Economic  Lot  Size  Model," 

Management  Science,  5,  89-96  (1958). 
[14]  Wilson,  L.O.,  and  A.J.  Kalotay,  "Alternating  Policies  for  Nonrearrangeable  Networks," 

INFOR,  14,  193-211  (1976). 
[15]  Zangwill,  W.I.,  "The  Piecewise  Concave  Function,"  Management  Science,   13,  900-912 

(1967). 
[16]  Zangwill,  W.I.,  "A  Backlogging  Model  and  a  Multiechelon  Model  for  a  Dynamic  Economic 
Lot  Size  Production  System— A  Network  Approach,"  Management  Science,  15,  506-527 
(1969). 


SOLVING  MULTIFACILITY  LOCATION  PROBLEMS 
INVOLVING  EUCLIDEAN  DISTANCES* 


Department  of  Systems  Design 

University  of  Waterloo 

Waterloo,  Ontario,  Canada 

Christakis  Charalambous 

Department  of  Electrical  Engineering 

Concordia  University 

Montreal,  Quebec,  Canada 

ABSTRACT 

This  paper  considers  the  problem  of  locating  multiple  new  facilities  in  order 
to  minimize  a  total  cost  function  consisting  of  the  sum  of  weighted  Euclidean 
distances  among  the  new  facilities  and  between  the  new  and  existing  facilities, 
the  locations  of  which  are  known.  A  new  procedure  is  derived  from  a  set  of 
results  pertaining  to  necessary  conditions  for  a  minimum  of  the  objective  func- 
tion. The  results  from  a  number  of  sample  problems  which  have  been  exe- 
cuted on  a  programmed  version  of  this  algorithm  are  used  to  illustrate  the 
effectiveness  of  the  new  technique. 


1.   BACKGROUND 

It  was  as  early  as  the  17th  century  that  mathematicians,  notably  Fermat,  were  concerned 
with  what  are  now  known  as  single  facility  location  problems.  However,  it  was  not  until  the 
20th  century  that  normative  approaches  to  solving  symbolic  models  of  these  and  related  prob- 
lems were  addressed  in  the  literature.  Each  of  these  solution  techniques  concerned  themselves 
with  determining  the  location  of  a  new  facility,  or  new  facilities,  with  respect  to  the  location  of 
existing  facilities  so  as  to  minimize  a  cost  function  based  on  a  weighted  interfacility  distance 
measure. 

If  one  studies  a  list  of  references  to  the  work  done  in  the  past  decade  involving  facility 
location  problems  it  becomes  readily  apparent  that  there  exists  a  strong  interdisciplinary  interest 
in  this  area  within  the  fields  of  operations  research,  management  science,  logistics,  economics, 
urban  planning  and  engineering.  As  a  result,  the  term  "facility"  has  taken  on  a  very  broad  con- 
notation in  order  to  suit  applications  in  each  of  these  areas.  Francis  and  Goldstein  [4]  provide  a 
fairly  recent  bibliography  of  the  facility  location  literature.  One  of  the  most  complete 
classifications  of  these  problems  is  provided  in  a  book  by  Francis  and  White  [5]. 


'This  work  was  supported  by  the  National  Research  Council  of  Canada  under  Grant  A4414  and  by  an  Ontario  Gradu- 
ate Scholarship  awarded  to  Paul  Calamai. 


609 


610  P.  CALAMAI  ANDC.  CHARALAMBOUS 

This  paper  concerns  itself  with  the  development  of  an  algorithm  for  solving  one  particular 
problem  in  the  area  of  facility  location  research.  The  problem  involves  multiple  new  facilities 
whose  locations,  the  decision  variables,  are  points  in  E2  space.  The  quantitative  objective  is  to 
minimize  the  total  cost  function  consisting  of  the  sum  of  weighted  Euclidean  distances  among 
new  facilities  and  between  new  and  existing  facilities.  The  weights  are  the  constants  of  propor- 
tionality relating  the  distance  travelled  to  the  costs  incurred.  It  is  assumed  that  the  problem  is 
"well  structured"  [3]. 

The  Euclidean  distance  problem  for  the  case  of  single  new  facilities  was  addressed  by 
Weiszfeld  [13],  Miehle  [10],  Kuhn  and  Kuenne  [8],  and  Cooper  [1]  to  name  a  few.  However, 
it  was  not  until  the  work  of  Kuhn  [7]  that  the  problem  was  considered  completely  solved.  A 
computational  procedure  for  minimizing  the  Euclidean  multifacility  problem  was  presented  by 
Vergin  and  Rogers  [12]  in  1967;  however,  their  techniques  sometimes  give  suboptimum  solu- 
tions. Two  years  later,  Love  [9]  gave  a  scheme  for  solving  this  problem  which  makes  use  of 
convex  programming  and  penalty  function  techniques.  One  advantage  to  this  approach  is  that  it 
considers  the  existence  of  various  types  of  spatial  constraints.  In  1973  Eyster,  White  and 
Wierwille  [2]  presented  the  hyperboloid  approximation  procedure  (HAP)  for  both  rectilinear 
and  Euclidean  distance  measures  which  extended  the  technique  employed  in  solving  the  single 
facility  problem  to  the  multifacility  case.  This  paper  presents  a  new  technique  for  solving  con- 
tinuous unconstrained  multifacility  location  problems  involving  Euclidean  distances. 

2.   PROBLEM  FORMULATION 

The  continuous  unconstrained  multifacility  location  problem  involving  the  lp  distance 
measure  can  be  stated  as  follows: 

Find  the  point  X*T  =  (X*{,  ...  ,  X*„r)in  E2n  to 
(PI)         minimize  f(X)  =      £      vJk  \  \Xj  -  Xk \  \p  +  £  £  wM  \  \Xj  -  At \  \p 

\Hj<kHn  7-1  /-l 

where 

n  A   number  of  new  facilities  (NF's) . 

m  A  number  of  existing  facilities  (EFs). 

X] '  =  (Xji  Xji)  A  vector  location  of  NFj  in  E2,  j  —  1,  •  •  •  ,  n. 

Aj —  (an  ai2)  A  vector  location  of  EFt  in  E2,  i  =  1, ,. . .  ,  m. 

\jk  A  nonnegative  constant  of  proportionality  relating  the  lp  distance  between  NFj  and  NFk 
to  the  cost  incurred  1  <  j  <  k  <  n. 

Wjj  A  nonnegative  constant  of  proportionality  relating  the  lp  distance  between  NFj  and  EFt  to 
the  cost  incurred  1  <  j  <  n,  1  <  /  <  m. 

\\Xj  -  Xk\\p  =  {\xn  -  xkl\P+  \xJ2  -  xk2Wlp  A  lp  distance  between  NF,  and  NFk. 

\\Xj  -  A,\\p  -  {\xji  -  an\"  +  \xJ2  -  ai2\pYlpb  lp  distance  between  NF,  and  EF,. 

Note  that  we  make  the  assumption  that  v^  =  vkJ  for  j,k  =  I,  ...  ,  n.  Substituting  p  =  1 
and  p  =  2  in  Problem  PI  respectively  yields  the  rectilinear  distance  problem  and  the  Euclidean 


SOLVING  MULTIFACILITY  LOCATION  PROBLEMS  611 

For  the  purpose  of  this  paper  Euclidean  distance  will  be  the  measure  used  between  facili- 
ties located  as  points  in  E2  space.  The  objective  function  becomes 

minimize  f(X)  =      £       vJk  {(xyl  -  xkl)2  +  (xj2  -  x^2)2)1/2 

X  \^j<k^n 

(P2)  +  £  £  wj,  {(*,,  -  an)2  +  (xJ2  -  a/2)2}1/2. 

7=1    i-1 

The  techniques  presented  in  this  paper  can  also  be  used  for  problems  involving  facilities  located 
in  three-dimensional  space. 

3.   NEW  FACILITY  CATEGORIZATION 

If  we  consider  a  current  solution  to  Problem  P2  we  can  think  of  each  new  facility  as  being 
in  one  of  the  following  distinct  categories: 

(1)  Unique  Point  (UP) 

A  new  facility  in  this  category  occupies  a  location  that  differs  from  all  other  facility 
locations. 

(2)  Coinciding  Point  (CP) 

A  new  facility  in  this  category  occupies  a  location  that  coincides  with  the  location  of 
an  existing  facility  but  differs  from  the  current  locations  of  all  other  new  facilities. 
Thus,  each  new  facility  in  this  category  has  associated  with  it  some  existing  facility 
which  has  the  same  vector  location. 

(3)  Unique  Clusters  ( UCX ,  ...,  UCNUC) 

All  new  facilities  in  the  /cth  unique  cluster  (k  =  1,  . . .  ,  NUC)  occupy  the  same  vec- 
tor location.  This  location  is  distinct  from  all  existing  facility  locations  as  well  as  the 
current  locations  of  new  facilities  that  are  not  classified  in  this  cluster. 

(4)  Coinciding  Clusters  (CC\,  ...  ,  CCNCc) 

All  new  facilities  categorized  in  the  /cth  coinciding  cluster  (k  =  1,  ...  ,  NCC)  occupy 
the  same  vector  location.  This  location  coincides  with  the  location  of  some  existing 
facility  and  differs  from  the  current  locations  of  all  new  facilities  that  are  not 
classified  in  this  cluster.  Each  of  these  coinciding  clusters  of  new  facilities  is  there- 
fore associated  with  some  existing  facility  with  which  it  shares  a  location. 

If  we  define  the  index  sets  J  A  {1,  . . .  ,  n)  and  /  A  {1,  . . .  ,  m)  and  let  the  subsets 
UC0  =  CC0  =  0  then  the  categorization  can  be  restated  as  follows: 

Partition  the  set  J  into  the  subsets  UP,  CP,  UC\ UCNUC,  CC\,  ...  ,  CCNCc  where 

(3.1)  UP  =  {V7.  €  j\Ai  *  Xj  *  Xk;  X/i  €  /,  Vk  €  J  -  {j}) 

(3.2)  CP  =  {\/j  €  j\Ar  =  Xj  *  Xk;  ij  e  I,  Vk  £  J  -  {j}} 
for  a  =  1,  ....  NUC 

(3.3) 


UCa  A    Vy  6  J  -  U    UC,\A,  *  Xj  =  Xk;  V;  €  /,  k  €7  -  {j}  -  U    UC,\ 


612  P  CALAMAI  ANDC.  CHARALAMBOUS 

for/3=  1,  ....  NCC 

(3.4)  CCP  A  I V,  €  /  -  V  Cq\Aip  =  ^  =  AT*;  fc  €  /,  Jfc  €  /  -  {j}  -  V  Cci 

NUC  A  number  of  unique  clusters. 
A^CC  A  number  of  coinciding  clusters. 
Note  that 

(a)  New  facility  j  coincides  with  existing  facility  ij  for  j  €  CP  (from  3.2). 

(b)  The  new  facilities  in  cluster  /3  coincide  with  existing  facility  ip  for  /3  =  1,  . . .  ,  A^CC 
(from  3.4). 

In  order  to  use  this  notation  for  the  derivation  of  the  new  algorithm  given  in  the  next  sec- 
tion define  a  unit  vector  D  in  E2n  as  follows: 

DT={D{,  ....  AH 

where 

(3.5)  Dj=  [dji  dj2],      7=1,  •••-  n 
and 

ILDll2=l. 

4.   THE  DIRECTIONAL  DERIVATIVE 

Using  the  notation  given  in  the  last  section  we  can  write  the  directional  derivative  of  the 
objective  function  at  X'm  the  direction  D  in  the  following  useful  manner: 


=    I    [G>--I>y] 

7  6  W 

,/eCP 
NUC 

+   11    [^•^+     I    v.JlA-  Dk\\2] 

a=\     \j£UCa  k^UCa 

NCC     . 

(4.1)  +11    \Gj-Dj+     £    VjJlDy-Alla+HUlDyl 

where 

(4.2a)  G,  =  £     | fy  7    y  |      +  I    I  '     J     .,         V,  6  W> 

(42b)        °'-£ife^t+£Tfe^t  V^€CP 


SOLVING  MULTIFACILITY  LOCATION  PROBLEMS  613 


<4-2c>  GJ-Z      ifr.ly.lL   +  X. 


vjk{Xj-Xk)    |  ^    w7,(A}-^) 


v,  €  f/Ca 


k<lUCn 


Xj-Xk\\2  +f?,  WXj-AMi       «-i.  ■■■■  ^c 


It  should  be  noted  that  in  each  case,  the  expression  for  Gj  is  the  gradient  of  that  part  of 
the  objective  function  f{X),  which  is  differentiate  with  respect  to  Xj.  In  the  case  where 
j  €  UP,  the  expression  is  the  exact  gradient  with  respect  to  Xj\  in  all  other  cases,  the  expres- 
sion for  Gj  can  be  considered  a  pseudo-gradient  of  f(X)  with  respect  to  Xj. 

Since  f(X)  is  a  convex  function,  the  point  X*  in  E2n  is  a  minimum  for  this  function  if 
and  only  if  the  directional  derivative  dDf{X*)  is  nonnegative  for  all  unit  vectors  D  in  Eln.  This 
fact  will  be  used  in  the  next  section. 

5.   NECESSARY  CONDITIONS  FOR  OPTIMALITY 

THEOREM  1:  If  the  following  conditions  are  not  satisfied  at  the  point  X  in  E2n,  then  the 
directional  derivative  given  by  expression  (4.1)  will  be  negative  for  some  unit  vector  D  in  E2n. 

(5.1)  (1)    ||G,||2-0         V7  €  UP 

(5.2)  (2)    ||G,||2<  wj,j        Vj  €  CP 

(5.3)  (3)    fora=  1,  ...  ,  NUC 

HZ  GA<  £       E      v;*      vsc  t/ca 

7€5  ,/€S    k£[UCa-S] 


(5.4)  (4)    for/i 


-M*6[cc„-n  p 


PROOF:  The  proofs  for  conditions  1  and  2  are  obvious. 
I  Dj  =  R    for  j  6  iS 


then  </fl/(Ar)  =£(?,-•/?  +£  £         v;j|/?||2 

y€5  y€5    /c€[f/C0  -  S] 

-H/?ll2lllZG/ll2cose+2:       I      V 

I    yes  yes  /ce[t/ca-s] 

Therefore,  rfD/U)  )0     VD  only  if 

III  Gj\\2<  I         I       ¥;»        V5C  UCa. 

j£S  j£S    k€[UCa-S] 

The  proof  for  condition  4  is  similar. 


614  P.  CALAMAI  AND  C   CHARALAMBOUS 

6.   UPDATE  FORMULAS 


As  a  result  of  the  preceeding  optimality  conditions  the  following  update  formulas  are  con- 
structed: 

CASE  1:    If  3/  €  UP  such  that  11(7,11  ^  0  then  the  direction  of  steepest-ascent  in  the 
subspace  defined  by  X,  is  Gj  =  Gr   We  therefore  use  the  following  update  formula  for  Xy 


Xj  * — Xj  —  Kj  Gj 


where 
(6.1) 


=    y  v* +  y  ZS. 


CASE  2:    If  3/  6  CP  such  that  llGylb  >  wjt  then  the  direction  of  steepest-ascent  in  the 
subspace  defined  by  Xj  is  Gj  =  Gj.  We  therefore  use  the  following  update  formula  for  Xy. 
Xj  —  Xj  -  \j  Gj 

where 


(6.2) 


+  1- 


fe  \\Xj-xk\\2     fe  \\Xj-AMt 


CASE  3:   If  35  C  UCa,a  =  1, 


NUC,  such  that 


IIIG,||2>I  £         v„ 

/€S  y€5    *€[{/CQ  -  5) 

then   the   direction   of  steepest-ascent    in   the   subspace   defined   by    the   subset   cluster   is 
Gs  =  £  Gj.   We  therefore  use  the  following  update  formula: 

Jts 

V,  €  5  J;  —  X,-ksGs 


where 
(6.3) 


y   na +  y zt 


CASE  4:   If  37C  CQ,  0  =  1,  . . .  ,  A^CC,  such  .that 

III^||2>   l[|       I         vJ  +  mJ 
jZT  j€T  \\k€lCCrT)  P] 

then  the  direction  of  steepest-ascent  in  the  subspace  defined  by  the  subset  cluster  is  GT  = 
£  Gj.   We  therefore  use  the  following  update  formula: 


Xj  <—  Xj  -  XT  GT 


SOLVING  MULTIFACILITY  LOCATION  PROBLEMS  615 


where 
(6.4) 


^tJcc,  WXj-XkWi  +k,  \\Zj-4\\2 


In  each  result,  the  expression  for  lambda  (A)  can  be  considered  a  weighted  harmonic 
mean  [8]  of  the  interfacility  distance  terms  appearing  in  the  equation  for  the  gradient  (Case  1) 
or  pseudo-gradients  (Cases  2  through  4) . 

7.  A  NEW  ALGORITHM 

Using  the  results  derived  in  the  preceeding  section  the  following  algorithm  can  be  used  to 
solve  Problem  P2: 

(1)  Find  a  current  solution  X  in  E2„- 

(2)  Try  to  obtain  a  better  solution  by  moving  single  new  facilities  by  using  Cases  1  and  2. 

(3)  For  a  =  1,  ....  NUC  try  to  obtain  a  better  solution  by  applying  the  special  form  of 
Case  3  where  \S\  =  1  (to  move  single  new  facilities)  or,  if  this  fails,  applying  the 
special  form  of  Case  3  where  |S|  =  I  UCa\  (to  move  entire  clusters  of  new  facilities). 
If  successful,  return  to  Step  2. 

(4)  For  p  =  1,  ....  NCC  try  to  obtain  a  better  solution  by  applying  the  special  form  of 
Case  4  where  |r|  =  1  (to  move  single  new  facilities)  or,  if  this  fails,  applying  the 
special  form  of  Case  4  where  I  T\  —  |CCp|  (to  move  entire  clusters  of  new  facilities). 
If  successful,  return  to  Step  2. 

(5)  Try  to  obtain  a  better  solution  by  moving  subset  clusters  using  Cases  3  and  4.  If  an 
improvement  is  made,  return  to  Step  2. 

8.  REMARKS  ON  IMPLEMENTATION 

The  following  rules  were  used  in  implementing  the  algorithm  described  in  the  last  section: 

(a)     New  facility  j  and  new  facility  k  were  considered  "clustered"  if: 
(8.1a)  ||*,.||2+|l*JI2<«i        Kj<k^n 

or 

where  e  i  A  inputted  cluster  tolerance, 

(b)    New  facility  j  and  existing  facility  /  were  considered  "coinciding"  if: 

(8.2a)  ll*,ll2+IU/[b<€,        j-  1 n;   ir  1 m 

or 

<8-2b)        iiiiif+liln! K  «•  J-1 n;'-1 m 


616  P.  CALAMAI  AND  C   CHARALAMBOUS 

where  e  i  A  inputted  cluster  tolerance, 

(c)  The  update  formulas  were  used  only  if: 

(8.3)  a||G||2>€2 

where  €2  A  inputted  step  tolerance.  This  helped  avoid  the  possibility  of  repeatedly  taking  small 
steps.  However,  the  step  tolerance  is  reduced  prior  to  the  termination  of  the  algorithm  as  out- 
lined by  the  next  rule. 

(d)  In  order  to  ensure  optimality,  the  following  check  is  made  prior  to  executing  Step  5  of  the 
algorithm: 

(8.4)  \f(X(h'l))  -  f(Xu'})]  *  100  <  e3  *  f(X{"-l)) 
where  €3  A  inputted  function  tolerance. 

If  this  condition  is  not  satisfied,  the  step  tolerance  (e2)  is  reduced  and  the  algorithm  res- 
tarted at  Step  2. 

9.   DISCUSSION 

The  new  algorithm  has  the  following  properties: 

(a)  It  makes  full  use  of  the  structure  of  the  facility  location  problem  thus  avoiding  the 
need  for  any  background  in  related  nonlinear  programming  areas. 

(b)  The  actual  objective  function,  and  not  an  approximation  to  it,  is  minimized  at  each 
step  in  the  algorithm. 

(c)  The  stepsize  used  in  this  algorithm  may  not  be  "optimal"  when  compared  with  step- 
sizes  obtained  from  line-search  techniques.  However,  the  use  of  this  stepsize  has  the 
following  advantages:  a)  ease  of  computation,  b)  maintenance  of  location  problem 
structure,  and  c)  reduced  computation  time  per  update. 

(d)  Although  Step  5  in  the  algorithm  is  combinatorial  in  complexity,  very  little  computa- 
tional work  is  necessary.  This  is  a  result  of  the  fact  that  all  the  information  needed 
for  this  step  has  already  been  computed  and  stored  in  previous  steps. 

(e)  The  algorithm  is  similar  to  the  technique  devised  by  Kuhn  for  solving  the  single- 
facility  location  problems  with  Euclidean  distances  [7]  and  the  method  devised  by 
Juel  and  Love  [6]  for  the  multifacility  location  problem  with  rectilinear  distances. 
This  makes  the  algorithm  attractive  to  those  with  experience  with  these  methods. 

(f)  Currently,  there  is  no  rigorous  proof  that  this  algorithm  converges.  In  1973,  Kuhn 
[7]  completed  the  proof  of  convergence  for  a  similar  scheme,  introduced  by 
Weiszfeld  [13]  in  1937,  for  the  case  of  single  new  facilities.  Based  on  computational 
experience  and  on  the  fact  that  the  algorithm  is  designed  to  minimize  the  objective 
function  in  all  new  facility  subspaces,  it  is  likely  that  the  algorithm  always  converges. 

(g)  The  main  disadvantage  of  the  algorithm  is  that  the  order  in  which  each  of  the  sub- 
spaces  is  checked  is,  currently,  not  optimal.  A  method,  based  on  projections,  that 
would  allow  us  to  determine  "a  priori"  which  subspace  to  update,  is  now  being  inves- 
tigated. 


SOLVING  MULTIFACILITY  LOCATION  PROBLEMS  617 

(h)  Most  existing  methods  for  solving  the  multifacility  problem  lack  any  consideration  of 
the  existence  of  constraints  on  the  solution  space  [9].  This  is  also  true  of  the  new 
method  outlined  in  this  paper;  however,  the  addition  of  constraints  should  not 
present  a  problem  to  the  projection  technique. 

(i)  It  has  yet  to  be  proven  that  the  necessary  conditions  for  optimality  for  Problem  P2, 
given  by  Equations  (5.1)  through  (5.4),  are  also  sufficient. 

10.   COMPUTATIONAL  EXPERIENCE 

The  performance  of  the  algorithm  described  in  this  paper  (MFLPV1)  was  tested  against 
the  hyperboloid  approximation  procedure  (HAP)  described  in  Eyster,  White  and  Wierwille  [2] 
and  a  modified  hyperboloid  approximation  procedure  (MHAP)  suggested  by  Ostresh  [11]. 

Two  parameters  were  used  as  a  basis  of  comparison:  1)  the  number  of  new  facility  loca- 
tion updates  needed  to  reach  optimality,  and  2)  the  required  CPU  time  in  minutes.  In  the  case 
of  program  MFLPV1,  two  counts  were  considered  necessary  for  specifying  the  first  parameter. 
The  first  count  represented  the  number  of  "attempted"  updates  (excluding  those  updates  from 
Step  5  of  the  algorithm).  The  second  count  represented  the  number  of  "successful"  updates. 
The  reason  for  excluding  the  number  of  attempted  updates  from  Step  5  of  the  algorithm  is  sim- 
ply this:  computationally,  very  little  work  is  done  at  this  step  in  the  procedure. 

Six  problems  were  used  for  the  comparison;  the  first  three  were  taken  from  [5]  (#5.23, 
#5.7  and  #5.6  respectively),  the  fourth  appears  in  [2]  and  the  last  two  problems  summarized  in 
Tables  1  and  2,  are  the  authors. 

HAP  and  MHAP  were  both  executed  using  two  different  initial  hyperbolic  constants  e<0) 
for  these  problems  in  order  to  emphasize  the  significance  of  this  parameter  to  the  performance 
of  these  algorithms.  The  stopping  criteria  used  in  each  case  was  the  same  as  that  outlined  in 
the  paper  introducing  HAP  [2].  Unless  otherwise  specified,  program  MFLPV1  also  made  use  of 
the  following  data. 

(1)  «i  A  cluster  tolerance  =  0.01  (from  Equations  (8.1)  and  (8.2)). 

(2)  e2  A  step  tolerance  —  0.05  (from  Equation  (8.3)). 

(3)  e3  A  function  tolerance  =  0.01  (from  Equation  (8.4)). 

The  results  of  these  tests  are  summarized  in  Table  3.  The  numbers  in  this  table  represent 
the  total  new  facility  updates  required  to  reach  optimality.  The  numbers  in  brackets  (  ),  under 
the  column  headed  MFLPV1,  represent  the  number  of  successful  updates  whereas  the  unbrack- 
eted  numbers  in  these  columns  represent  the  number  of  attempted  updates.  The  following 
observations  and  comments  can  be  made  about  the  results  summarized  in  this  table: 

(a)  In  all  but  Problem  5,  the  number  of  attempted  updates  required  to  reach  optimality 
using  MFLPV1  is  less  than  the  number  of  updates  required  by  HAP  and  MHAP. 
These  numbers  are  directly  comparable. 

(b)  The  new  procedure  (MFLPV1)  used  considerably  less  CPU  time  in  solving  the  six 
problems  than  did  HAP  and  MHAP. 


618 


P.  CALAMAI  AND  C.  CHARALAMBOUS 

TABLE  1  —  Input  Parameters  for  Problem  5 


i 

«/i 

an 

! 

0.0 

0.0 

2 

2.0 

4.0 

3 

6.0 

2.0 

4 

6.0 

10.0 

5 

8.0 

8.0 

J 

v<P) 

^0) 

1 

0.0 

0.0 

2 

0.0 

0.0 

3 

6.0 

10.0 

4 

1.0 

3.0 

5 

6.0 

10.0 

6 

8.0 

8.0 

7 

2.0 

4.0 

8 

2.0 

4.0 

9 

6.0 

10.0 

(a)    EF  Locations 


(b)  Initial  NF  Locations 


\     i 

12     3     4     5 

J  \ 

1 

1.0  1.0  1.0  1.0  1.0 

2 

1.0  1.0  1.0  1.0  1.0 

3 

1.0  1.0  1.0  1.0  1.0 

4 

1.0  1.0  1.0  1.0  1.0 

5 

1.0  1.0  1.0  1.0  1.0 

6 

1.0  1.0  1.0  1.0  1.0 

7 

1.0  1.0  1.0  1.0  1.0 

8 

1.0  1.0  1.0  1.0  1.0 

9 

1.0  1.0  1.0  1.0  1.0 

s   * 

1    2      3 

4     5      6     7     8     9 

y\ 

1 

Xj.o  1.0 

1.0  1.0  1.0  1.0  1.0  1.0 

2 

^vl.O 

1.0  1.0  1.0  1.0  1.0  1.0 

3 

1.0  1.0  1.0  1.0  1.0  1.0 

4 

V     1.0  1.0  1.0  1.0  1.0 

5 

\     1.0  1.0  1.0  1.0 

6 

N.         1.0  1.0  1.0 

7 

Nv        l.o  l.o 

8 

>v                1.0 

9 

(c)  Wji  Weights 


(d)  v,*  Weights 


TABLE  2  —  Input  Parameters  for  Problem  6 


*j\0)        Xjf 


i 

at\ 

aa 

1 

2.0 

5.0 

2 

10.0 

20.0 

3 

10.0 

10.0 

(a)  EF  Locations 


5.0 
5.0 


15.0 
15.0 


(b)  Initial  NF  Locations 


1 

2 

0.16      0.56      0.16 
0.16      0.56      0.16 

1  Sv     1.5 

2  1      \ 

(c)  Wjj  Weights 

(d)  vjk  Weights 

SOLVING  MULT1FACILITY  LOCATION  PROBLEMS 


TABLE  3  —  Comparative  Test  Results  for  Six  Problems 


# 

MFLPV1 

e(o)=  10o 

6(0)=    1Q-4 

X* 

fix*) 

HAP 

MHAP 

HAP 

MHAP 

1 

564  (77) 

1661 

1381 

2027 

1407 

(1.0,0.0) 
(1.0,0.0) 
(1.0,0.0) 
(2.0,0.0) 
(2.0,0.0) 

38.990 

2 

148  (34) 

647 

546 

4641 

2281 

(10.0,20.0) 
(10.0,20.0) 

186.798 

3 

63  (16) 

87 

70 

770 

197 

(8.0,7.0) 
(8.0,7.0) 

43.351 

4 

31  (15) 

45 

45 

45 

45 

(2.832,2.692 
(5.096,6.351) 

67.250 

5 

223  (40) 

142 

114 

1763 

975 

(4.045,4.281) 
(4.045,4.281) 
(4.045,4.281) 
(4.045,4.281) 
(4.045,4.281) 
(4.045,4.281) 
(4.045,4.281) 
(4.045,4.281) 
(4.045,4.281) 

201.878 

6 

63  (7) 

242 

164 

3743 

1869 

(10.0,20.0) 
(10.0,20.0) 

8.540 

TOTAL 

1092  (189) 

2824 

2320 

12989 

6774 

CPU 

0.07 

0.45 

0.50 

1.88 

1.48 

(c)  Five  of  the  six  problems  have  solutions  at  cluster  points.  This  appears  to  be  the  case 
in  many  other  problems.  This  suggests  that  methods  using  clustering  information, 
such  as  MFLPV1,  will  perform  better  than  methods  that  disregard  this  information. 

10.  CONCLUDING  REMARKS 

To  date,  many  of  the  methods  designed  for  solving  the  multifacility  location  problem  have 
been  either  poorly  structured,  suboptimal  or  haphazard.  In  this  paper,  a  new  method  is 
developed  for  solving  the  multifacility  location  problem  involving  Euclidean  distances.  This 
new  method  can  easily  be  extended  to  accommodate  problems  involving  item  movements  that 
are  other  than  Euclidean.  Computational  experience  shows  that  this  method  outperforms  tech- 
niques currently  in  use.  In  addition,  the  proposed  method  takes  full  advantage  of  the  structure 
of  the  location  problem. 


Most  current  techniques  used  for  solving  location  problems,  including  those  proposed  in 
this  paper,  are  designed  to  minimize  an  unconstrained  objective  function.  This  is  an  incom- 
plete treatment  since  most  practical  problems  involve  some  form  of  spatial  constraints.    It  is 


620  P  CALAMAI  AND  C.  CHARALAMBOUS 

proposed  that  these  constraints  be  handled  and  the  performance  of  the  algorithm  improved 
through  the  use  of  projection  techniques.  This  approach  is  currently  being  investigated  by  the 
authors. 

BIBLIOGRAPHY 

[1]  Cooper,  L.,  "Location- Allocation  Problems,"  Operations  Research,  77,  331-344  (1963). 
[2]  Eyster,  J.W.,  J. A.  White  and  W.W.  Wierwille,  "On  Solving  Multifacility  Location  Problems 

Using   a   Hyperboloid   Approximation   Procedure,"    American   Institute   of  Industrial 

Engineers  Transactions,  5,  1-6  (1973). 
[3]  Francis,  R.L.  and  A.V.  Cabot,  "Properties  of  a  Multifacility  Location  Problem  Involving 

Euclidean  Distances,"  Naval  Research  Logistics  Quarterly,  79,  335-353  (1972). 
[4]  Francis,  R.L.  and  J.M.  Goldstein,  "Location  Theory:  A  Selective  Bibliography,"  Operations 

Research,  22,  400-410  (1974). 
[5]    Francis,    R.L.   and  J. A.   White,   "Facility  Layout  and  Location:  An  Analytic  Approach" 

Prentice-Hall,  Englewood  Cliffs,  New  Jersey  (1974). 
[6]  Juel,  H.  and  R.F.  Love,  "An  Efficient  Computational  Procedure  for  Solving  the  Multi- 
Facility  Rectilinear  Facilities  Location  Problem,"  Operational  Research  Quarterly,  27, 

697-703  (1976). 
[7]   Kuhn,  H.W.,  "A  Note  on  Fermat's  Problem,"  Mathematical  Programming,  4,  98-107 

(1973). 
[8]  Kuhn,  H.W.  and  R.E.  Kuenne,  "An  Efficient  Algorithm  for  the  Numerical  Solution  of  the 

Generalized  Weber  Problem  in  Spatial  Economics,"  Journal  of  Regional  Science,  4,  21- 

33  (1962). 
[9]  Love,  R.F.,  "Locating  Facilities  in  Three-Dimensional  Space  by  Convex  Programming," 

Naval  Research  Logistics  Quarterly,  76,  503-516  (1969). 
[10]  Miehle,  W.,  "Link-Length  Minimization  in  Networks,"  Operations  Research,  6,  232-243 

(1958). 
[11]  Ostresh,  L.M.,  "The  Multifacility  Location  Problem:  Applications  and  Descent  Theorems," 

Journal  of  Regional  Science,  17,  409-419  (1977). 
[12]  Vergin,  R.C.  and  J.D.  Rogers,  "An  Algorithm  and  Computational  Procedure  for  Locating 

Economic  Activities,"  Management  Science,  13,  240-254  (1967). 
[13]  Weiszfeld,  E.  "Sur  le  Point  pour  Lequel  la  Somme  des  Distances  de  n  Points  Donnes  Est 

Minimum,"  Tohoku  Mathematical  Journal,  43,  355-386  (1936). 


AN  EASY  SOLUTION  FOR  A  SPECIAL 
CLASS  OF  FIXED  CHARGE  PROBLEMS 


Patrick  G.  McKeown 

College  of  Business  Administration 

University  of  Georgia 

Athens,  Georgia 

Prabhakant  Sinha 

Graduate  School  of  Management 

Rutgers—  The  State  University 

Newark,  N.J. 

ABSTRACT 

The  fixed  charge  problem  is  a  mixed  integer  mathematical  programming 
problem  which  has  proved  difficult  to  solve  in  the  past.  In  this  paper  we  look 
at  a  special  case  of  that  problem  and  show  that  this  case  can  be  solved  by  for- 
mulating it  as  a  set-covering  problem.  We  then  use  a  branch-and-bound  in- 
teger programming  code  to  solve  test  fixed  charge  problems  using  the  set- 
covering  formulation.  Even  without  a  special  purpose  set-covering  algorithm, 
the  results  from  this  solution  procedure  are  dramatically  better  than  those  ob- 
tained using  other  solution  procedures. 


1.   INTRODUCTION 

The  linear  fixed  charge  problem  may  be  formulated  as: 

(1)  Min  I.ejXj  +  'Z.fjyj 

(2)  Subject  to      £  djjXj  >  bt      i  €  /, 

(F) 

j  1  ifxj  >  0 

(3)  yJ  =  \  0  otherwise  J '^  7' 

(4)  and  Xj  >  0,  j  €  J. 

for  /=  {1,  ....  m)  and/  =  {1,  ...  ,  n). 

In  addition  to  continuous  costs,  the  variables  have  fixed  costs  which  are  incurred  when 
the  corresponding  continuous  variable  becomes  positive.  All  cost  are  assumed  to  be  nonnega- 
tive.  Problem  (F)  is  very  similar  to  the  standard  linear  programming  problem,  differing  only  in 
the  presence  of  the  fixed  costs.  In  spite  of  this  similarity,  it  has  proven  to  be  an  extremely 
difficult  problem  to  solve. 

If  all  the  continuous  costs  are  zero,  we  have  a  special  case  of  the  fixed  charge  problem 
which  we  will  refer  to  as  problem  (PF).  Problems  of  this  type  can  occur,  for  example,  when- 
ever there  is  a  need  to  find  solutions  with  the  least  number  of  basic,  nondegenerate  variables. 


622  P.  MCKEOWN  AND  P.  SINHA 

In  a  network  context,  Kuhn  and  Baumol  [4]  discuss  the  need  to  know  the  least  number  of  arcs 
necessary  to  carry  a  desired  flow.  Also,  in  the  survey  processing  field,  it  often  becomes  neces- 
sary to  check  a  record  of  replies  to  a  questionnaire  and  to  determine  changes  to  make  the 
record  consistent.  In  this  case,  it  is  necessary  to  know  the  minimum  number  of  such  changes 
that  are  necessary  for  consistency.  Both  of  these  problems  are  examples  of  problem  (PF)  with 
the  former  having  the  standard  transportation  constraint  matrix  and  the  latter  having  a  general 
constraint  matrix  which  depends  upon  the  consistency  conditions. 

A  special  case  of  problem  (PF)  occurs  when  all  the  constraint  coefficients  are  nonnega- 
tive,  i.e.,  a,j  ^  0  for  all  /,  j.  We  will  refer  to  this  problem  as  (PF+)  since  we  retain  the  condi- 
tion that  all  continuous  costs  are  equal  to  zero.  In  this  paper,  we  will  demonstrate  a  solution 
procedure  for  (PF+)  based  on  a  revised  formulation  for  the  problem.  We  then  use  a  branch- 
and-bound  integer  programming  code  to  solve  the  revised  formulation.  The  results  from  this 
approach  will  be  compared  to  those  obtained  using  other  procedures. 

2.  A  REVISED  FORMULATION 

The  problem  in  which  we  are  interested  may  be  formulated  as  follows: 

(5)  Min  J^fjyj 

JtJ 

(PF+) 

subject  to      (2)  -  (4) 

(6)  where  av  >  0  for  /  €  /,  j  €  J 

(PF+)  remains  a  special  case  of  the  fixed  charge  problem  (F)  so  any  results  that  are  applicable 
to  problem  (F)  will  also  be  applicable  to  (PF+). 

Two  previously  derived  results  for  (F)  that  are  of  particular  interest  to  (PF+)  are: 

1)  any  optimal  solution  to  (PF+)  will  occur  at  a  vertex  of  the  continuous  constraint 
set  (2)  and  (4)  (Hirsh  and  Dantzig  [3]); 

2)  a  lower  bound,  L0,  on  the  sum  of  the  fixed  costs  can  be  found  by  solving  the 
set-covering  problem,  Ps,  below  (McKeown  [5]). 

Min  L0  -  J  fjyj 

w 

(7)  Subject  to     £8^  >  1,    i  €/ 

(8)  yj  €  (0, 1),  j  €  J 

i  1  if  av  >  0 

(9)  where  *(/- j  0  otherwise, 


EASY  SOLUTION  FOR  FIXED  CHARGE  PROBLEMS 


We  will  combine  these  two  results  to  develop 
summarized  in  Theorem  1  below. 


i  solution  procedure,  the  essence  of  which  is 


THEOREM  1:  Let  Bg  =  [j\yj  =  1  in  an  optimal  solution  to  P8},  then  there  exists  a  feasi- 
ble solution  to  (PF+)  such  that  xt  >  0  for  j  6  Bg.  Furthermore,  this  solution  will  be  optimal 
for  (PF+). 

PROOF:  Given  an  optimal  solution  to  P8,  we  must  show  that  there  exists  a  correspond- 
ing solution  to  (PF+).  The  first  thing  to  note  is  that  each  column  of  the  constraint  matrix  (7) 
of  Ph  in  Bg  has  at  least  one  nonzero  element  that  is  the  only  nonzero  element  in  that  row. 
Otherwise,  the  set  would  be  over-covered  and  we  could  reduce  the  objective  value  of  /*§  by 
removing  that  column  from  the  optimal  solution.  We  may  use  this  result  together  with  the 
nonnegativity  of  the  ai:i  elements  to  construct  a  solution  to  (PF+)  using  Bg. 

Assume,  without  loss  of  generality,  that  \Bg\  =  k  and  that  the  decision  variables  have 

been  reindexed  such  that  {1 k]  €  Bg,  i.e.,  the  first  k  variables  of  (PF  +  )  correspond  to 

the  optimal  basic  variables  of  Ph.   We  can  now  construct  a  feasible  solution  to  (PF+)  using  the 
following  two  rules: 


1) 

X]  =  Max  {bjaix} 
an  *  0 
/  €  / 

Max     \bi- 

-zVJ 

2) 

xk  =  Max 

n            -^  a    ' 

j—  1           J 

u,  aik  ?=  u 
/  €  / 

aik 

This  proves  the  existence  of  a  solution  to  (PF  +  )  corresponding  to  Bg.  The  optimality  of 
this  solution  is  guaranteed  by  the  fact  that  both  {Ph)  and  (PF+)  have  the  same  objective  value 
and  that  this  objective  value  for  Ps  is  a  lower  bound  on  (PF+).  Hence,  Bg  corresponds  to  an 
optimal  solution  to  Ph. 

3.  COMPUTATIONAL  COMPARISONS 

Since  the  optimal  set  of  variables  for  (PF+)  can  be  found  by  solving  the  set-covering 
problem,  />§,  we  should  be  able  to  use  this  result  to  reach  quicker  solutions  to  (PF+).  We 
used  a  mixed  integer  programming  code  based  on  the  approach  of  Tomlin  [7]  as  extended  by 
Armstrong  and  Sinha  [1]  to  solve  the  set-covering  problems.  Special-purpose  set-covering  algo- 
rithms can  be  expected  to  perform  even  better.  Fixed  charge  test  problems  first  generated  by 
Cooper  and  Drebes  [2]  were  used  as  a  basis  of  comparison  between  this  set-covering  approach 
and  two  other  procedures.  The  first  such  procedure  is  a  branch-and-bound  code  developed  by 
McKeown  [6]  specifically  for  fixed  charge  problems  while  the  second  procedure  used  the  same 
mixed  integer  code  as  before,  but  solved  (PF+)  as  a  mixed  integer  problem. 


The  original  test  problems  were  of  dimension  5x10,  but  larger  problems  were  generated 
by  putting  these  smaller  problems  on  the  diagonal.  Using  these  problems,  the  results  of  our 
comparisons  are  shown  in  Table  1  below. 


P.  MCKEOWN  AND  P.  SINHA 


Problem 

Set 

Size 

Number 

of 
Problems 

Average  Solution  Time  per  Problem  in 
CPU  Seconds  on  CDC  70/74 

Armstrong 
and 
Sinha 

McKeown 

r  Set. 
Covering 

Solutions 

1 

2 
3 

5  x  10 
10  x  20 
15  x  30 

12 
6 
4 

0.132 
0.856 
3.039 

0.049 
0.345 
1.357 

0.017 
0.046 
0.101 

10 

4 
2 

From  the  table  we  can  see  that  the  set  covering  formulation  is  almost  three  times  faster 
than  the  best  alternative  approach  for  the  small  (5  x  10)  problems  and  up  to  13  times  faster  for 
the  larger  problems  (15  x  30).  We  have  also  noted  the  number  of  problems  for  which  the 
linear  programming  solution  was  integer  feasible  for  the  set  covering  problems.  This  occurred 
in  over  half  of  the  cases. 

4.   CONCLUSIONS 

In  this  paper,  we  have  shown  that  a  fixed  charge  problem  with  nonnegative  constraint 
matrix  coefficients  and  all  continuous  costs  equal  to  zero  can  be  solved  by  solving  a  related  set- 
covering  problem.  Computational  experience  confirms  that  this  procedure  yields  dramatically 
better  solution  times  than  any  other  available  solution  procedure.  Even  quicker  solution  times 
can  be  expected  to  result  if  special  purpose  set-covering  codes  are  used. 

REFERENCES 


[1]  Armstrong,  R.D.  and  P.  Sinha,  "Improved  Penalty  Calculations  for  a  Mixed  Integer 
Branch-and-Bound  Algorithm,"  Mathematical  Programming,  21,  474-482  (1974). 

[2]  Cooper,  L.  and  C.  Drebes,  "An  Approximate  Solution  Method  for  the  Fixed  Charge  Prob- 
lem," Naval  Research  Logistics  Quarterly,  8,  101-113  (1976). 

[3]  Hirsch,  W.M.  and  G.B.  Dantzig,  "The  Fixed  Charge  Problem,"  Naval  Research  Logistics 
Quarterly,  15,  413-424  (1968). 

[4]  Kuhn,  H.W.  and  W.J.  Baumol,  "An  Approximative  Algorithm  for  the  Fixed-Charges  Tran- 
sportation Problem,"  Naval  Research  Logistics  Quarterly,  9,  1-15  (1962). 

[5]  McKeown,  P.G.,  "A  Vertex  Ranking  Procedure  for  Solving  the  Linear  Fixed-Charge  Prob- 
lem," Operations  Research,  23,  1183-1191  (1975). 

[6]  McKeown,  P.G.,  "A  Branch-and-Bound  Algorithm  for  the  Linear  Fixed  Charge  Problem," 
Working  Paper,  University  of  Georgia  (1978). 

[7]  Tomlin,  J. A.,  "Branch  and  Bound  Methods  for  Integer  and  Non-Convex  Programming," 
Integer  and  Nonlinear  Programming,  437-450,  J.  Abadie,  Editor,  (American  Elsevier  Pub- 
lishing Company,  New  York,  1970). 


THE  BOUNDED  INTERVAL  GENERALIZED  ASSIGNMENT  MODEL 

G.  Terry  Ross 

University  of  Georgia 
Athens,  Georgia 

Richard  M.  Soland 

The  George  Washington  University 
Washington,  D.C. 

Andris  A.  Zoltners 

Northwestern  University 
Evanston,  Illinois 

ABSTRACT 

The  bounded  interval  generalized  assignment  model  is  a  "many-for-one"  as- 
signment model.  Each  task  must  be  assigned  to  exactly  one  agent;  however, 
each  agent  can  be  assigned  multiple  tasks  as  long  as  the  agent  resource  con- 
sumed by  performing  the  assigned  tasks  falls  within  a  specified  interval.  The 
bounded  interval  generalized  assignment  model  is  formulated,  and  an  algo- 
rithm for  its  solution  is  developed.  Algorithms  for  the  bounded  interval  ver- 
sions of  the  semiassignment  model  and  sources-to-uses  transportation  model 
are  also  discussed. 


1.   INTRODUCTION 

In  general  terms,  assignment  models  represent  problems  in  which  indivisible  tasks  are  to 
be  paired  with  agents.  Given  a  measure  of  utility  (or  disutility)  associated  with  each  possible 
pairing,  the  objective  of  the  model  is  to  optimize  the  collective  utility  associated  with  assigning 
a  set  of  tasks  to  a  set  of  agents.  In  practical  applications,  the  number  of  tasks  typically  exceeds 
the  number  of  agents,  and  at  least  one  agent  must  be  assigned  two  or  more  tasks  if  all  tasks  are 
to  be  completed.  Examples  of  such  "many-tasks-for-one-agent"  problems  include  the  assign- 
ment of  engagements  to  a  firm's  personnel  [20],  points  of  distribution  to  facilities  [15],  geo- 
graphic units  to  district  centers  [21],  products  to  plants  [1],  inventory  items  to  warehouses  [8], 
harvestable  forest  compartments  to  a  labor  force  [12],  ships  to  shipyards  [11],  scholarships  to 
students  [18],  storage  compartments  to  commodities  [19],  jobs  to  computers  [3],  files  to  mass 
storage  devices  [2,13],  defect  checkpoints  to  inspectors  [17],  and  trips  to  ships  [7].  The  feasi- 
bility of  many-for-one  assignments  will  depend  on  the  agents'  abilities  to  complete  the  collec- 
tions of  tasks  assigned  to  them.  That  is,  the  subsets  of  tasks  that  can  be  assigned  to  each  agent 
are  determined  by  the  total  amount  of  effort  available  to  the  agent  and  the  amount  of  effort 
that  each  individual  task  requires. 


626  G.T.  ROSS,  R.M.  SOLAND  AND  A. A   ZOLTNERS 

Several  many-for-one  assignment  models  have  been  developed  which  take  into  account 
only  upper  limits  on  the  total  amount  of  effort  that  each  agent  may  expand.  Each  of  these 
models  is  a  special  case  of  a  model  developed  by  Balachandran  [3]  and  Ross  and  Soland  [14] 
called  the  generalized  assignment  model.  This  model  has  the  form: 

(1)  (P)     minimize      z  =  £  £  CyXy 

(2)  subject  to  £  xy  =  1  for  all  j  €  7, 

/€/ 

(3)  £  rijXij  <  b,  for  all  i  <E  /, 

(4)  xu  =  0  or  1  for  all  i  €  /,  j  £  J. 

where  /=  {1,2,  ...  ,  m)  is  an  agent  index  set,  J  =  {1,2,  ...  ,  n)  is  a  task  index  set,  ch 
represents  the  disutility  associated  with  an  agent  /',  task  j  assignment,  fy  >  0  denotes  the 
resource  burden  incurred  by  agent  /  in  completing  task  j,  and  b,  is  the  resource  available  to 
agent  /.   The  decision  variable  x„  is  interpreted  as 

II  if  agent  /  performs  task  j 
0  otherwise 

Constraints  (2)  and  (4)  insure  that  each  task  is  uniquely  assigned  to  a  single  agent,  and  con- 
straints (3)  insure  that  each  agent  expends  no  more  than  b,  resource  units  in  accomplishing 
assigned  tasks.  Differences  in  the  difficulty  of  tasks  and  differences  in  agents'  abilities  to  per- 
form the  tasks  are  reflected  in  the  values  of  the  parameter  r#. 

The  special  cases  of  (P)  place  various  restrictions  on  the  form  of  the  agent  resource  con- 
straint (3).  Francis  and  White  [9]  and  Barr,  Glover  and  Klingman  [5]  have  addressed  the  prob- 
lem in  which  constraints  (3)  have  the  form: 

(3a)  £  Xij  <  ht    for  all  i  €  /. 

Here  b,  denotes  the  number  of  jobs  agent  /  can  complete,  for  all  jobs  consume  only  one  unit  of 
an  agent's  resource  when  the  agent  performs  the  task  (i.e.,  ru  =  1  for  all  /  €  /,  j  €  /).  The 
model  (l,2,3a,4)  is  a  generalization  of  the  standard  assignment  problem  of  linear  programming 
in  that  it  permits  an  agent  to  undertake  more  than  one  task.  It  has  been  called  the  generalized 
assignment  problem  by  Francis  and  White  and  the  semi-assignment  problem  by  Barr,  Glover, 
and  Klingman. 

Caswell  [6],  DeMaio  and  Roveda  [8],  and  Srinivasan  and  Thompson  [16]  studied  the 
problem  in  which  (3)  is  replaced  by: 

(3b)  £  rjXy  <  b,     for  all  i  €  /. 

The  model  (l,2,3b,4)  explicitly  considers  differences  in  the  difficulty  of  tasks  incorporated  in 
the  parameter  /}.  Srinivasan  and  Thompson  called  this  model  the  sources-to-uses  problem  to 
reflect  the  interpretation  of  the  model  as  a  transportation  problem  in  which  the  demand  at  the 
y'-th  location,  /},  is  to  be  supplied  by  a  single  source. 

Practical  considerations  frequently  require  that  the  agents  expend  a  minimum  total 
amount  of  effort  in  completing  assigned  tasks.  Placing  both  minimum  and  maximum  restric- 
tions on  the  resources  each  agent  can  expend,  yield  assignments  which  neither  overburden  nor 


BOUNDED  INTERVAL  GENERALIZED  ASSIGNMENT  PROBLEM  627 

underutilize  the  agents.  Such  restrictions  arise  in  most  personnel  planning  applications  [20]. 
Managerial  policies  usually  require  an  equitable  distribution  of  work  across  agents.  Analagous 
restrictions  crop  up  in  other  contexts  as  well.  For  example,  in  machine  loading  models,  it  usu- 
ally is  desirable  to  balance  machine  workloads  rather  than  allowing  some  heavily  loaded  and 
some  lightly  loaded  machines.  In  facility  location  models,  capacity  constraints  may  restrict  both 
the  minimum  and  maximum  size  of  a  facility  to  avoid  diseconomies  of  scale  associated  with 
plant  sizes  outside  of  a  reasonable  range,  to  permit  piecewise  linear  approximation  of  concave 
cost  functions,  or  to  restrict  both  the  minimum  and  maximum  number  of  facilities  [15].  Simi- 
larly, territory  design  procedures  for  problems  of  political  districting,  school  districting,  and 
sales  districting  require  an  equitable  distribution  of  some  entity  (such  as  voters,  minority  stu- 
dents, or  sales  potential)  among  the  districts.  Finally,  in  some  applications,  upper  limits  on  the 
effort  an  agent  can  expend  may  be  irrelevant,  and  only  lower  limits  need  be  considered.  Such  a 
situation  arises  in  the  segregated  storage  problem  [19]  which  requires  only  that  a  minimal 
amount  of  storage  space  be  allocated  to  store  commodities  and  no  maximum  allocation  is 
specified. 

Thus,  from  the  standpoint  of  modeling  flexibility,  it  is  desirable  that  assignment  models 
consider  explicitly  upper  and/or  lower  bounds  on  the  efforts  agents  must  expend  in  completing 
assigned  tasks.  While  most  "many-for-one"  assignment  models  consider  upper  bounds,  lower 
bounds  have  largely  been  overlooked.  In  this  paper,  we  introduce  the  bounded  interval  gen- 
eralized assignment  model  and  discuss  how  existing  algorithms  can  be  modified  to  accommo- 
date lower  bounds  on  agent  workloads  for  this  model  and  its  special  cases. 

2.  THE  BOUNDED  INTERVAL  GENERALIZED  ASSIGNMENT  MODEL  AND 
ALGORITHMIC  CONSIDERATIONS 

The  bounded  interval  generalized  assignment  model  may  be  formulated  as  follows: 

(5)  IP*)    minimize      z  =  £  £  c^ 

/€/./€./ 

(6)  subject  to  £  xu :  =  *  f°r  a^  J  €  ^ 

/€/ 

(7)  a,  <  £  r0Xjj  <  b-,  for  all  /  €  /, 

(8)  x,j  =  Oor  1  for  all  /  €  /,  j£  J. 

Notice  that  (P*)  derives  from  (P).  Fortunately,  the  modeling  flexibility  achieved  through 
the  introduction  of  lower  bounds  a,  >  0  in  constraints  (3),  (3a),  or  (3b)  does  not  complicate 
significantly  the  computational  effort  required  to  solve  any  of  the  models  described  above. 
Rather,  as  we  shall  show,  straightforward  modifications  can  be  made  to  the  existing  algorithms 
for  the  semi-assignment  problem,  sources-to-uses  transportation  problem,  and  the  generalized 
assignment  problem.  The  interested  reader  should  consult  the  cited  references  for  the  details 
of  the  original  algorithms. 

In  the  case  of  the  semi-assignment  problem,  the  constraint  matrix  is  totally  unimodular, 
and  integer  solutions  can  be  obtained  using  the  simplex  method.  To  impose  the  lower  limit, 

(7a)  J\x,j  >  at     for  all/  €  /, 


628  G  T   ROSS,  R.M.  SOLAND  AND  A. A.  ZOLTNERS 

one  need  only  add  an  upper  bounded  slack  variable  s,  <  bt  —  a,  to  each  of  the  constraints  (3a) 
and  rewrite  them  as  equality  constraints.  Optimal  solutions  to  the  resultant  bounded  variable 
linear  program  will  be  integer  valued. 

Models  with  constraints  (3)  or  (3b)  are  not  totally  unimodular.  Hence,  the  solutions  of 
the  linear  programming  relaxation  (i.e.,  x,j  >  0  for  all  i,j)  need  not  be  integer.  Branch  and 
bound  approaches  have  been  developed  for  deriving  integer  optimal  solutions  which  solve 
linear  programming  relaxations  for  fathoming  and  to  compute  lower  bounds.  In  the  case  of 
(3b),  a  linear  programming  relaxation  is  the  standard  transportation  problem  [16];  and  in  the 
case  of  (3),  a  linear  programming  relaxation  is  the  generalized  transportation  problem  [3].  As 
in  the  case  of  the  semi-assignment  problem,  to  impose  constraints  (7)  or 

(7b)  aj  <  £  rjXiJ  <  bi    for  all/  €  / 

in  a  linear  programming  relaxation,  one  need  only  add  upper  bounded  slack  variables 
Sj  ^  bj  —  d)  to  constraints  (3)  or  (3b)  and  rewrite  them  as  equality  constraints. 

The  algorithm  developed  by  Ross  and  Soland  [14]  for  the  generalized  assignment  problem 
does  not  solve  a  linear  programming  relaxation  to  determine  the  lower  bounds.  Instead,  a 
Lagrangian  relaxation  is  solved  in  the  form  of  a  series  of  separable  binary  knapsack  problems. 
The  Lagrangian  relaxation  has  the  form: 

(8)    (PK)    minimize      ZK  =  £  £  cuxv  +  £  \j(\  -  £x„) 

HIJSJ  j£J  /€/ 

subject  to  £  rl}  Xy  <  b,    for  all  /  €  / 

ju 

xu  =  0  or  1  for  all  /  €  /,  j  €  J. 

The  value  of  each  \j  is  set  equal  to  c2y,  the  second  smallest  value  of  Cy  for  all  /'  €  /.  These  \j 
are  optimal  dual  multipliers  for  the  problem: 

(PL)    minimize      £  £  CyXu 

i£IJ€J 

subject  to      £  xu ■  =  1  for  all  j  €  7, 

/£/ 

0  <  xi}  <■  1       for  all  /  €  I  j  €  J. 

Thus,  determining  a  lower  bound  requires  two  steps.  First,  solve  (//.),  then  solve  (/\).  If  the 
primal  solution  X  =  (5ey)  to_(PL)  should  also  satisfy  (8),  then  Z  =  ZL  =  Zx,  and  (Pk)  need 
not  be  solved.   Frequently,  J  will  not  satisfy  (8),  and  (Pk)  must  be  solved  to  find  Zx. 

To  incorporate  the  lower  bounds  a,  into  the  algorithm,  one  need  only  replace  constraints 
(8)  by  constraints  (7)  giving  rise  to  the  problem  (P*)  with  knapsack  constraints  bounded  both 
from  below  and  from  above.  Seemingly,  this  minor  modification  to  the  form  of  (PK)  should 
have  little  effect  on  the  algorithm.  However,  it  must  be  noted  that  (Px)  will  involve  fewer  0-1 
variables  and  may  be  easier  to  solve  than  (P*).  The  reason  is  best  explained  by  considering  an 
equivalent  form  of  the  objective  function  of  (PK): 

ZK  =  X  ^j  ~  maximum  [£  £  (\,  -  cy)  xJ. 


BOUNDED  INTERVAL  GENERALIZED  ASSIGNMENT  PROBLEM  629 

Clearly,  with  constraints  (8),  one  can  set  any  xu  equal  to  zero  which  has  an  objective  function 
coefficient  ikj  -  cu)  ^  0.  Thus,  using  the  values  of  Xj  calculated  from  solving  (Pi),  (PK) 
reduces  to  a  problem  involving  at  most  n  0  —  1  variables.  Such  a  reduction  is  not  possible  for 
(Pt)- 

In  addition  to  providing  a  lower  bound,  the  solutions  to  (PL)  and  (P*)  may  be  used  to 

select  a  branching  (or  separation)  _yariable  for  defining  subsequent  candidate  problems.    As 

noted  above,  the  solution  to  (PL),  X,  is  usually  not  feasible  to  (7).    In  essence,  the  solution  to 

(P*),  X=  (Xjj),  may  be  interpreted  as  recommending  changes  in  X  which  must  be  made  in 

order  to  satisfy  (7).    That  is,  it  is  possible  that  for  some  j  €  7,  £5c,7  =  0  to  avoid  overloading 

,€/ 
any  agent  or  £  xu  >  1  to  insure  every  agent  uses  a  minimum  amount  of  his  resource.   Those 

variables  xu  with  an  optimal  value  of  one  indicate  agent-task  pairings  that  should  be  made; 
whereas,  those  x,7  with  an  optimal  value  of  zero  indicate  pairings  that  should  be  avoided.  Thus, 
these  variable  values  indicate  changes  that  will  reduce  the  aggregate  infeasibility  of  Jin  (7), 
and  they  are  helpful  in  choosing  a  branching  variable. 

To  formalize  the  concept  of  reducing  aggregate  infeasibility,  we  define  the  infeasibility  in 
constraint  /  prior  to  taking  a  branch  to  be 

A  =  max  {0,  df,  d~) 

where    d,+  =  £  r,y3cy  —b„ 


The  set  /+  =  {/  €  l\d,+  >  0}  identifies  those  constraints  (7)  for  which  A^exceeds  the  upper 
bound,  and  /"  =  {/€  l\d~  >  0}  identifies  those  constraints  (7)  for  which  A' fails  to  satisfy  the 
lower  bound. 

Suppose  I+  ^  0  and  k€[j  €  J\xjj  =  1  and  /  €  /+};  if  xik  is  set  to  0  then  d*  and  d~ 
become: 

di+  =  Z  ruXij  ~  bj  -  rik 

d,r  =  a, •  ~  £  ru  Xy  +  rik 

and  the  resulting  infeasibility  in  constraint  /  becomes 

A*=  max{0,  di+,  d~\. 

Assuming  that  task  k  is  reassigned  to  the  second  least  costly  agent,  (say  agent  /?,  where 
chk  =  min  C\k)  then  the  infeasibility  in  constraint  h  becomes 

A*=  max  {0,  d£,  d^\ 
where 

dh  =  Z  n,j  *hj  -  bh  +  rhk 

dh  =  a,,  -  ZoyX/y  -  rhk. 


630  G.T.  ROSS.  R.M.  SOLAND  AND  A. A.  ZOLTNERS 

Hence,  the  net  difference  in  total  infeasibility  is: 

kD*=  (Z>,  +  D,,)-  (Df  +  Dlj) 

If  AZ)*  >  0  then  setting  xik  =  0  yields  a  reduction  in  aggregate  infeasibility,  and  if  AZ)*  <  0 
then  such  a  branch  will  not  reduce  aggregate  infeasibility. 

Similarly,  suppose  that  l~  ^  0  and  k  €  [j  €  y|xy  =  0  and  /  €  /""};  if  xik  were  set  to  1 
then  df  and  d~  become: 

d,+  =  Z  rv  xu  -  b,  +  rik 

d,~  =  a, ■  -  £  ry  xy  -  rik 
J& 

and  the  resulting  infeasibility  in  constraint  /  would  be 

Z)/*-  max  {0,  #\  4-}. 

If  Xft   is  set  to   1   then  task  k  is  assigned  to  agent  /  and  agent  ik   relinquishes  it,  where 
ik  =  min  cik.   Hence,  the  infeasibility  for  constraint  ik  becomes 

Dtk  -  max  {0,  df ,  4" } 

where 

df  =  £  V*V€  ~  *'*  ~  V 

/€/     ' 

di~  =  «/A  -  £  r,^  *,y  +  r,  *. 

/€./ 

The  net  difference  in  infeasibility  is 

AA*=  (A  +  D,k)-  (D^+  Z)£) 

where  Z),  and  Z>,   are  the  infeasibilities  in  constraints  /and  4  prior  to  any  branch.   As  before,  if 
AD*  >  0  then  there  is  reduced  infeasibility  following  a  branch  on  variable  xlk. 

Several  rules  for  selecting  the  branching  variable,  *,.,.,  are  formulated  as  follows: 

I.    a)  Xpp  is  that  variable  for  which 

A  £>/.*=        max       [ADf] 

(i,j)€  H+UH~ 

where       H+  =  { (/,  j)  \x0  =  0  and  /  €  /+} 
//-  =  {(/,y)|xy=  1  and/  €  /"} 
b)  If  AZ>/.*  =  0  in  a)  then  xfy.  is  that  variable  for  which 
AZ>/«*  =  max  {AZ>/} 

(/,  y)€  (G+~H+)  U(G--H-) 

where       G+  =  {(/,  y)|3cy  =  1  and  /  6  I+} 
Gr  =  {(i,j)\xu  =  Oand/  €  /"}. 


BOUNDED  INTERVAL  GENERALIZED  ASSIGNMENT  PROBLEM 

II.  a)  x(*j*  is  that  variable  for  which 

min       i . 


piy  =  min 


min 


(/,  j)  e  G+  ' 


(/,  j)  €  G~ 


ADf 


b)  If  AZ)/=  0  for  all  (/,  j)  €  G+UG~  then  xiT  is  that  variable  for  which 


max 


JdF: 


where      £+  =  {(/,  j)\xu  =  1  and  /  €  /+}, 

F/  denotes  the  set  of  tasks  assigned  to  agent  /  by  prior  branching. 

Rules  la  and  lb  are  designed  to  choose  that  variable  which  reduces  the  post  branch  aggregate 
infeasibility  by  the  greatest  amount.  Rule  Ha  conditions  the  choice  of  branching  variable  on  the 
additional  cost  incurred  per  unit  reduction  in  infeasibility.  Rule  lib  is  the  one  used  in  [14];  the 
variable  chosen  by  this  rule  represents  an  agent-task  pairing  which  should  be  made  considering 
the  penalty  for  not  doing  so  weighted  by  the  fraction  of  the  agent's  remaining  free  resources 
consumed  by  the  assignment. 

As  the  algorithm  progresses  and  new  candidate  problems  (CPs)  are  defined  by  the  branch- 
ing process,  the  additional  steps  given  below  may  be  taken  to  facilitate  fathoming.  These  steps 
are  specialized  adaptations  of  more  general  forcing  (or  variable  fixing)  tests  suggested  by  Balas 
[4]  and  Glover  [10]. 

In  solving  any  (CP),  any  x?y  for  which  r,y  >  b,- —  £   xj'jrj'j  mav  De  set  equal  to  zero. 

JtFr 
Here  F?  denotes  those  j  €  J  for  which  x,-7  has  been  assigned  a  value  of  zero  or  one  by  prior 
branching    or    variable    fixing    tests.     Similarly,    if   there    is    an   x,y    for    which    at>  —  £ 


I 

jzj-f; 


then  Xj'j>  must  be  set  equal  to  one  in  the  solution  to  (CP).    These 


variable  forcing  tests  may  subsequently  result  in  other  variables  being  forced  to  zero  or  to  one 
when  all  of  the  resultant  implications  are  considered.  Moreover,  forcing  certain  variables  to 
zero  or  to  one  in  the  solution  to  (CP)  may  affect  the  values  of  some  of  the  Xj  obtained  from 
solving  (Pi).  This  change  may,  in  turn,  increase  the  value  of  the  lower  bound  provided  by 
(P*)- 

Another  test  may  be  used  to  check  the  feasibility  of  (P*)  (or  any  candidate  subproblem). 
Summing  the  constraints  (7)  together  yields  the  constraint  (9): 

(9) 


B. 


This  new  constraint,  together  with  constraints  (6),  implies  that  for  any  feasible  solution  to  (P*) 
we  must  have: 


(10) 


Zo: 


A  and  J)  r 

JZJ 


632  O.T.  ROSS.  R.M.  SOLAND  AND  A  A.  ZOLTNERS 

where 

r',  =  max  {/•..}  and  r"=  min  {/•„}. 

The  values  necessary  for  the  tests  (10)  can  be  updated  easily  as  part  of  the  branching  process  in 
order  to  apply  this  test  to  each  (CP). 

The  algorithm  terminates  in  the  usual  way  when  all  candidate  problems  have  been 
fathomed. 

3.  CONCLUSION 

This  note  has  described  an  efficient  branch  and  bound  algorithm  for  the  bounded  interval 
generalized  assignment  problem.  The  algorithm  serves  as  a  useful  tool  for  solving  a  large 
number  of  applications  of  this  assignment  model,  a  representative  sample  of  which  is  men- 
tioned in  the  introduction. 

REFERENCES 

[1]  Abella,  R.J.  and  T.E.  Bova,  "Optimal  Plant  Allocation  of  Stockkeeping  Units,"  presented  at 
TIMS/ORSA  Joint  National  Meeting,  San  Francisco,  California  (May  1977). 

[2]  Babad,  J.M.,  V.  Balachandran  and  E.A.  Stohr,  "Management  of  Program  Storage  in  Com- 
puters," Management  Science,  23,  380-390  (1976). 

[3]  Balachandran,  V.,."An  Integer  Generalized  Transportation  Model  for  Optimal  Job  Assign- 
ment in  Computer  Networks,"  Operations  Research,  24,  742-759  (1976). 

[4]  Balas,  E.,  "An  Additive  Algorithm  for  Solving  Linear  Programs  with  Zero-one  Variables," 
Operations  Research,  13,  517-545  (1965). 

[5]  Barr,  R.S.,  F.  Glover  and  D.  Klingman,  "A  New  Alternating  Basis  Algorithm  for  Semi- 
assignment  Networks,"  Research  Report  CCS-264,  Center  For  Cybernetic  Studies, 
University  of  Texas,  Austin,  Texas  (January  1977). 

[6]  Caswell,  W.,  "The  Transignment  Problem,"  Unpublished  Ph.D.  Thesis,  Rensselaer 
Polytechnic  Institute  (1972). 

[7]  Debanne,  J.G.  and  J-N  Lavier,  "Management  Science  in  the  Public  Sector— The  Estevan 
Case,"  Interfaces,  9,  66-77  (1979). 

[8]  DeMaio,  A.  and  C.  Roveda,  "An  All  Zero-One  Algorithm  for  a  Certain  Class  of  Transpor- 
tation Problems,"  Operations  Research,  19,  1406-1418  (1971). 

[9]    Francis,    R.L.   and  J. A.    White,   Facility  Layout  and  Location:  An  Analytical  Approach, 
(Prentice-Hall,  Englewood  Cliffs,  New  Jersey,  1974). 
[10]  Glover,  F.,  "A  Multiphase-Dual  Algorithm  for  the  Zero-One  Integer  Programming  Prob- 
lem," Operations  Research,  13,  879-919  (1965). 
[11]  Gross,  D.  and  C.E.  Pinkus,  "Optimal  Allocation  of  Ships  to  Yards  for  Regular  Overhauls," 
Technical  Memorandum  63095,  Institute  for  Management  Science  and  Engineering, 
The  George  Washington  University,  Washington,  D.C.  (May  1972). 
[12]  Littschwager,  J.M.  and  T.H.  Tcheng,  "Solution  of  a  Large-scale  Forest  Scheduling  Problem 

by  Linear  Programming  Decomposition,"  Journal  of  Forestry,  65,  644-646  (1967). 
[13]  Morgan,  H.L.,  "Optimal  Space  Allocation  on  Disk  Storage  Devices,"  Communications  of 

the  ACM,  17,  139-142  (1974). 
[14]  Ross,  G.T.  and  R.M.  Soland,  "A  Branch  and  Bound  Algorithm  for  the  Generalized  Assign- 
ment Problem,"  Mathematical  Programming,  8,  91-103  (1975). 
[15]  Ross,  G.T.  and  R.M.  Soland,  "Modeling  Facility  Location  Problems  as  Generalized  Assign- 
ment Problems,"  Management  Science,  24,  345-357  (1977). 


BOUNDED  INTERVAL  GENERALIZED  ASSIGNMENT  PROBLEM  633 

[16]  Srinivasan,  V.  and  G.L.  Thompson,  "An  Algorithm  For  Assigning  Uses  to  Sources  in  a 
Special  Class  of  Transportation  Problems,"  Operations  Research,  21,  284-295  (1973). 

[17]  Trippi,  R.R,  "The  Warehouse  Location  Formulation  as  a  Special  Type  of  Inspection  Prob- 
lem," Management  Science,  21,  986-988  (1975). 

[18]  Wagner,  H.M.,  Principles  of  Operations  Research,  (Prentice-Hall,  Englewood  Cliffs,  N.J., 
1968). 

[19]  White,  J. A.  and  R.L.  Francis,  "Solving  A  Segregated  Storage  Problem  Using  Branch  and 
Bound  and  Extreme  Point  Ranking,"  AIIE  Transactions,  3,  37-44  (1971). 

[20]  Zoltners,  A. A.,  "The  Audit  Staff  Assignment  Problem:  An  Integer  Programming 
Approach,"  Working  Paper  74-34,  School  of  Business  Administration,  University  of 
Massachusetts,  Amherst,  Massachusetts  (September  1974). 

[21]  Zoltners,  A. A.,  "A  Unified  Approach  to  Sales  Territory  Alignment,"  Sales  Management: 
New  Developments  from  Behavioral  and  Decision  Model  Research  R.  Bagozzi,  Editor, 
(Cambridge,  Massachusetts  Marketing  Science  Institute,  1979),  360-376. 


THE  M/G/l  QUEUE  WITH  INSTANTANEOUS 
BERNOULLI  FEEDBACK* 

Ralph  L.  Disney 

Virginia  Polytechnic  Institute  and  State  University 
Blacksburg,  Virginia 

Donald  C.  McNickle 

University  of  Canterbury 
Christchurch,  New  Zealand 


Bell  Laboratories 
Holmdel,  New  Jersey 

ABSTRACT 

In  this  paper  we  are  concerned  with  several  random  processes  that  occur  in 
M/G/l  queues  with  instantaneous  feedback  in  which  the  feedback  decision  pro- 
cess is  a  Bernoulli  process.  Queue  length  processes  embedded  at  various  times 
are  studied.  It  is  shown  that  these  do  not  all  have  the  same  asymptotic  distri- 
bution, and  that  in  general  none  of  the  output,  input,  or  feedback  processes  is 
renewal.  These  results  have  implications  in  the  application  of  certain  decompo- 
sition results  to  queueing  networks. 


1.   INTRODUCTION 

In  this  paper  we  are  concerned  with  several  random  processes  that  occur  within  the  class 
of  M/G/l  queues  with  instantaneous  feedback  in  which  the  feedback  decision  process  is  a  Ber- 
noulli process.  Such  systems  in  the  case  G  =  M  are  among  the  simplest,  nontrivial  examples 
of  Jackson  networks  [8].  Indeed,  they  are  so  simple  that  they  are  usually  dismissed  from  con- 
sideration in  queueing  network  theory  as  being  obvious.  We  will  show  that  far  from  being 
obvious,  they  exhibit  some  important  unexpected  properties  whose  implications  raise  some 
interesting  questions  about  Jackson  networks  and  their  application. 

In  particular,  Jackson  [8]  observed  that  in  his  networks  the  vector-valued  queue  length 
process  behaved  as  if  the  component  processes  were  independent,  M/M/l  systems.  Since  those 
results  appeared  there  has  developed  a  mythology  to  explain  them.  These  arguments  usually 
rest  on  three  sets  of  results  that  are  well  known  in  random  point  process  theory:  superposition, 
thinning,  and  stretching.  By  examining  the  network  flow,  it  will  be  shown  that  the  applications 
of  these  results  are  inappropriate  for  queueing  networks  with  instantaneous,  Bernoulli  feedback. 
These  flows  are  considerably  more  complicated  than  one  expects  based  on  such  arguments. 


The  research  was  supported  under  ONR  Contracts  N00014-75-C-0492  (NR042-296)  andN00014-77-C-0743  (NR042-296). 


635 


636  RL.  DISNEY.  DC.  MCNICKLE  AND  B.  SIMON 

It  is  shown  that  in  general,  both  the  input  and  output  processes  of  the  M/M/l  queue  with 
feedback  are  Markov-renewal,  and  the  kernels  of  those  Markov-renewal  processes  are  given. 
The  output  of  the  M/G/l  queue  with  feedback  is  also  Markov-renewal,  and  that  kernel  is  given. 
It  is  shown  that  in  general  these  processes  are  never  renewal.  The  implications  of  these  facts 
are  discussed  in  Section  4. 

1.1  The  Problem  and  Notation 

We  assume  the  usual  apparatus  of  an  M/G/l  queue  with  unlimited  waiting  capacity.  The 
new  idea  is  that  a  unit  which  has  received  service  departs  with  probability  q  and  returns  for 
more  service  with  probability  p.  p  +  q  =  1 .  Without  loss  of  generality  for  the  processes  stu- 
died here,  the  returning  customer  can  be  put  anywhere  in  the  queue. 

To  establish  notation  it  is  assumed  that  the  arrival  process  is  a  Poisson  process  with  param- 
eter X  >  0.  The  arrival  epochs  are  the  elements  of  {Wn:  n  =  1,2,  ...}.  Service  times  are 
independent,  identically  distributed,  nonnegative,  random  variables,  S„  with 

Pr  [Sn  <  t]  -  Hit),       t  >  0, 
E[S„]  <  «>. 
We  define  H*is),  the  Laplace-Stieltjes  transform  of  Hit),  by 

H*is)=  f°°  e-s'dHit),    Res  ^  0. 
The  arrival  process  and  service  times  are  independent  processes. 


Service  completions  occur  at  T0  <  T]  <  T2  ...   called  the  output  epochs.   Let 


0,  if  the  n  -th  output  departs, 

1,  if  the  n-th  output  feeds  back. 


{ Yn}  is  a  Bernoulli  process. 


Elements  of  the  subset  {tn}  C  [Tn]  are  called  the  departure  epochs  and  are  the  times  at 
which  an  output  leaves  the  system.  The  elements  of  the  subset  {t,,}  c  {Tn}  are  called  the  feed- 
back epochs  and  are  the  times  at  which  an  output  returns  to  the  queue.  {/„}  U  {t„}  =  \Tn). 

The  times  T'n  are  the  times  at  which  a  unit  enters  the  queue.  {7^}  is  called  the  input  pro- 
cess, {t;,}  =  [w„)\j  Wn}. 

There  are  five  queue  length  processes  to  be  studied.  They  are  closely  related  as  will  be 
shown.  Let  Qit)  be  the  queue  length  (number  in  the  system)  at  t.  Then, 
Q-  („)  =  Q(wn  -  0);  Qi  in)  =  QiVn  -  0);  Qt  in)  =  QiTn  +  0);  Qt  in)  =  Qitn  +  0)  are 
respectively  the  embedded  queue  lengths  at  arrival  epochs,  input  epochs,  output  epochs,  depar- 
ture epochs. 

2.   QUEUE  LENGTH  PROCESSES 

The  queue  lengths  listed  in  Section  1.1  are  closely  related.  The  steady  state  versions  of 
[Qi  in)}  and  {Q^  in)}  are  of  primary  concern.  They  are  studied  in  Sections  2.1  and  2.2 
separately.  They  are  related  to  the  other  processes  in  Section  2.3.  The  important  special  case 
for  G  =  Mis  then  studied  in  2.4. 


M/GA  QUEUE  WITH  INSTANTANEOUS  BERNOULLI  FEEDBACK 


feedback  process 


J- 


arrival 

input 

output 

process 

process 

Figure  1. 

process 

2.1  The{(?4+(/!)}  Process 

There  are  several  ways  to  study  this  process.  The  following  appears  to  be  direct,  correct, 
and  may  help  explain  why  these  feedback  problems  have  received  such  little  attention  in  the 
queueing  literature.    First,  it  is  clear  that 

j  r„_,  +  S;,,  ifQfin  -  1)  >  0, 

'"  =  |  /„_,  +  /„  +  S^,   if  Q+  (n-\)  =  0. 

Here  S'n  is  the  total  service  time  consumed  between  the  (n  -  1)  -  st  and  «-th  departure.  /„  is 
the  idle  time  following  tn_x  when  Qf  (n  —  1)  =  0.  For  the  M/G/l  queue,  the  /'s  are  indepen- 
dent, identically  distributed,  random  variables  that  are  exponentially  distributed  with  parameter 


Without  loss  of  generality,  since  customers  are  indistinguishable, 

S'„  =  S]  +  s2  +  ...  sm, 

where  m  is  the  number  of  services  performed  between  the  (n  —  1)  -  st  and  n-th  departure. 
Since  [Yn}  is  a  Bernoulli  process,  m  is  geometrically  distributed  and  it  follows  that  {S'n}  is  a 
sequence  of  independent,  identically  distributed,  random  variables.  Thus,  the  Laplace-Stieltjes 
transform  of  the  distribution  function  of  S'„  is  easily  found  to  be 

G*(s)  =  qH*(s)/[\-  pHHs)]. 


Using  standard  embedded  Markov  chain  methods  [3,  167-174]  one  finds  that  the  probabil- 
ity generating  function  of  Jp,  the  limiting  probability  distribution  of  {^4"  («)},  is  given  by 

(0)  (z  -  1)  G*(k  -kz) 


(1) 

and 

(2) 


g(z)- 


■  G*(\  -kz) 


ir'(0)=  \-kE[Sn]/q. 


638 


R.L.  DISNEY,  DC.  MCNICKLE  AND  B.  SIMON 


If  one  is  willing  to  assume  that  the  M/G/l  queue  with  instantaneous,  Bernoulli  feedback 
has  a  queue  length  process  which  asymptotically  has  the  same  distribution  as  another  M/G/l 
queue  without  feedback,  then  (1)  and  (2)  follow  immediately.  This  assumption  is  valid  since  if 
customers  feedback  to  the  front  of  the  queue,  the  total  service  time  of  the  n-th  customer  is  S'„. 
{S'n}  is  a  sequence  of  independent,  identically  distributed,  random  variables  with  mean  E[S„]/q. 
Alternatively,  one  can  argue  that  the  M/G/l  queue  with  feedback  (as  defined  here)  has  the 
same  asymptotic  distribution  for  its  queue  length  process  as  an  M/G/l  queue  without  feedback 
if  one  takes  the  arrival  process  parameter  in  the  latter  case  to  be  k/q.  Indeed,  both  of  these 
assumptions  and  several  others  that  are  used  to  "prove"  that  these  queues  with  feedback  are 
trivial  have  now  been  proven  by  the  arguments  leading  up  to  (1)  and  (2).  That  these  argu- 
ments can  be  applied  more  generally  is  easily  proven.  In  the  remainder  of  this  paper,  in  Takacs 
[10]  and  in  Disney  [6]  it  is  shown  that  while  these  arguments  may  imply  that  the  study  of 
queue  lengths  at  departure  times  is  trivial,  the  same  cannot  be  said  for  other  processes  of 
interest. 

2.2  The  {Q?(n)}  Process 

This  is  the  queue  length  process  embedded  at  output  points.  Since  {/„}  C  {Tn},  {Q4  (n)} 
is  a  process  on  a  coarser  grid  than  {Q?  (n)}.  Since  one  is  ultimately  to  be  concerned  with  both 
{Qt  (n))  and  [Tn  -  r„_i},  the  following  study  is  for  the  joint  process  {Qfin),  Tn  -  T„_|}. 
The  marginal  results  for  [Q^  (n)}  then  will  be  easy  to  determine. 


THEOREM  1:   The  process  (C?3+  (n),   Tn  -  Tn_x)  is  a  Markov-renewal  process  with  kernel 
A  (i,j,x)  =  Pr{Q?  (n)  *  j,  Tn  -  fw_,  <  x\Q$  (n  -  1)  =  /}.   If  one  defines 


Pj(y)=  (ky)Je  Ky/jl, 


j  =  0,1,2,  .. 


A  (i,j,x)  =    < 


L 


if;  <  /  -  1, 


i+l(y)q)dH(y), 


if  /  ^  0, 


C  (1  -  e-k(x-y))  (Pj-](y)p  +  Pi{y)q)dH(y),\n  =  0, 

J  >  0, 


u; 


(1  _  e-i(x-y))  pQ(y)qdH(y),    if  j  =  /  =  0. 


PROOF: 

j  S„,  if  Q$(n  -  1)  >  0, 

Tn  -  rn_,  -  j  Jn  +  Sn>  .f  Q+  {n  _  1}  =  0 

where  /„  is  the  exponentially  distributed  idle  time  preceeding  S„  if  Q3+  (n  ■ 
then  follows  directly  using  arguments  as  in  [5].         □ 


1)  =  0.   The  result 


As  x  —  00,  A  (ij.x)  —  A  (ij)  the  one  step  transition  probability  for  the  {Q3  (n)}  process. 
Then  using  standard  embedded  Markov  chain  results  [3,  167-174]  one  can  show  that  the  proba- 
bility generating  function  g(z)  for  the  limiting  probabilities  n(j)  of  Q3+  («)  are  given  by 
tt(0)  (2  -  1)  (pzH*(k  -  \z)  +  qH*(k  -  kz)) 


(3) 


?(z)  = 


z  -  pzH*(k  -  kz)  -  qH*(k  -  kz) 


M/GA  QUEUE  WITH  INSTANTANEOUS  BERNOULLI  FEEDBACK  639 


and 

(4)  ir{0)  =  q-\E[S„l 


2.3   Other  Queue  Length  Processes 


The  queue  length  and  limiting  probabilities  for  the  queueing  processes,  {(?f(n)}, 
[Qi  (n)}  now  follow  from  a  theorem  found  in  Cooper  [3,  155].  From  this  it  follows  that 
{QU)},  {(?f  (/»)},  and  [Qfin)}  are  asymptotically,  identically  distributed  (see  Cooper  [3,  65]) 
and  Qi(n),  and  {Qi  in)}  are  asymptotically,  identically  distributed.  Clearly,  [Qf  in)}  and 
{Qf  in)}  are  not  asymptotically,  identically  distributed.  That  {Qt  in)}  and  [Qf  in)}  are  not 
asymptotically,  identically  distributed  can  be  seen  as  follows.  First,  in  the  set  up  of  studying  the 
(C?3+  in)}  process  one  must  decide  how  to  count  the  feedback  customer  when  he  appears.  The 
clean  way  to  do  this  is  to  use  Yn  as  defined  in  Section  1.1  and  *Qf  in)  as  the  number  in  the 
queue  not  including  the  outputting  customer.  Then  one  can  study  the  process  [Y„,  *Q?in)}. 
Indeed,  this  is  precisely  the  direction  used,  for  example,  in  d' Avignon  and  Disney  [4].  Then 
the  {Qi  in)}  of  Theorem  1  above  would  be  the  {*Qy  in)  +  Y„}  process  of  [4].  It  then  follows 
that  {*(?3+(«)}  and  (Q4+(n)}  are  asymptotically,  identically  distributed.  Thus,  if  one  does  not 
count  the  feedback  customer  in  the  queue  length  process,  the  queue  length  processes  defined  in 
Section  1.1  are  all  asymptoticaly,  identically  distributed. 

2.4   The  Ml  MIX  Case 

If  one  assumes  that  the  service  time  distribution  is 
Hit)  =  1  -  e-*',  t  >  0, 
some  further  clarification  is  possible  here.   From  the  results  of  Jackson  [8], 

ir'O')-    1-  —     —    ,  7  =  0,1,2,  .... 


From  (3)  and  (4)  one  obtains 


77  (0)  = 

T 

-  — 1 

<7M  j 

7T(/)  = 

l 

A_ 

[  x  )J 
\qfl) 

\p 

+ 

— 1 
p  1 

j=  1,2,  .... 

Comments  in  Section  2.3  explain  this  difference  between  tt(j)  and  ir'(j). 

3.   FLOW  PROCESSES 

To  further  clarify  the  problems  here,  it  is  useful  to  study  the  flow  processes  in  this  sys- 
tem. There  are  five  processes  of  interest:  the  arrival  process,  the  input  process,  the  output  pro- 
cess, the  departure  process,  and  the  feedback  process. 

There  have  been  some  questions  since  the  publication  of  the  Jackson  results  concerning 
the  interpretation  of  his  results  [2].  In  his  paper  Jackson  showed  that  for  his  networks  the  joint 
limiting  probability  for  the  vector  of  queue  lengths  at  each  server  could  be  factored  into  limit- 
ing probabilities  for  the  queue  length  at  each  server.  This  imples  that  the  queue  lengths  are 
independent  in  the  limit.    Furthermore,  the  marginal  limiting  probabilities  were  found  to  be 


640  R  L    DISNEY.  DC.  MCNICKLE  AND  B.  SIMON 

precisely  those  of  an  M/M/l  queue.  Burke  (2],  has  argued  that  the  Jackson  results  are  surpris- 
ing. Burke's  argument  is  based  on  showing  that  the  input  to  a  single  server  queue  with  feed- 
back is  not  Poisson  because  the  interinput  times  (our  [T'n  -  T'„-\l)  are  not  exponentially  distri- 
buted.   [2]  gives  the  precise  result 

Pr\r„  -  r;_,  <  /}  =  1  -  qP  ~  X  e~kl P±-e-»',  t  >  0. 

fJL  —  k  (X  —  A 

In  this  section  we  will  study  some  of  the  flows  in  this  network  and  show  indeed  that  the 
Jackson  results  are  surprising. 

3.1  Departures 

The  departure  process  {t„}  can  be  studied  as  in  Disney,  Farrell,  deMorais  [5]  upon  using 
the  mapping  in  Section  2.1.  Thus  we  know  that  whenever  {Sn}  is  a  renewal  process  with 
exponential  distribution  this  departure  process  is  a  renewal  process,  and  is  a  Poisson  process. 
This  is  the  Jackson  case.  So  we  conclude  that  the  departure  process  from  the  Jackson  network 
is  a  Poisson  process. 

From  the  results  of  Section  2.1  it  would  seem  possible  that  the  departure  process  is  Pois- 
son even  if  S„  is  not  exponentially  distributed.  The  result  that  is  needed  for  the  results  of  [5] 
to  follow  is  that  5^  be  exponentially  distributed  (since  it  is  known  that  {S'„}  is  a  sequence  of 
mutually  independent,  identically  distributed,  random  variables). 

LEMMA  1:  The  departure  process  from  the  M/G/l  queue  with  feedback  is  a  renewal  pro- 
cess if  and  only  if  S„  is  exponentially  distributed  for  every  n.  In  that  case  the  departure  process 
is  Poisson. 

PROOF:  From  Section  2.1  we  have  G*is),  the  Laplace-Stieltjes  transform  of  the  distribu- 
tion functions  of  S{,  is  given  by 

/-«/   ^  qH*(s) 

G  (5)  =   i — u*t  \  ■ 
1  -  pH*(s) 

From  [5],  when  the  queue  capacity  is  infinite  the  departure  process  will  be  a  renewal  pro- 
cess if  and  only  if  S'n  is  exponentially  distributed  with  parameter  a,  and  will  be  Poisson  in  that 
case.   But  this  implies  that  H*(s)  must  satisfy 

alia  +  s)  =  qH*is)/[\-  pH*is)]. 

The  only  solution  here  is 

HHs)  =  -^- 
aj q  +  s 

which  implies  Hit)  is  exponential.   □ 

3.2  Outputs  and  Inputs 

From  Section  2.2  it  is  clear  that  the  output  process  is  a  Markov-renewal  process  whose 
distributions  are  given  by  A  iij.x).   From  these,  the  following  results  are  obtained. 

THEOREM  2:  The  output  process  [T„  -  r„_,}  is  a  renewal  process  if  and  only  if  q  =  1 
and  Hit)  =  1  -  e^'. 


M/GA  QUEUE  WITH  INSTANTANEOUS  BERNOULLI  FEEDBACK  641 

PROOF:  If  q  =  1  and  Hit)  =  1  -  e_M',  the  output  process  and  departure  process  are 
identical  processes.  Furthermore,  the  processes  are  both  departure  processes  from  a  M/M/l 
queue  without  feedback.  From  [5]  we  have  that  this  departure  process  is  a  Poisson  process  and 
"if  follows.  To  prove  "only  if  we  consider  the  contrapositive  statement  and  assume  q  ^  1. 
(The  other  side  of  the  contrapositive  would  have  Hit)  ^  1  —  e-M'.  But  then  "only  if  follows 
trivially  from  [5].  Thus,  we  need  only  consider  the  case  of  q  ^  1.)  Equations  (3.1)  and  (3.2) 
in  [5]  can  be  modified  in  such  a  way  that  one  can  show  that  if  q  ^  1,  there  is  no  solution  to 
both  of  those  equations  simultaneously.  Then  using  the  same  arguments  as  in  [5]  one  has  that 
{T„  -  Tn_x]  is  not  a  renewal  process  and  therefore  "only  if  is  proven.   □ 

To  be  more  specific,  Theorem  2  can  be  particularized  as 

COROLLARY  1:  The  output  process  [Tn  —  Tn_:}  for  the  M/M/l  queue  is  a  Poisson  pro- 
cess if  and  only  if  q  =  1.  One  can  prove  this  result  (in  fact  it  is  obvious)  directly  from 
Theorem  2.  The  following  is  an  alternate  proof  that  exposes  a  bit  more  of  the  properties  of 
these  systems.   Again  we  use  a  contrapositive  proof  for  "only  if. 

PROOF:   Define 

F(x)=  Pr{T„-  r„_,  <  x). 

F(x)  =  77 AU where  (/is  a  column  vector  all  of  whose  elements  are  1,  n  is  the  vector  of  limit- 
ing probabilities  given  in  Section  2.4  for  (Q3+  in)}  and  A  is  the  matrix  of  A  (i,j,x).  Then  from 
Theorem  1  one  obtains  after  some  algebraic  manipulations: 

(5)  Fix)  =  \q  -  -\   C  [1  -  e-^-y']  dHiy)  +  \p  +  ~\  Hix) 

M  I      °  Ml 

for  any  M/G/l  queue  with  instantaneous,  Bernoulli  feedback. 
For  Hiy)  =  1  -  e_w,  it  follows  that 

(6)  Fix)  =  1  -  qfX~k  e~Kx  -  -££—  e~»\  x  >  0. 

/JL   —  \  /J,   —  A. 

Thus,  single  intervals  are  not  exponentially  distributed  and  the  output  process  is  not  a  Poisson 
process  if  q  ^  1.  On  the  other  hand  if  q  =  1,  then  we  fulfill  the  conditions  of  Theorem  2. 
Hence,  [Tn  —  Tn_x)  is  a  renewal  process.  But  from  (6)  this  renewal  process  has  exponentially 
distributed  intervals  and  thus  is  a  Poisson  process.  □ 

Formula  (6)  was  previously  found  by  Burke  [2]  for  the  distribution  of  times  between 
inputs.   The  input  process  can  be  analyzed  as  follows: 

THEOREM  3:  If  Hix)  =  \  -  e~»\  the  process  {Qfin),  T'„  -  T'n_x)  is  a  Markov- 
renewal  process  with  kernel 

YiiJ.x)  =  PrlQi  in)  =  j,  T'„  -  T'n_x  <  x\Qi  (n  -  1)  -  /] 

given  by 


YiiJ,x)  - 


0;    ;  > 

i  +  1, 

J>- 

-  qe~ 

<*)q 

dHu+x)is); 

J  =  0,  i 

>  0, 

X'H 

k   +  fX 

u- 

e-(\+n)(x-s 

>)  +  p\  Q' 

~jdH( 

er?%\  - 

-  e-kx); 

j  = 

-  i  +  1, 

642  R  L    DISNEY.  DC    MCNICKLE  AND  B   SIMON 

where  dH("+])is)  =  ^5)"g~M'  ds. 
n\ 

PROOF:  Clearly,  if  j  >  /  +  1  then  YiiJ.x)  =  0.  If  j  =  0  then  Y(i,j,x)  is  the  probability 
that  the  /'  +  1  customers  in  line  all  depart  before  x  and  the  first  arrival  occurs  after  the  last 
departure,  but  before  x,  or,  the  first  /  -  1  customers  depart,  but  the  last  one  feeds  back  before 
x,  and  there  are  no  arrivals  while  this  is  happening. 

If  1  <  j  <  /  then  Y(iJ,x)  is  the  probability  that  i  —  j  +  1  customers  depart  before  x,  no 
arrivals  occur  during  this  time,  but  between  the  last  departure  and  x,  an  arrival  occurs  before  a 
departure;  or,  /  —  j  +  1  customers  are  served  before  x,  the  first  /  -  j  depart,  the  last  one  feeds 
back,  and  there  are  no  arrivals  while  this  is  happening. 

If  j  =  i  +  1  then  Y(i,j,x)  is  the  probability  that  there  is  an  arrival  before  x  and  no  depar- 
tures before  x.  Since  Y(i,j,x)  never  depends  on  {Q2  ik);  k  <  n  —  1}  or  \Tk\  k  <  n},  the 
process  [Qi  in),  T'n  —  T'n_\\  is  a  Markov-renewal  process.  □ 

Now,  if  Y(x)  is  the  matrix  whose  elements  are  Y(i,j,x),  tt  is  the  vector  of  probabilities 
found  in  (3)  and  U  is  a  vector  all  of  whose  elements  are  1  then  it  is  easy  to  see  that 

Fix)  =  Pr[T'n  -  r;,_,  <  x]  =  ttY(x)U 
and 

G(x,y)  =  Pr[r„  -  r;_,  <  x,   T'n+X  -  T'„  <  y]  =  tt  Y(x)  Y(y)U. 

where  Fix)  is  the  Fix)  given  by  (6).   Of  course,  if  { T'n  —  T'n_\)  is  to  be  a  renewal  process  then 
it  is  necessary  (but  not  sufficient)  that 

Gix,y)  =  Fix)Fiy). 

From  this  we  can  conclude: 

COROLLARY  2:  The  input  process  to  the  M/M/l  queue  with  instantaneous,  Bernoulli 
feedback  is  not  a  renewal  process  unless  q  =  \. 

PROOF:   If  q  =  1  then  the  input  process  is  just  the  arrival  process  which  is  Poisson. 

If  the  input  process  is  a  renewal  process  for  q  ^  1  then  it  must  be  true  that 
y/x\TrYix)U  =  Fix) 
Vxv;  7rYix)Yiy)U  =  Fix)Fiy)   where 
Fix)  is  given  by  (6)  and  (/is  a  column  of  l's.   Thus, 

Vxv;   U-^ry-    Yiy)U=0. 
Fix) 


y PH-     P-y-y  +  np-(jx+\)y\ 


Some  algebra  yields 

f 

7TYix)\ 

Fix)   J 

Yiy)U  = 

Fix] 

1-  (1 

-  e^*)  \ 

r 

Fix 

) 

\  1 

if?  *  1, 

Fix) 

-i\-e~ 

< 

e-nx  - 

fiq 

-  k 
-X 

-  A 

fj.  —  X 


e  *y  +  pe 


-MdS^k 


M/GA  QUEUE  WITH  INSTANTANEOUS  BERNOULLI  FEEDBACK  643 

Thus,  to  show  that  the  input  process  is  not  renewal,  it  suffices  to  show  that  for  some  y, 

(X  —  X  /JL  —  A 

The  third  term  of  the  Taylor  expansion  of  this  expression  is 

M      >i2  _     PH     M2    ,    P(p-  +X)  =  p\fji     ,  0 
fi-\    2        fi  -\    2  2  2 

so  (by  [1,  198]  for  instance),  it  cannot  be  identically  zero  unless  p  =  0  (i.e.,  g  =  1).  □ 

It  seems  obvious  that  the  arrival  process  and  feedback  process  are  not  independent 
processes.  One  can  show,  using  the  above  arguments: 

COROLLARY  3:*  Either  the  feedback  process  is  not  a  Poisson  process  or  the  arrival  pro- 
cess and  feedback  process  are  not  independent  processes  (or  both)  for  the  M/M/l  queue  with 
instantaneous,  Bernoulli  feedback. 

PROOF:  This  result  follows  immediately  from  Burke's  result  [2]  on  the  distribution  of 
the  interinput  arrivals.  For  if  the  feedback  process  is  both  independent  of  the  arrival  process 
and  is  itself  a  Poisson  process,  the  input  process  is  Poisson.  Thus,  Burke's  result  contradicts 
the  assumption.   □ 

3.3  Feedback 

The  feedback  stream  seems  to  be  quite  difficult  to  work  with.  From  the  previous  section 
we  know  that  it  is  either  not  independent  of  the  arrival  stream  or  not  a  Poisson  stream. 
Melamed  [9]  has  shown  that  this  feedback  process  is  not  a  Poisson  process.  We  conjecture 
further  that  it  is  not  independent  of  the  arrival  process.  If  so,  then  the  known  superposition 
theorems  cannot  be  used  to  study  feedbacks  in  terms  of  the  arrival,  feedback  and  input 
processes. 

Since  the  feedback  stream  is  the  result  of  applying  a  filter  to  the  Markov-renewal  output 
process,  it  is  itself  Markov-renewal  on  the  state  space  {1,2,  . . .}.  Even  in  the  M/M/l  case,  the 
form  of  the  feedback  stream  does  not  appear  to  reduce  to  that  of  any  simpler  process. 

4.   CONCLUSIONS 

There  are  several  conjectures  that  one  can  pose  concerning  networks  based  on  the  results 
of  this  paper.  First  with  respect  to  queue  length,  busy  period,  and  departure  processes,  if  one 
adopts  the  "outsiders"  view  [3]  these  processes  appear  to  be  those  generated  by  an  M/G/l 
queue  without  feedback.  However,  if  one  adopts  the  "insider"  view  the  queue  length  process 
does  not  appear  to  behave  as  seen  by  the  "outsider." 

Flow  processes  in  this  network  cannot  be  explained  by  appeal  to  superposition,  stretching, 
and  thinning  results  for  Poisson  processes.  The  requisite  independence  assumptions  both 
within  and  between  streams  of  events  are  not  satisfied  here.  Thus,  one  cannot  assume  that 
these  queues  which  act  "as  if  they  were  M/M/l  queues  to  the  "outsider"  are  M/M/l  queues 
to  the  "insider."  In  particular,  this  hints  at  the  possibility  that  in  these  networks,  even  as  simple 
as  Jackson  networks,  any  attempt  to  decompose  the  network  into  independent  M/M/l  queues 
is  doomed  to  failure.  This  decomposition  must  account  for  the  internal  flows  and  these  not 
only  appear  to  be  non  Poisson,  they  are  nonrenewal  and  are  dependent  on  each  other. 


*Melamed  [91  has  shown,  using  other  arguments,  that  the  feedback  stream  is  not  a  renewal  process. 


644  R.L    DISNEY.  DC.  MCNICKLE  AND  B.  SIMON 

In  [9],  it  is  shown  that  in  the  Jackson  structure,  the  flow  along  any  path  that  returns  a 
customer  to  a  node  that  he  has  previously  visited  is  not  only  not  Poisson,  it  is  not  renewal. 
Thus,  if  Jackson  networks  have  loops,  (direct  feedback  as  in  this  paper  being  the  simplest 
example),  they  cannot  be  decomposed  into  sub-networks  of  simple  M/M/l  queues.  In  particu- 
lar, these  results  imply  that  a  node-by-node  analysis  of  waiting  times  depending  as  they  do  on 
the  "insiders"  view  is  not  valid  if  one  simply  uses  M/M/l  results  at  each  server.  Takacs  [10] 
studies  the  waiting  time  problems  in  the  system  discussed  in  this  paper.  Disney  [6]  presents 
another  view  of  the  same  problem. 

ACKNOWLEDGMENTS 

We  would  like  to  thank  Dr.  Robert  Foley  and  Dr.  Robert  B.  Cooper  for  their  helpful  com- 
ments on  this  paper.  In  particular  Foley  first  brought  to  our  attention  that  {Q3+  (n)}  and 
{(?4+  («)}  are  asymptotically,  identically  distributed  if  one  does  not  include  the  feedback  custo- 
mers in  the  queue  length.   This  point  was  made  in  his  paper  [7]. 

REFERENCES 

[1]  Buck,  R.C.,  Advanced  Calculus,  (McGraw-Hill,  New  York,  1956). 

[2]  Burke,  P.J.,  "Proof  of  a  Conjecture  on  the  Interarrival-Time  Distribution  in  an  M/M/l 

Queue  with  Feedback,"  IEEE  Transactions  on  Communications,  575-576  (May  1976). 
[3]  Cooper,  R.B.,  Introduction  to  Queueing  Theory,  (MacMillan,  New  York,  1972). 
[4]  d' Avignon,  G.R..and  R.L.  Disney,  "Queues  with  Instantaneous  Feedback,"  Management 

Science,  24,  168-180  (1977). 
[5]  Disney,  R.L.,  R.L.  Farrell,  and  P.R.  deMorais,  "Characterization  of  M/G/l  Queues  with 

Renewal  Departure  Processes,"  Management  Science,  19,  1222-1228  (1973). 
[6]  Disney,  R.L.,  "Sojourn  Times  in  M/G/l  Queues  with  Instantaneous,  Bernoulli  Feedback," 

Proceedings  Conference  on  Point  Processes  and  Queueing  Theory,   Keszthely,   Hungary 

(September  1978). 
[7]  Foley,  R.D.,  "On  the  Output  Process  of  an  M/M/l  Queue  with  Feedback,"  Talk  given  at 

San  Francisco  Meeting,  Operations  Research  Society  of  America  (May  1977). 
[8]  Jackson,  J.R.,  "Networks  of  Waiting  Lines,"  Operations  Research,  5,  518-521  (1957). 
[9]   Melamed,  B.,  "Characterizations  of  Poisson  Traffic  Streams  in  Jackson  Queueing  Net- 
works," Advances  in  Applied  Probability,  11,  422-438  (1979). 
[10]  Takacs,  L.,  "A  Single  Server  Queue  with  Feedback,"  Bell  System  Technical  Journal,  42, 

509-519  (1963). 


AN  INVENTORY  MODEL  WITH  SEARCH 
FOR  BEST  ORDERING  PRICE* 


Woodward-Clyde  Consultants 
San  Francisco,  California 

ABSTRACT 

This  paper  presents  a  single-item  inventory  model  with  deterministic 
demand  where  the  buyer  is  allowed  to  search  for  the  most  favorable  price  be- 
fore deciding  on  the  order  quantity.  In  the  beginning  of  each  period,  a  sequen- 
tial random  sample  can  be  taken  from  a  known  distribution  and  there  is  a  fixed 
cost  per  search.  The  decision  maker  is  faced  with  the  task  of  deciding  when  to 
initiate  and  when  to  stop  the  search  process,  as  well  as  determining  the  optimal 
order  quantity  once  the  search  process  is  terminated.  The  objective  is  to 
minimize  total  expected  costs  while  satisfying  all  demands  on  time.  We 
demonstrate  that  a  set  of  critical  numbers  determine  the  optimal  stopping  and 
ordering  strategies.  We  present  recursive  expressions  yielding  the  critical 
numbers,  as  well  as  the  minimal  expected  cost  from  the  beginning  of  every 
period  to  the  end  of  the  horizon. 


1.   INTRODUCTION 

This  research  is  an  attempt  to  marry  some  aspects  of  search  theory  and  optimal  stopping 
with  inventory  theory.  Following  the  pioneering  work  of  Stigler  [11],  [12],  searching  for  the 
lowest  price  is  considered  a  basic  feature  of  economic  markets.  By  citing  examples  based  on 
real  data,  Stigler  [11]  asserted  that  prices  change  with  varying  frequency  in  all  markets,  and 
unless  a  market  is  completely  centralized,  the  buyer  will  not  know  for  certain  the  prices  that  the 
various  sellers  quote  at  any  given  time.  This  suggests  that  at  any  time  there  will  be  a  frequency 
distribution  of  the  prices  quoted  by  sellers.  If  the  dispersion  of  price  quotations  by  sellers  is 
large  compared  to  the  cost  of  search,  it  will  pay— on  average— to  obtain  price  quotations  from 
several  sellers  before  taking  an  "action."  The  vast  literature  on  search  theory  (a  survey  of 
which  can  be  found  in  Lippman  and  McCall  [8],  DeGroot  [5],  and  Rothschild  [10])  is  con- 
cerned with  rules  that  the  searchers  should  follow  when  the  "action"  is  accepting  or  rejecting  a 
price.  Once  the  price  has  been  accepted,  the  decision  process  terminates.  In  many  dynamic 
models,  the  action  is  more  complicated.  In  inventory  models,  for  example,  the  decision  not 
only  involves  accepting  or  rejecting  an  ordering  price  but  how  much  to  order,  an  action  which 
will  affect  the  search  and  ordering  policies  in  future  periods.  In  this  paper  we  study  such  a 
model.  We  seek  the  best  search  and  ordering  policies  for  a  simple  dynamic  inventory  problem 
with  deterministic  demands  where,  in  the  beginning  of  each  period,  the  purchaser  can  search 
for  the  lowest  price  before  placing  an  order. 

*This  research  was  partially  supported  by  the  National  Science  Foundation  through  Grant  NSF  ENG74-13494  and  the 
Air  Force  Office  of  Scientific  Research  (AFOSR  72-2349C). 


645 


646  K   GOLABI 

Classical  optimal  search  considers  the  following  problem:  A  purchaser  can  take  a  sequen- 
tial random  sample  X],X2,  ...  from  a  continuous  distribution  with  a  known  distribution  func- 
tion F.  There  is  a  fixed  cost  s  per  observation.  Suppose  that  if  the  decision  maker  stops  the 
sampling  (search)  process  after  the  values  X\  —  x\,  X2  =  x2,  . . .,  Xn  =  x„  have  been  observed, 
his  cost  is  x„  +  sn.  Hence,  the  problem  is  to  find  a  stopping  rule  which  minimizes  E{XN  +  sN) 
where  N  indicates  the  random  number  of  observations  that  are  taken  under  a  specified  stopping 
rule.  It  can  be  shown  that,  whether  sampling  is  with  or  without  recall,  the  optimal  stopping 
rule  is  characterized  by  a  unique  critical  number  v*  (usually  called  the  reservation  price)  so  that 
an  optimal  sampling  rule  is  to  continue  sampling  whenever  an  observed  value  exceeds  v*  and  to 
stop  the  process  as  soon  as  some  observed  value  does  not  exceed  v*.  Various  versions  of  this 
problem  have  been  studied  by  MacQueen  and  Miller  [9],  Derman  and  Sacks  [6]  and  Chow  and 
Robbins  [2],  [3]  among  others. 

The  above  search  model  can  be  visualized  as  a  one  period  purchasing  problem  in  which 
one  unit  of  some  commodity  has  to  be  purchased  at  the  beginning  of  the  period.  Now  consider 
a  dynamic  multiperiod  version  of  this  problem  where  a  demand  of  one  unit  has  to  be  satisfied 
in  each  period  and  inventory  holding  cost  is  charged  for  items  held  over  for  use  in  subsequent 
periods.  As  in  the  classical  search  problem,  in  the  beginning  of  each  period  a  sequential  ran- 
dom sample  XhX2,  ...  can  be  taken  from  a  distribution  with  known  distribution  function  F, 
but  the  decision  process  is  not  terminated  as  soon  as  an  acceptable  value  is  observed.  The  deci- 
sion maker  is  faced  with  the  task  of  deciding  how  much  to  order  so  as  to  minimize  total 
expected  costs  while  satisfying  all  demands  on  time.  When  the  inventory  level  is  sufficient  to 
satisfy  the  immediate  demand,  he  has  also  the  burden  of  deciding  whether  to  initiate  search  at 
all.  This  multiperiod  rriodel  is  the  subject  of  our  study  in  this  paper. 

In  Section  2,  we  present  the  model.  In  Section  3,  we  give  the  optimal  search  policy  and 
in  Section  4,  the  optimal  ordering  policy.  We  show  the  intuitive  result  that  an  optimal  strategy 
prescribes  that  search  should  be  initiated  only  when  the  inventory  level  is  zero.  Furthermore, 
we  show  that  the  reservation  price  property  of  the  classical  search  problem  still  holds.  That  is, 
when  the  inventory  level  is  zero  (and  therefore  search  has  to  be  initiated)  and  n  periods  remain 
to  the  end  of  the  problem,  there  exists  a  reservation  price  an  such  that  a  price  should  be 
accepted  if  it  does  not  exceed  a„  and  rejected  otherwise.  In  Section  4,  we  show  that  once  a 
price  has  been  accepted,  a  finite  number  of  critical  numbers  specify  the  optimal  strategy:  The 
critical  numbers  divide  the  interval  [O.aJ  into  segments  so  that  the  interval  in  which  the 
accepted  price  falls  determines  the  optimal  order  quantity.  We  give  recursive  expressions  which 
yield  a„  as  well  as  the  minimal  expected  cost  for  any  period  to  the  end  of  the  horizon.  We  will 
also  obtain  expressions  describing  the  critical  numbers  when  the  holding  cost  function  is  con- 
vex. 

2.   THE  MODEL 

Consider  a  multi-period  single-item  inventory  model  in  which  a  demand  of  one  unit  has 
to  be  satisfied  in  the  beginning  of  each  period  and  an  inventory  holding  cost  is  charged.  In 
each  period,  a  sequential  random  sample  X\,X2,  ...  of  ordering  prices  can  be  generated  from  a 
continuous  distribution  with  known  cumulative  distribution  function  F(-),  E(X\)  <  «>,  and  the 
Xj's  are  mutually  independent.  The  cost  of  generating  each  random  price  is  s  and  there  is  no 
limit  on  the  number  of  observations  which  can  be  made  in  each  period.  After  receiving  a  price, 
the  decision  maker  has  to  decide  whether  to  accept  that  price  or  generate  another  offer.  If  he 
accepts  the  offered  price,  he  is  faced  with  the  decision  of  how  much  to  order.  When  the  inven- 
tory level  is  sufficient  to  satisfy  the  immediate  demand,  he  also  has  to  decide  whether  to  initiate 
search  at  all.   The  objective  is  to  minimize  the  total  expected  costs. 


INVENTORY  MODEL  WITH  SEARCH  647 

We  assume  that  the  length  N  of  the  planning  horizon  is  finite,  initial  inventory  is  zero, 
backlogging  of  demand  is  not  allowed,  the  cost  of  holding  z  units  for  one  period,  h(z),  is  non- 
decreasing  in  z  and  h  (0)  =  0,  the  purchasing  cost  is  linear  in  the  quantity  ordered,  and  only 
integer  quantities  can  be  ordered.  We  also  assume  prices  that  are  not  accepted  immediately  are 
lost;  in  view  of  our  results,  sampling  with  recall  (of  prices  in  the  same  period)  extends  no  addi- 
tional advantage  over  sampling  without  recall,  and  hence  would  not  affect  the  search  policy. 
Note  that  when  TV  =  1,  this  model  reduces  to  the  classical  search  problem. 

Let  n  be  the  number  of  periods  remaining  until  the  end  of  the  horizon,  z  the  inventory  on 
hand  with  n  periods  remaining  and  x  the  last  price  received.  In  each  period,  our  state  space 
consists  of  numbers  (z)  and  pairs  (z,x)  corresponding  respectively  to  the  state  of  the  system 
before  a  search  is  placed  and  the  state  when  a  search  has  been  placed  and  an  offer  x  has  been 
received.  A  policy  for  period  n  prescribes  a  search  decision  for  state  (z),  and  a  reject-accept 
and  ordering  decision  for  state  (z,x).  We  assume  that  for  each  period  an  optimal  policy  exists. 
Moreover,  we  restrict  our  attention  to  history-independent  policies;  that  is,  once  the  price  x  has 
been  rejected,  we  are  in  the  same  position  as  having  not  placed  the  search  at  all.  Schematically, 
remembering  that  demand  in  each  period  equals  one,  the  period-state  pairs  correspond  to  each 
other  as  follows: 

Forz  ^  1: 


(nz)      Search  ,  (nzx)AccQpi  x  and  order  an  amount  a     (n-\  z+a-\) 

^    ^(A7-1(Z-1) 

and 


/ -i  qx     aearcn         ^  /■    ^    \ Accept  x  and  order  an  amount  a     /-^-i  a_jv 

3.  OPTIMAL  SEARCH  POLICY 

In  this  section,  we  present  the  optimal  search  policy.  We  show  that  search  should  only  be 
initiated  when  the  inventory  level  is  zero,  and  prove  that  in  each  period  a  single  reservation 
price  determines  the  stopping  rule.  We  also  give  a  recursive  expression  which  describes  the 
sequence  of  reservation  prices. 

To  being,  define 

Vn(z,x)  =  the  minimal  (conditional)  expected  cost  during  the  last  n  periods  when  the 
inventory  level  with  n  periods  remaining  is  z  and  the  last  price  offered  is  x. 

v„(z)  =  the  minimal  expected  cost  during  the  last  n  periods  before  the  decision  to 
search  for  an  offer  is  made,  and  when  the  inventory  level  with  n  periods 
remaining  is  z. 


un{z)  =  the  minimal  expected  cost  during  the  last  n  periods  after  the  decision  to 
search  for  an  offer  in  this  period  has  been  made,  and  when  the  inventory 
level  with  n  periods  remaining  is  z. 

w„(z)  =  the  minimal  expected  cost  during  the  last  n  periods  after  the  decision  not  to 
search  for  an  offer  in  this  period  has  been  made,  and  when  the  inventory 
level  with  n  periods  remaining  is  z,  z  ^  1 . 

H{z)       =  the  total  holding  cost  of  z  units  to  be  used  in  z  consecutive  periods. 
Hence,  we  will  have  the  following  relationships: 

[ax  +  h(z  +  a  -  1)  +  v„_,(z  +  a  -  1)]}, 
1),  z  ^  1, 


(1) 

v„(z)  =  min[«„(z),  wn(z)] 

(2) 

V„(z,x)  =  min{v„(z),         min 

a€[1.2 1 

(3) 

u„(z)  =  s  +  Ex[V„{z,x)], 

(4) 

w„{z)  =  hiz  -  1)  +  v„_,(z  - 

and 

(5) 

Define 

//(z)=£//(z-  /)=  £  //(/ 

(6a) 

In[x,a]  =  ax  +  h(a  -  l)  +  v, 

and 

(6b) 

I  Ax)  =       min      I.,(x,a), 

a€[1.2 »] 

-l(fl-  1) 


and  let  an(x)  be  the  minimizing  value  of  a  in  (6a),  that  is, 

(6c)  In(x)  =  lAx,an(x)}. 

The  quantity  In(x)  is  the  minimal  expected  cost  attainable  during  the  last  n  periods  when  the 
inventory  level  with  n  periods  remaining  is  zero  and  it  has  been  decided  to  accept  x,  the  last 
price  offered. 

At  this  point  it  is  natural  to  ask  whether  when  n  periods  remain,  there  exists  a  single  criti- 
cal price  an  which  dictates  the  acceptance  or  rejection  of  a  price  x  when  the  inventory  level  is 
zero.  In  other  words,  is  there  ana„  such  that  it  is  optimal  to  accept  the  price  x  (and  order  a 
positive  amount)  if  jc  <  a„  and  to  continue  the  search  if  x  >  <xn.  That  this  is  indeed  the  case, 
is  verified  in  Theorem  1. 

Define 

(7)  q„e/;'[v„(0)], 

and  the  sequence  {An}"=0  by  the  following  recursion: 

(8a)  ^o=0 

and 

(8b)  A„F(a„)  =  5  +  C"      min     [ax  +  H(a)  +  An_a]dF(x)   for  n  >  1. 

J0      a6[i,2 „] 


INVENTORY  MODEL  WITH  SEARCH 


649 


We  will  show  later  that  an  exists  and  that  A„  equals  v„(0),  so  that  an  =  I~HAn).  These  pro- 
perties are  exploited  to  verify  that  an  optimal  policy  prescribes  that  search  be  initiated,  and  ord- 
ers be  placed,  only  when  the  inventory  level  is  zero.  Furthermore,  we  will  show  that  if  the  set 
of  prices  at  which  it  is  optimal  to  order  one  unit  is  nonempty,  a„  =  An  —  An-\  so  that  Equation 
(8b)  can  be  written  as 


(9) 


A„FU„  -  Att-0  =  s  - 


[ax  +  H(a)  +An_a]dF(x), 


enabling  us  to  obtain  the  minimal  expected  cost  from  the  beginning  of  any  period  to  the  end  of 
the  horizon  by  finding  [Aj}j!Lo,  the  unique  set  of  solutions  to  Equation  (9). 

THEOREM  1:  If  thexV  inventory  level  with  n  periods  remaining  is  zero,  it  is  optimal  to 
continue  the  search  if  x,  the  last  price  offered,  is  greater  than  an  and  accept  the  price  if 
x  <  a„,  where  n  =  1,  2,  . . .  ,  N. 

PROOF:  Clearly,  I„{x,a)  is  continuous  in  x  for  each  n  and  a,  and  therefore,  In(x)  is  a 
continuous  function  of  x.  Furthermore,  for  all  positive  numbers  e, 

I„{x  +  e)  =  In[x  +  e,  an(x  +  c)]  >  In[x,an{x  +  e)]  >  I„[x,a„(x)]  =  /„(*), 

and  hence  In{x)   is  strictly  increasing  in  x.    Let  a„(v)   be  such  that  I„[an(y)]  =  v,  i.e., 
a„(y)  =  /„-1(y).  Since 


v„(0)  ^  v„_!(0)  ^      min 

fl€[l,2,..., 


[h (a-  l)  +  v„_,(a-  l)]  =  /„(0), 


it  follows  that  an  =  a„[v„(0)]  exists  and,  as  In(x)  is  strictly  increasing  in  x,  it  is  unique  (see 
Figure  1).  The  first  inequality  of  the  above  expression  follows  from  the  fact  that  for  the  n  —  1 
period  problem  we  can  always  follow  the  optimal  policy  for  the  n  period  problem,  so  that  at 
each  stage  m,  n  —  l^w^l,  we  would  adopt  the  action  prescribed  by  the  n  period  optimal 
policy  for  stage  m  +  1.  Thus,  the  expected  cost  for  the  n  -  1  period  problem  under  this  pol- 
icy, v^_!(0),  would  be  equal  to  the  expected  cost  of  the  first  n  -  1  periods  of  the  n  period  prob- 
lem, and  hence  v^^O)  <  v„(0).  Since  v„_!(0)  <  v^^O),  it  follows  that  v„(0)  is  nondecreas- 
ing  in  n. 


W 


vn(0) 
ln(0) 


<*n 


650  K.  GOLABI 

From  (2)  and  (6)  we  have 

K„(0,x)  =  min  [v„(0),  /„(*)]. 

If  x  <  a„,  then  In(x)  <  v„(0)  so  that  V„(0,x)  =  In(x)  and  search  terminates.  If  x  >  a„, 
then  I„(x)  >  v„(0)  so  that  Vn(0,x)  =  v„(0),  in  which  case  it  is  optimal  to  continue  the  search. 

Q.E.D. 

Thus,  when  the  inventory  level  is  zero,  a  single  critical  number  determines  whether  a 
price  should  be  accepted.  We  are  also  interested  in  finding  the  optimal  strategy  when  the 
inventory  level  is  positive.  It  seems  intuitive  that  if  the  immediate  demand  can  be  satisfied  by 
the  current  inventory,  it  would  be  best  to  postpone  the  search— since  it  is  possible  to  incur  the 
same  amount  of  expected  search  cost  in  a  later  period  while  saving  on  the  holding  cost.  The 
next  result,  the  proof  of  which  is  given  in  the  Appendix,  verifies  this  observation.  In  addition, 
it  shows  that  the  expected  cost  from  any  period  k  in  which  the  inventory  level  is  zero  to  the 
end  of  the  horizon  equals  Ak.  Thus,  the  expected  cost  from  any  period  can  be  determined  by 
computing  the  sequence  [A„)  from  Equation  (8b). 

THEOREM  2.  Under  the  assumptions  of  the  model,  for  all  k,  k  =  1, 2,  ...  ,  N, 

(a)   vk(0)  =  Ak 

.  (b)   vfc(z)  =  H(z)  +  \k-2(0)         for  1  <  z  ^  k. 

Theorem  2  verifies  that  search  should  be  initiated  only  when  the  inventory  level  is  zero, 
and  Theorem  1  gives  a  rule  for  accepting  or  rejecting  an  offered  price  once  search  is  initiated. 
These  two  results  however,  do  not  completely  specify  the  optimal  strategy.  Given  that  an 
acceptable  price  is  received,  we  would  like  to  know  how  much  should  be  purchased  at  that 
price.  This  question  is  investigated  in  the  next  section. 

4.   OPTIMAL  ORDERING  POLICY 

In  this  section  we  present  the  optimal  ordering  policy  once  an  acceptable  price  has  been 
received.  In  Corollary  3  we  show  that  a  nonincreasing  sequence  of  critical  numbers  characterize 
the  optimal  order  quantity.  In  other  words,  once  a  price  is  received  that  is  less  than  the  reser- 
vation price  for  that  period,  the  interval  in  which  the  offered  price  falls  determines  the  quantity 
that  should  be  ordered  at  that  price.  In  Theorem  5  we  obtain  expressions  which  describe  these 
critical  numbers  when  the  holding  cost  function  is  convex. 

Before  presenting  the  next  result  we  note  that  when  n  periods  remain,  the  inventory  level 
is  zero,  and  an  acceptable  price  has  been  received,  the  optimal  order  quantity  is  equal  to  a„(x). 
To  see  this,  note  that 

K„(0,x)  =  min[(v„(0),  /„(*)] 

by  (2)  and  (6).  This  fact  coupled  with  Theorem  1  yields  Vn(0,x)  =  In(x)  whenever  x  <  a„. 
Finally,  'since 

(10)  /„W  =  /„b„W]=    min    [ax  +  h(a-l)  +  w„-i(a-l)]t 

it  follows  that  ordering  a„(x)  minimizes  the  expected  cost  attainable  during  the  last  n  periods 
when  the  inventory  level  is  zero  and  x  <  a„.  Note  also  that  by  Theorem  2(b),  Equation  (10) 
can  be  written  as 


INVENTORY  MODEL  WITH  SEARCH  651 

(11)  /„(*)=    min    [ax  +  H(a)  +  A„.a]. 

COROLLARY  3:  If  n  periods  remain,  the  inventory  level  is  zero,  and  an  acceptable  price 
has  been  received,  then  the  optimal  order  quantity  is  nonincreasing  in  the  price  offered,  i.e., 
a„(x')  <  a„(x)  whenever  x'  >  x,  n  =  1,2,  ...  ,  N.  Consequently,  a  nonincreasing  sequence 
of  critical  numbers  [Bj{n)}-L\  characterize  the  optimal  order  quantity.  Specifically,  it  is  optimal 
to  order  k  units  whenever  Bk(n)  <  x  <  Bk_x{n). 

PROOF:  From  (6c)  and  (11),  we  have 

I„(x)  =  I„bc,a„(x)]  =  an(x)  ■  x  +  H[an(x)]  +  An_an(x) 
<  In[x,an(x')]  =  aa(x')  •  x  +  H[a„(x')]  +  An^ajxV 
giving 

(12)  x[a„(x')  -  an{x)]  >  An_Qn{x)  -  A^w  +  H[a„(x)]  -  H[an(x')]. 
If  an{x')  >  an(x),  then  (12)  implies 

x'[an(x')  -  a„(x)]  >  An_anix)  -  ^.Crf  +  H[an(x)]  -  H[an(x')}, 
which  yields 

In(x')  =  an{x')  ■  x'  +  An^an(x')  +  H[a„(x')]  >  an(x)  •  x'  +  An_an{x) 
+  H[an(x)]  =  In[x',an(x)}, 
contradicting  the  fact  that  an(x')  is  optimal  when  x'  is  offered.  Q.E.D. 

Intuitively,  we  would  expect  that  when  an  offered  price  equals  the  critical  number  an,  we 
would  be  indifferent  between  ordering  one  unit  and  not  ordering  at  all.  If  this  were  indeed  the 
case,  the  expected  cost  when  the  price  is  rejected,  v„(0),  would  be  equal  to  a„  +  v„_!(0)  yield- 
ing a„  =  A„  —  An_\.  This  result  could  then  be  used  to  obtain  a  simple  expression  for  the 
Bk(nVs  when  h  (•)  is  convex.  As  we  will  show  in  Lemma  4,  the  above  result  holds  if  the  set  of 
prices  at  which  it  is  optimal  to  order  one  unit  is  nonempty.  Unfortunately,  as  seen  from  the 
following  example,  this  is  not  always  the  case. 

EXAMPLE  1.   Let  n  =  5,  5  =  5,  h  (z)  =  0  for  all  z  and  the  price  distribution  be  such  that 

P(X  =  2)  =  1  -  e,    and    P(a  <  X  <  b)    =^-{b-  a)    for    0  <  a  <  b  <  4,    where    2    is 

4 
excluded  from  all  intervals  and  e  is  an  arbitrary  small  number.   Suppose  the  offered  price  in  the 
beginning  of  the  fifth  period  is  3. 

The  expected  cost  of  rejecting  the  offered  price  is  (approximately) 

5  +  2x  5=  15, 

as  one  would  pay  the  search  cost  of  5  and  almost  definitely  receive  the  price  of  2,  at  which  one 
would  order  5  units.   However,  the  expected  cost  of  ordering  /  units,  i  <  4,  is  (approximately) 

3/  +  5  +  2(5-  /)=  15  +  /, 

while  the  cost  of  ordering  5  units  is  15.   Hence,  we  would  be  indifferent  between  not  ordering 
and  ordering  at  x  =  3,  which  implies  that  a5  =  3. 


Since  at  x  =  3  we  order  5  units,  any  price  above  3  is  rejected,  and  the  optimal  order  quan- 
tity a„(x)  is  nonincreasing  in  x,  it  follows  that  [x  :a„(x)  =  1]  is  empty. 

LEMMA  4:  If  [x  :  a„(x)  =  1]  is  nonempty,  then  an  =  An  —  A„_x. 

PROOF:  Let  x  be  the  largest  x  such  that  an{x)  =  1.  By  Theorem  1,  a „  is  the  highest 
price  at  which  it  is  optimal  to  order  a  positive  quantity.  Therefore,  3c  <  a„.  Consequently,  we 
can  conclude  from  Corollary  3  that  an(an)  ^  1,  but  an(a„)  is  positive  so  that  a„{a„)  =  1. 
From  Theorem  2  and  Equations  (7)  and  (11),  we  have 


=  (a„(a„)  •  a„  +  H[an(an)]  +  4,-a>n))  =  <*n  +  H{\)  +  A„.h 
which  yields  a„  =  An  -  A„_x.  Q.E.D. 

Whereas  we  cannot  determine  in  advance  the  conditions  under  which  Lemma  4  would 
hold,  we  can  proceed  by  assuming  that  the  lemma  holds,  and  determine  the  sequence  [4„}!Lp 
that  satisfies  Equation  (9).  We  then  can  obtain  {<*„}  from  an  =  I~x{An).  If  {«„}  and  \An)  also 
satisfy  Equation  (8b),  by  uniqueness  of  the  solution,  an  is  indeed  equal  to  A„  -  A„_x. 

It  is  interesting  to  note  that  contrary  to  what  one  might  expect,  an  is  not  monotone  in  n. 
Before  Theorem  5,  we  give  examples  where  an  is  not  monotone  irrespective  of  whether 
[x  :  an  (x)  =  1]  is  empty  or  not 

EXAMPLE  2:  (a)  Consider  again  Example  1.  Since  we  would  almost  definitely  receive 
the  price  of  2  after  the  first  search,  we  have 

v„(0)  =  s+    min    [ax  +  H(a)  +  vfl_fl(0)]. 

Thus, 

Vl(0)  =  5  +  2=7 

v2(0)  =  5  + min  (2  +  7,4)  =  9. 
From  v„(0)  =  I„(a„),  we  have  ax  =  vj(0)  =  7  and 

9=  min  (a2  +  7,2  a2) 
yielding  a2  =  4.5.   As  shown  earlier,  a5  =  3.  Therefore,  an  is  not  monotone  in  n. 

(b)  We  note  that  an  is  not  necessarily  monotone  even  if  [x  :an(x)  =  1]  is  nonempty.  Con- 
sider the  case  where  the  price  distribution  is  the  same  as  Example  1 .  However,  there  is  a  hold- 
ing cost  of  1  per  unit  per  period  and  5  =  2.  Then,  H{\)  —  0,  H(2)  =  1,  HO)  =  3  and 
7/(4)  -  6  and 

Vl(0)  =  2  +  2=4 

v2(0)  =  2  + min  [4+  1,2  +  4]  =  7 

v3(0)  =  2  +  min[6  +  3,4+  1  +  4,2  +  7]=  11 

v4(0)  =  2  + min  [8  +  6,6  +  3  +  4,4+  1  +  7,2+  11]=  14. 


INVENTORY  MODEL  WITH  SEARCH  653 

From  v„(0)  =  /„(«„),  we  have 
4-ai 

7  =  min  [2a2  +  l.«2  +  4] 

11  =  min  [3a3  +  3,2a3  +  1  +  4,a3  +  7] 

14=  min  [4a  4  +  6,  3a4  +  3  +  4,2a4  +  1  +  7,a4  +  11] 

yielding 

a\  =  4,  a2  =  3,  a3  =4,  a4  =  3. 

Note  that  in  this  example,  an{an)  =  1  for  1  <  n  <  4  and  the  condition  for  Lemma  4  holds.   It 
can  be  easily  verified  that  a„  =  v„(0)  -  v^CO)  for  all  1  <  n  <  4. 

THEOREM  5:  If  the  condition  for  Lemma  4  holds  and  if  the  holding  cost  function  h(-) 
is  convex,  then 

(13)  Bk(n)  =  an_k-  h(k),    where  1  <  k  <  n. 

PROOF:  We  have  to  show  that 

(a)  The  RHS  of  (13)  is  nonincreasing  in  k. 

(b)  It  is  optimal  to  order  k  units  if  x,  the  price  offered,  satisfies 

(14)  otn-k-  h(k)  <  x  <  «„_<*_,)-  h(k  -  1). 

To  show  (a) ,  we  note  that 

An_k+X  =  v„_£+1(0)  =  /„_*+,  (a„_fc+1)  =        min        [aan-k+x 

+  H(a)  +  4,-fc+i-al  ^  2a„_^+1  +  7/(2)  +  ^„_^_! 

=  lUn-k+i  -  A„.k)  +  h(l)  +  A„-k-V, 

where  the  first  equality  follows  from  Theorem  2,  the  second  from  (7),  the  third  from  (11)  and 
the  last  from  Lemma  4.  Thus,  by  convexity  of  h  (•), 

h(k)-  h(k-  \)  >  h(l)  >  An_k  -  An-k.x  -  An_k+X  +  An_k 

=  <*n-k  -  <*n-k+l 

and,  therefore,  (a)  is  true. 

To  show  (b),  suppose  x  is  such  that  (14)  holds.  We  show  that  I„(x,k  —  j)  < 
I„(x,k  -  j  -  I)  for  each  j  ^  0,  and  therefore  ordering  k  units  is  at  least  as  good  as  ordering 
any  amount  less  than  k.  Suppose  In(x,k  -  j)  >  I„(x,k  -  j  -  1).  Then 

(k-j-  l)x  +  ^-(jfc-y-1)  +  H(k  -j-\)<(k-  j)X  +  An_(k_j)  +  H(k  -  j) 

which  yields 

x  >  Amfr-j-Q  -  An_(k_j)  -  h(k  -  j  -  1) 

=  an-(k-j-\)  -  h (k  -  j  -  1)  ^  a„-0fc-i)  -  h(k  -  1), 
where  the  last  inequality  follows  from  (a).  This  contradicts  the  right  inequality  of  (14).  There- 
fore, /„ (x,k  -  j)  <  In(x,k-  j  -  1) . 


I„(x,k  +  j)  <  I„(x,k+  J  +  1)  for  each  j  >  0  by  a  similar  proof. 
Hence,  it  is  optimal  to  order  k  units  whenever  (14)  holds.  Q.E.D. 

5.    REMARKS 

The  purpose  of  this  study  has  been  to  investigate  optimal  search  policies  in  the  context  of 
a  sequential  model.  The  underlying  inventory  model  has  been  chosen  as  a  rather  simple  one. 
There  are  no  setup  costs  involved  and  the  demand  equals  one  unit  in  each  period.  It  would  be 
interesting  to  investigate  more  general  problems.  We  suspect  that  both  the  reservation  price 
property  of  Theorem  1  and  the  Wagner- Whitin  [13]  type  result  of  Theorem  2  (order  only  when 
current  inventory  level  is  zero)  would  still  hold  for  models  with  setup  costs  and  arbitrary  deter- 
ministic demands.  The  optimal  policy  would  be  a  function  of  setup  costs  as  well  as  the  holding 
cost  and  price  distribution.  The  results  should  also  hold  when  the  price  distributions  are  non- 
stationary.  Given  that  the  initial  inventory  is  zero,  the  ordering  policy  will  be  such  that  there  is 
no  inventory  in  the  beginning  of  periods  with  favorable  price  distributions. 

Another  interesting  extension  is  the  case  wherein  the  search  process  is  adaptive.  The 
searcher  does  not  know  the  exact  distribution  of  price;  the  price  offer  is  used  not  only  as  an 
opportunity  to  order  at  that  price  but  also  as  a  piece  of  information  to  update  the  prior  distribu- 
tion. When  the  distribution  of  prices  is  not  known  exactly,  the  form  of  the  optimal  policy  is 
not  obvious.  As  Rothschild  [10]  points  out,  the  reservation  price  property  of  Theorem  1  would 
not  necessarily  hold  even  for  a  one  period  problem.  Rothschild  presents  the  following  example. 
Suppose  there  are  three  prices,  $1,  $2,  and  $3,  and  that  the  cost  of  search  is  $0.01.  Prior 
beliefs  admit  the  possibility  of  only  two  distributions  of  prices.  Either  all  prices  are  $3  or  they 
are  distributed  between  $1  and  $2  in  the  proportions  99  to  1.  A  man  with  these  beliefs  should 
accept  a  price  of  $3  (as  this  is  a  signal  that  no  lower  prices  are  to  be  had)  and  reject  a  quote  of 
$2  (which  indicates  that  the  likelihood  that  a  much  better  price  will  be  observed  on  another 
draw  is  high). 

However,  when  the  distribution  is  a  member  of  certain  families  of  distributions  but  has 
one  or  more  unknown  parameters,  Rothschild  [10],  DeGroot  [5]  and  Albright  [1]  have  shown 
that  the  reservation  price  property  holds  for  the  one-period  problem.  We  conjecture  that  when 
the  distribution  of  price  is  stationary  but  is  not  known  exactly,  search  should  be  initiated  only 
when  the  inventory  level  is  zero.  If  this  is  the  case  and  the  distribution  belongs  to  one  of  the 
families  of  distributions  studied  by  Rothschild  [10]  and  Albright  [1],  then  the  reservation  price 
property  as  well  as  the  ordering  policy  presented  in  Section  4  should  still  hold. 

ACKNOWLEDGMENTS 

This  paper  is  essentially  Chapter  3  of  the  author's  dissertation  (1976)  at  the  University  of 
California,  Los  Angeles.  The  author  expresses  his  appreciation  to  Professor  Steven  Lippman 
for  his  guidance  and  encouragement.  He  also  appreciates  several  helpful  comments  by  Profes- 
sor Sheldon  Ross  and  the  referee. 

BIBLIOGRAPHY 

[1]  Albright,  C.S.,  "A  Bayesian  Approach  to  a  Generalized  House  Selling  Problem,"  Manage- 
ment Science  24,  432-440  (1977). 

[2]  Chow,  Y.S.  and  H.  Robbins,  "A  Martingale  System  Theorem  and  Applications,"  Proceedings 
of  the  4th  Berkeley  Symposium  on  Mathematical  Statistics  and  Probability,  University  of  Cali- 
fornia Press,  Berkeley,  California  (1961). 


INVENTORY  MODEL  WITH  SEARCH  655 

[3]  Chow,  Y.S.  and  H.  Robbins,  "On  Values  Associated  with  a  Stochastic  Sequence,"  Proceed- 
ings of  the  5th  Berkeley  Symposium  on  Mathematical  Statistics  and  Probability,  University 
of  California  Press,  Berkeley,  California  (1967). 

[4]  DeGroot,  M.H.,  "Some  Problems  of  Optimal  Stopping,"  Journal  of  the  Royal  Statistical  So- 
ciety 30,  108-122  (1968). 

[5]  DeGroot,  M.H.,  Optimal  Statistical  Decisions  (McGraw-Hill,  1970). 

[6]  Derman,  C.  and  J.  Sacks,  "Replacement  of  Periodically  Inspected  Equipment,"  Naval 
Research  Logistics  Quarterly  7,  597-607  (1960). 

[7]  Golabi,  K.,  "Optimal  Inventory  and  Search  Policies  with  Random  Ordering  Costs,"  Work- 
ing Paper  No.  252,  Western  Management  Science  Institute,  University  of  California, 
Los  Angeles  (1976). 

[8]  Lippman,  S.A.  and  J.J.  McCall,  "The  Economics  of  Job  Search:  A  Survey,"  Economic  In- 
quiry 14,  155-189  (1976). 

[9]  MacQueen,  J.B.  and  R.G.  Miller,  Jr.,  "Optimal  Persistence  Policies,"  Operations  Research 
8,  362-380  (1960). 
[10]  Rothschild,  M.,  "Models  of  Market  Organization  with  Imperfect  Information:  A  Survey," 

Journal  of  Political  Economy  81,  1283-1308  (1973). 
[11]  Stigler,  G.J.,  "The  Economics  of  Information,"  Journal  of  Political  Economy  69,  213-225 

(1961). 
[12]  Stigler,  G.J.,  "Information  in  the  Labor  Market,"  Journal  of  Political  Economy  70,  94-104 

(1962). 
[13]  Wagner,  H.M.  and  T.M.  Whitin,  "Dynamic  Version  of  the  Economic  Lot  Size  Model," 
Management  Science  5,  89-96  (1958). 

APPENDIX 

THEOREM  2:   Under  the  assumptions  of  the  model,  for  all  k,  k  =  1,2,  ...  ,  N, 

(a)  v*(0)  =  Ak 

(b)  vA.(z)  =  H{z)  +  vfc_z(0)    for   1  <  z  <  k. 
Consequently,  the  search  process  is  initiated  only  when  the  inventory  level  is  zero. 

Before  proving  Theorem  2,  we  establish  two  elementary  facts. 

LEMMA  A:    For  any  two  positive  integers  /and  j,  H{i  +  j)  >  //(/)  +  H{j). 

PROOF: 

HU  +  j)  =  '  £    h  ik)  =  £  h  (k)  +  '  £    h (Jd  >  J  h  (k)  +  £  h (k) 

A=l  k=\  k=i  k=\  k=\ 

=  HU)  +  Hij).  Q.E.D 

LEMMA  B:  The  integral  J  "  [y  —  ln(x)]  dF(x)  =  G„(y)  is  strictly  increasing  in  y, 
continuous,  and  unbounded  above. 

PROOF:  Since  In[a„(y)]  =  y  and  /„(*)  is  strictly  increasing  in  x,  it  follows  that  an{y)  is 
strictly  increasing  in  y.  Hence,  G„(y)  is  strictly  increasing,  continuous  (as  Fis  continuous)  and 
unbounded  above.  OFF) 


PROOF  OF  THEOREM  2:   The  proof  is  by  induction  on  k.   From  Equations  (1),  (3),  (2) 
and  (6),  we  have 

(A-l)  vA(0)  =  uk(0)  =  s  +  Ex[Vk(0,x)} 

=  5  +  E  min  |vA(0),    min    [ax  +  h(a  -  1)  +  vfc_,  {a  -  \)]\ 


=  5  +  E  min  [vk(0),  Ik(x)]. 

For  A:  =  1,  (b)  is  obvious.    To  show  (a),  note  that  by  (6),  /|(.v)  =  x.    Next,  from  (A-l) 
we  have 

rv,(0)  r~ 

v,(0)  =  s  +  E  min  [v,(0),x]  =  5  +  Jq        xdE(x)  +  Jv  (Q)  v,(0)  dF{x), 
from  which  we  obtain 

-v   (0) 

(A-2)  v,(0)F[v,(0)]  =  s  +  Jo        xdF(x). 

(Note  the  close  connection  between  V|(0)  and  the  maximizing  price  in  the  house  selling  prob- 
lem.) In  order  to  determine  whether  v,(0)  is  the  unique  solution  to  (A-2),  note  that  it  is 
equivalent  to  verify  that  s  =  J  (y  -  x)  dF(x)  =  G|(v)  has  a  unique  solution.  The  latter 
result  follows  from  Lemma  B. 

From  (7)  we  have  /i(c*i)  =  v,(0)  and  therefore  a,  =  v,(0).   Thus,  (A-2)  becomes 
a,F(a,)  =  5  +  J    '  xdF(x), 
which  coupled  with  (8b)  for  n  =  1,  gives  A  |  =  ot\  =  v,(0)  so  that  (a)  holds  for  k  =  1. 

Assume  (a)  and  (b)  hold  for  k  =  1,2,  ...  ,  n  —  1.    We  show  that  the  theorem  holds  for 

k  =  n. 

From  (A-l),  we  have 
v„(0)  =  s  +  E  min  [v„(0),  I„(x)] 

=  s  +  fj"  In(x)  dFix)  +  £~  v„(0WFU) 

=  [F(a„)]-] \s  +  J""  [  min     [ax  +  h(a  -  1)  +  v/;_,(o  -  l)]|rfF0c) 

=  [F(a„))"'L  +  Jp  (  min    [ax  +  h'(a  -  1)  +  H(a  -  1)  +  v„_o(0)]]  dF(x)\ 

=  [F(a„)]-^ \s  +  Jo""  (  min    [ax  +  H(a)  +  A„_a]\  dF(x)\ 

-  [F(a„)]-*AnF(a„) 

=  A„f 

where  the  second  equality  follows  from  Theorem  1,  the  third  from  a  simple  rearrangement  of 
terms,  the  fourth  and  fifth  equalities  from  the  induction  hypothesis  and  the  sixth  equality  from 
(8b).   Therefore  (a)  is  true  for  k  =  n. 


INVENTORY  MODEL  WITH  SEARCH 


657 


Since  we  are  assuming  that  (b)  holds  for  k  =  n  —  1,  it  follows  that 


AnF(an)  =  5  +    C"  I„(x)dF(x), 
which  gives 
(A-4)  fj"  [A„  -  Itl(x)]  dF(x)  =  s. 

We  note  that  by  (4)  and  the  induction  hypothesis,  for  1  <  z  <  n 

w„(z)  =  h(z  -  1)  +  v„_,(z  -  1)  =  h(z  -  1)  +  H(z  -  1)  +  vn_z(0) 

=  H(z)  +v„_z(0), 

and  therefore  to  prove  (b)  for  k  =  n,  it  suffices  to  show  v„(z)  =  w„(z)  whenever  z  >  1.   That 
is,  we  need  to  show  u„{z)  ^  //(z)  +  v„_-(0)  whenever  z  ^  1. 

We  can  write 

«.(z)  =  s  +  E  mi 


=  s  +  E  m 
=  5  +  E  m 
^  5  +  E  mi 

=  s  +  E  mi 


where  the  first  equality  follows  from  (3)  and  (2),  the  second  from  (1),  the  fourth  from  induc- 
tion hypothesis  and  the  last  equality  from  the  induction  hypothesis  and  (A-3).  The  inequality 
follows  from  Lemma  A.    Hence, 

(A-5)  y  =  u„{z)  -  H(z)  >  s  +  E  min  \u„(z)  -  H(z),  min  [vn_r(0),/„--U)]  . 

If  y  were  less  than  v„_.(0),  then  from  (A-5)  we  would  have 
y  ^  s  +  E  min  [-y,/„_-(x)] 


nj«„(z),//(z)  +     min      [ax  +  H{a)  +  A„_:_a]{ 
n|w„(z),//(z)  +  minU„_r,      min      (ax  +  H(a)  +  A„_:_a)][ 
n \u„(z),H(z)  +  min  [v„_r(0),/„_zU)]  , 


658  K  CiOLABI 

giving 

fQa"~:  \    [y  -  I„-:(x))dF(x)  >  s  =  fQa"~:  [A„_z  -  I„_Ax)]dF(x), 
where  the  equality  follows  from  (A-4)-.    Hence, 

G„.z(y)  >  G„_:(A„-:)=  G„_r[v„_r(0)], 
contradicting  Lemma  B.   Therefore,  y  ^  v„_2(0),  which  completes  the  induction  argument. 

Q.E.D. 


THE  UNITED  STATES  COAST  GUARD  COMPUTER-ASSISTED  SEARCH 
PLANNING  SYSTEM  (CASP)* 

Henry  R.  Richardson 

Daniel  H.  Wagner,  Associates 
Paoli,  Pennsylvania 

Joseph  H.  Discenza** 
U.S.  Coast  Guard 

ABSTRACT 

This  paper  provides  an  overview  of  the  Compuler-Assisied  Search  Planning 
(CASP)  system  developed  for  the  United  Stales  Coast  Guard.  The  CASP  in- 
formation processing  methodology  is  based  upon  Monte  Carlo  simulation  to 
obtain  an  initial  probability  distribution  for  target  location  and  to  update  this 
distribution  to  account  for  drift  due  to  currents  and  winds.  A  multiple  scenario 
approach  is  employed  to  generate  the  initial  probability  distribution.  Bayesian 
updating  is  used  to  reflect  negative  information  obtained  from  unsuccessful 
search.  The  principal  output  of  the  CASP  system  is  a  sequence  of  probability 
"maps"  which  display  the  current  target  location  probability  distributions 
throughout  the  time  period  of  interest.  CASP  also  provides  guidance  for  allo- 
cating search  effort  based  upon  optimal  search  theory. 


1.    INTRODUCTION 

This  paper  provides  an  overview  of  the  computer-assisted  search  planning  (CASP)  system 
developed  for  the  United  States  Coast  Guard  to  assist  its  search  and  rescue  (SAR)  operations. 
The  system  resides  on  a  CDC  3300  located  in  Washington,  D.C.,  and  can  be  used  by  all  USCG 
Rescue  Coordination  Centers  (RCCs)  in  the  continental  United  States  and  Hawaii  via  remote 
access  terminals. 

The  Coast  Guard  is  engaged  daily  in  search  and  rescue  missions  which  range  from  simple 
to  complex.  The  amount  of  information  available  to  predict  the  position  of  the  search  target 
ranges  from  extremely  good  to  almost  no  information  at  all.  The  process  of  planning,  com- 
manding, and  evaluating  these  searches  takes  place  in  Rescue  Coordination  Centers  (RCCs) 
located  throughout  the  United  States  in  major  coastal  cities. 

The  entire  planning  process  begins  with  the  awareness  that  a  distress  on  the  water  may 
exist.  This  awareness  usually  results  from  a  telephone  call  from  a  friend  or  relative  or  from  a 
radio  communication  from  the  boat  or  vessel  itself. 

This  work  was  supported  in  part  by  USCG  Contract  DOT-CG-32489-A  and  ONR  Contraci 
**The  opinions  or  assertions  contained  herein  are  the  private  ones  of  the  author  and  are  r 
or  reflecting  the  view  of  the  Commandant  or  the  Coast  Guard  at  large 


660  H.R.  RICHARDSON  AND  J.H    DISCENZA 

Next  all  available  information  has  to  be  evaluated  to  decide  whether  or  not  to  begin  a 
search,  and  what  level  of  effort  is  required  given  the  search  begins.  At  this  point  a  great  deal  of 
effort  goes  into  deciding  where  the  distress  incident  occurred.  This  might  be  considered  the 
first  phase  of  planning. 

The  next  phase  involves  computing  where  the  search  target  will  be  when  the  first  search 
units  arrive  on  scene.  Among  other  things,  this  requires  the  prediction  of  ocean  drift  and  wind 
velocity  and  the  estimation  of  uncertainties  in  these  predictions. 

The  next  questions  pertain  to  the  effort  allocation  process— how  much  effort  must  be 
expended  and  in  what  areas?  Prior  to  the  advent  of  computer  search  programs,  SAR  planners 
relied  upon  various  rules  of  thumb  as  presented  in  the  National  Search  and  Rescue  Manual 
[11].  Simplicity  was  necessary  to  facilitate  hand  computation,  but  at  the  same  time  prevented 
adequate  treatment  of  the  many  sources  of  uncertainty  which  characterize  a  SAR  incident. 

The  search  phase  is  the  actual  deployment  of  aircraft  and  vessels,  the  conduct  of  preset 
search  patterns,  and  the  report  of  results  back  to  the  RCC. 

If  the  search  is  unsuccessful  for  that  day,  then  the  results  must  be  reevaluated  and  a  new 
search  planned  for  the  following  day. 

This  process  continues  until  the  target  is  found  or  until  the  search  is  terminated.  In  brief 
(and  in  slightly  more  technical  terms),  the  planning  phases  are  as  follows: 

(1)  Determine  the  target  location  probability  distribution  at  the  time  of  the  distress 
incident. 

(2)  Update  the  target  location  probability  distribution  to  account  for  target  motion  prior  to 
the  earliest  possible  arrival  of  a  search  unit  on-scene. 

(3)  Determine  the  optimal  allocation  of  search  effort,  and  estimate  the  expected  amount 
of  search  effort  required  to  find  the  target. 

(4)  Execute  the  search. 

(5)  If  the  search  is  unsuccessful,  evaluate  the  results  and  update  the  target  location  proba- 
bility distribution  to  account  for  this  negative  information. 

(6)  Repeat  the  planning  procedures  in  Steps  (2)  through  (5)  until  the  target  is  found  or 
the  search  is  terminated. 

These  planning  phases  are  illustrated  in  the  CASP  case  example  given  in  Section  3. 

The  first  efforts  at  computerization  concentrated  on  the  target  location  prediction  process. 
Oceanographic  models  were  used  to  compute  drift  and  to  estimate  target  position.  The  Mon- 
terey Search  Planning  Program  and  the  Coast  Guard's  own  Search  and  Rescue  Planning  Sys- 
tem, SARP,  represented  early  computer  assisted  search  efforts.  Even  today,  in  cases  where  the 
information  available  makes  the  planning  straightforward,  the  SARP  program  does  nicely. 

In  1970,  the  Office  of  Research  and  Development  in  Washington  funded  development  of 
a  more  comprehensive  approach  to  search  planning  based  in  part  on  lessons  learned  in  the 
Mediterranean  H-bomb  search  in  1966  (Richardson  [5])  and  in  the  Scorpion  search  in  1968 


COASTGUARD  COMPUTER-ASSISTED  SEARCH  (CASP)  661 

(Richardson  and  Stone  [6]).  In  1972,  the  CASP  system  was  delivered  to  the  Operations 
Analysis  Branch  of  Commander  Atlantic  Area  in  New  York  for  evaluation,  implementation, 
and  training.   The  system  was  made  operational  early  in  1974. 

CASP  is  now  in  use  in  11  Coast  Guard  rescue  centers.  In  addition,  CASP  has  been  used 
at  the  Air  Force  Central  Rescue  Headquarters  at  Scott  AFB,  Illinois,  to  help  plan  and  coordi- 
nate search  missions  for  lost  airplanes  within  the  continental  United  States.  A  modification  of 
the  CASP  system  has  also  been  provided  to  the  Canadians  for  inland  SAR  planning. 

At  the  present  time,  the  use  of  CASP  is  limited  to  open  ocean  searches.  Even  though 
these  searches  represent  but  a  small  percentage  of  the  total  U.S.  Coast  Guard  search  operations, 
CASP  has  been  credited  with  saving  over  a  dozen  lives. 

Section  2  provides  a  description  of  the  CASP  methodology.  Section  3  illustrates  the  use 
of  CASP  in  an  actual  SAR  incident  involving  the  1976  sinking  of  the  sailing  vessel  S/V  Spirit  in 
the  Pacific,  and  Section  4  describes  CASP  training. 

2.   CASP  METHODOLOGY 

The  CASP  information  processing  methodology  is  based  upon  Monte  Carlo  simulation  to 
obtain  an  initial  probability  distribution  for  target  location  and  to  update  this  distribution  to 
account  for  drift  due  to  currents  and  winds.  A  multiple  scenario  approach  is  employed  to  gen- 
erate the  initial  probability  distribution.  In  the  sense  used  here,  a  scenario  is  a  hypothetical 
description  of  the  distress  incident  which  provides  quantitative  inputs  for  the  CASP  programs. 
Bayesian  updating  is  used  to  reflect  negative  information  obtained  from  unsuccessful  search. 

The  principal  output  of  the  CASP  system  is  a  sequence  of  probability  "maps"  which 
display  the  current  target  location  probability  distributions  throughout  the  time  period  of 
interest.  CASP  also  provides  guidance  for  allocating  search  effort  based  upon  optimal  search 
theory. 

The  CASP  system  is  composed  of  a  number  of  different  programs,  each  designed  for  a 
different  information  processing  function.  The  program  components  are  MAP,  POSITION, 
AREA,  TRACKLINE,  COMBINATION,  DRIFT,  RECTANGLE,  PATH,  and  MULTI;  the 
functions  are  as  follows: 

(1)  display  the  probability  maps  (MAP), 

(2)  generate  an  initial  distribution  of  target  location  at  the  time  of  distress  (POSITION, 
AREA,  TRACKLINE,  and  COMBINATION), 

(3)  update  the  target  location  probability  distributions  for  motion  subsequent  to  the  time 
of  distress  (DRIFT), 

(4)  update  the  target  location  probability  distributions  for  negative  search  results  and  com- 
pute the  cumulative  detection  probability  (RECTANGLE  and  PATH),  and 

(5)  calculate  optimal  allocations  of  search  effort  (MAP  and  MULTI). 

These  programs  will  be  described  below  following  presentation  of  an  overview  of  the  general 
system  design. 


662  H.R.  RICHARDSON  AND  J .11.  D1SCENZA 

CASP  System  Design 

The  CASP  system  design  was  motivated  by  a  desire  to  provide  a  highly  realistic  probabilis- 
tic description  for  the  target's  location  at  the  time  of  the  distress  incident  and  for  the  target's 
substantial  motion.  In  view  of  the  success  achieved  in  the  Mediterranean  H-bomb  search  [12] 
in  1966,  and  in  the  Scorpion  search  [5]  in  1968,  it  seemed  evident  that  a  Bayesian  approach 
would  provide  a  practical  method  for  incorporating  information  gained  from  unsuccessful 
search. 

Target  motion  modeling  posed  a  more  difficult  problem.  Models  which  were  amenable  to 
an  "analytic"  approach  were  not  flexible  enough  to  give  a  good  representation  of  the  search 
facts.  For  example,  Gaussian  motion  processes  (or  mixtures  of  Gaussian  processes)  were  unsa- 
tisfactory in  cases  where  the  search  facts  required  a  uniform  or  annular  shaped  target  location 
probability  density.  Markov  chains  based  on  transitions  among  search  grid  cells  were  unsatis- 
factory in  cases  where  one  desired  to  change  the  grid  in  the  course  of  an  operation.  In  general, 
these  models  tended  to  force  the  facts  to  fit  the  mathematics  to  an  undesirable  extent. 

It  was  also  desired  to  develop  a  modular  system  so  that  additional  features  and  improve- 
ments could  be  made  as  time  went  on.  In  order  to  gain  the  confidence  of  the  users,  the  system 
had  to  be  simple  to  understand  and  require  a  minimum  of  unfamiliar  inputs.  The  design  which 
seemed  best  suited  in  view  of  the  above  considerations  is  a  hybrid  approach  which  uses  Monte 
Carlo  to  simulate  target  motion  and  analytic  methods  to  compute  detection  probabilities. 

A  motivation  for  use  of  Monte  Carlo  was  the  recognition  that  computation  of  the  poste- 
rior target  location  probability  distribution  can  be  viewed  as  the  numerical  evaluation  of  a  mul- 
tivariate integral  of  high  dimensionality.  In  such  cases  (i.e.,  high  dimensionality),  classical 
numerical  integration  techniques  perform  poorly  (see,  for  example,  Shreider  [7])  especially 
when  the  integrands  can  have  jump  discontinuities  and  are  not  of  a  simple  analytic  form. 
These  problems  are  typical  of  CASP  applications.  Discontinuities  occur  when  the  "target" 
moves  into  a  region  where  search  effort  is  concentrated,  and  the  joint  probability  density  for 
target  position  at  several  specified  times  during  the  search  is  a  very  complicated  function. 

The  underlying  structure  of  CASP  is  a  Markov  process,  with  a  three-dimensional  state 
space  consisting  of  points  (X,  Y,  $).  The  variables  Zand  Y  denote  latitude  and  longitude  and 
<E>  denotes  search  failure  probability.  For  j=  1,  ...  ,  7,  the  yth  Monte  Carlo  replication 
(Xj,  Yj,  $>j)  represents  the  target's  current  position  (time  is  implicit)  together  with  the  cumula- 
tive probability  of  search  failure  for  that  particular  target  replication  computed  for  its  entire  his- 
tory. Target  motion  is  assumed  to  be  Markovian  and  successive  increments  of  search  are 
assumed  to  be  statistically  independent.  Thus  {Xjt  Yj,  <!>,)  completely  describes  the  state  of  the 
yth  target  replication  at  a  given  moment. 

Figure  1  provides  a  schematic  diagram  for  the  operation  of  the  CASP  system.  All  of  the 
programs  mentioned  will  be  discussed  individually  in  subsequent  subsections.  The  first  step  is 
to  construct  a  file  (called  the  "target-triple  file")  consisting  of  samples  from  the  target  location 
probability  distribution  at  the  time  of  the  distress  incident.  This  file  is  stored  on  computer  disc 
and  processed  sequentially  by  various  programs. 

These  initial  points  {Xr  Yr  1)  have  failure  probabilities  <& , ■  =  1,  since  no  search  has  yet 
been  carried  out.  The  target  positions  (Xr  Yj)  are  sampled  from  a  probability  density  function 
Fof  the  form 


COAST  GUARD  COMPUTER-ASSISTED  SEARCH  (CASP) 


Scenario  Generation 


:)  Update  for  Negative  Search  Results 


Figure  1.   Casp  syste 


f=  E  *kfk, 


where  fk  is  the  density  corresponding  to  the  kth  "scenario,"  and  wk  >  0  is  the  scenario's  subjec- 

Ik  | 


Monte  Carlo  samples  from  a  probability  density  F  are  obtained  by  first  using  one  of  the 
"generation  programs"  POSITION,  AREA,  or  TRACKLINE.  Averages  of  densities  of  different 
types  are  obtained  by  forming  preliminary  target  triple  files  with  two  or  more  "generation"  pro- 
grams and  then  combining  them  with  the  program  COMBINATION.  The  construction  of  the 
prior  target  location  probability  distribution  is  shown  schematically  in  Figure  1(a). 

Updates  for  target  motion  (Figure  Kb))  or  to  account  for  negative  search  results  (Figure 
1(c))  are  carried  out  by  reading  the  "old"  target  triple  file  from  disc  into  the  appropriate  pro- 
gram and  outputting  a  "new"  target  triple  file.  When  program  DRIFT  is  used  (Figure  Kb)),  the 
values  of  Xj  and  Y}  are  modified,  but  the  value  of  4>;  remains  unchanged.    For  an  update  for 


664 


H.R.  RICHARDSON  AND  J.H.  DISCENZA 


negative  search  results,  the  file  is  first  updated  for  motion  by  use  of  program  DRIFT.  The  tar- 
get triple  file  is  frozen  at  the  mid-search  time  and  then  modified  by  RECTANGLE  or  PATH. 
These  programs  modify  <!>,  by  use  of  Bayes'  theorem  but  the  position  variables  Xj  and  Yj 
remain  the  same  since  motion  is  frozen. 

The  probability  distributions  and  optimal  allocations  of  search  effort  are  displayed  using 
program  MAP  or  MULTI  (Figure  1(d)).  In  both  cases,  this  is  a  read-only  operation,  and  the 
target  triple  file  is  not  modified. 


Display 

The  MAP  program  displays  the  target  location  probability  distributions  in  a  two  dimen- 
sional format.  Figure  2  shows  an  example  of  a  probability  map  corresponding  to  an  actual  SAR 
case.  The  geographical  region  is  divided  into  cells  oriented  north-south  and  east-west  and  the 
target  location  probabilities*  for  each  cell  are  multiplied  by  10,000  and  displayed.  Thus,  the 
number  1800  in  a  cell  indicates  that  the  target  location  probability  is  .18.  Equal  probability  con- 
tours are  usually  sketched  to  make  it  easier  to  visualize  the  probability  distribution. 


Figure  2.  Target  location  probability  distribution  probabilities  are  multiplied  at  10,000  and  truncated 

A  "quick  map"  in  which  symbols  are  used  to  represent  ranges  of  probabilities  can  also  be 
output.  The  quick  map  provides  a  compact  version  of  the  probability  distribution  which  is  suit- 
able for  a  quick  appraisal  of  the  search  situation  and  is  convenient  for  inclusion  in  after-action 
reports. 

Finally,  MAP  can  output  an  ordered  list  of  the  highest  probability  cells  and  the  amount  of 
effort  to  be  placed  in  each  cell  in  order  to  maximize  detection  probability.  More  will  be  said 
about  search  optimization  in  the  last  subsection. 


The  format  implies  higher  accuracy  than  is  warranted  in  view  of  the  Monte  Carlo  procedures  employed. 


COAST  GUARD  COMPUTER-ASSISTED  SEARCH  (CASP)  665 

Initial  Target  Location  Probability  Distribution 

The  initial  target  location  probability  distribution  is  constructed  from  "building  block  dis- 
tributions" using  a  weighted  scenario  approach.  The  individual  building  block  distributions  are 
generated  by  the  use  of  one  or  more  of  the  programs  POSITION,  TRACKLINE,  and  AREA. 
Program  COMBINE  is  used  to  combine  the  outputs  of  the  individual  "generation"  programs. 

In  most  SAR  cases,  there  is  scant  information  available  about  the  target's  position  at  the 
time  of  distress.  Sometimes,  for  example,  a  fisherman  simply  is  reported  overdue  at  the  end  of 
a  day.  He  may  have  been  planning  to  fish  in  one  of  several  fishing  grounds  but  did  not  make 
his  precise  intentions  known. 

In  other  cases,  more  information  is  available.  For  example,  it  might  be  known  that  a 
vacationer  was  intending  to  sail  from  one  marina  to  another  but  never  arrived  at  the  intended 
destination.  In  some  cases,  it  might  also  be  known  that  there  was  bad  weather  along  the 
intended  route.  This  would  make  some  positions  along  track  more  likely  for  a  distress  than 
others. 

In  order  to  encourage  inclusion  of  diverse  possibilities  in  these  scenarios,  it  is  a  recom- 
mended practice  for  two  or  three  search  planners  to  work  out  the  details  together.  The 
remainder  of  this  subsection  will  describe  the  programs  POSITION,  AREA,  and  TRACKLINE 
which  are  used  to  simulate  the  scenarios  and  generate  the  initial  target  location  probability  dis- 
tribution. 

Position.  A  POSITION  scenario  has  two  parts,  an  initial  position  and  a  subsequent  dis- 
placement.  POSITION  can  be  used  to  generate  a  weighted  average  of  as  many  as  ten  scenarios. 

The  initial  position  probability  distribution  is  modeled  as  a  bivariate  normal  distribution, 
and  the  displacement  is  modeled  as  a  distribution  over  an  annular  sector.  In  the  latter  distribu- 
tion, the  angle  and  distance  random  variables  are  assumed  to  be  independent  and  uniformly 
distributed  between  minimum  and  maximum  values  input  by  the  user.  The  displacement  distri- 
bution is  useful,  for  example,  in  cases  where  the  initial  position  corresponds  to  the  last  fix  on 
the  target  and  where  one  can  estimate  the  course  and  speed  of  subsequent  movement  prior  to 
the  occurrence  of  the  distress  incident. 

The  displacement  option  can  also  be  used  in  cases  involving  a  "bail  out"  where  it  can 
describe  the  parachute  drift.  The  amount  of  displacement  in  this  case  will  depend  upon  the 
altitude  of  the  aircraft  and  the  prevailing  winds  at  the  time.  Since  these  factors  are  rarely 
known  precisely,  the  capability  to  "randomize"  direction  and  distance  is  an  important  feature. 

Area.  The  second  generation  program  is  AREA.  This  program  is  used  to  generate  an  ini- 
tial target  location  probability  distribution  in  cases  where  a  general  region  can  be  postulated  for 
the  location  of  the  distress  incident  but  where  a  normal  distribution  simulated  by  POSITION 
would  be  a  poor  representation  of  the  uncertainty.  Each  scenario  for  program  AREA  deter- 
mines a  uniform  probability  density  within  a  convex  polygon.  AREA  might  be  used,  for  exam- 
ple, when  a  lost  fisherman's  usual  fishing  ground  is  known  from  discussions  with  friends  and 
relatives.    As  with  POSITION,  AREA  can  generate  a  weighted  average  of  10  scenarios. 


666 


H.R.  RICHARDSON  AND  J.H.  DISCENZA 


Trackline.  The  third  and  last  generation  program  is  TRACKLINE.  This  program  is  the 
most  complex  of  the  generation  programs  and  is  used  when  target  track  information  is  available 
from  a  float  plan  or  some  other  source.  TRACKLINE  creates  a  probability  distribution  about  a 
base  track.  This  track  can  be  constructed  from  as  many  as  10  segments,  each  of  which  can  be  a 
portion  of  a  rhumb  line  or  of  a  great  circle. 

The  motion  of  the  target  about  each  base  track  segment  is  specified  by  three  circular  nor- 
mal probability  distributions  corresponding  to  target  position  at  the  initial,  mid-point,  and  end- 
point  of  each  segment.  Each  simulated  target  track  is  obtained  by  drawing  random  numbers  for 
target  position  from  these  distributions  and  then  connecting  the  points  with  straight  lines. 

Figure  3  illustrates  a  typical  situation.  The  target's  point  of  departure  and  intended  desti- 
nation are  assumed  known,  and  a  base  track  is  constructed  between  these  points.  The  base 
track  might  be  taken  from  the  target's  float  plan  or  hypothesized  from  the  target's  past  habits. 
In  the  case  illustrated  by  Figure  3,  there  are  three  track  segments.  The  50%  circles  of  uncer- 
tainty are  assumed  to  grow  in  size  to  about  midway  along  the  track  and  then  diminish.  Since 
the  point  of  departure  and  intended  destination  are  assumed  to  be  known,  the  extreme  end- 
points  of  the  entire  track  have  zero  uncertainty. 


Figure  3.  Description  of  irackli 


COAST  GUARD  COMPUTER-ASSISTED  SEARCH  (CASP)  667 

In  some  cases,  there  is  information  which  leads  one  to  suspect  that  the  distress  is  more 
likely  to  have  occurred  on  one  part  of  the  track  than  on  another.  For  example,  as  mentioned 
above,  the  track  may  have  passed  through  an  area  of  storms  and  heavy  seas.  If  desired,  the  tar- 
get location  probability  distribution  generated  by  TRACKLINE  can  be  made  to  have  a  higher 
density  in  such  an  area.  This  is  done  by  specifying  the  highest  probability  point  along  base 
track  together  with  the  odds  that  the  distress  occurred  there  rather  than  at  the  extreme  end- 
points  of  the  track.  These  inputs  determine  a  truncated  triangular  probability  density  for  the 
fraction  of  track  covered  before  the  distress  incident  occurred. 

Updating  for  Target  Motion 

The  DRIFT  program  is  used  to  alter  a  target  location  probability  distribution  to  account 
for  the  effects  of  drift.  Normally,  the  DRIFT  program  will  cause  the  center  of  the  distribution 
to  move  to  a  new  location  and  the  distribution  to  become  more  diffuse. 

Target  motion  due  to  drift  complicates  the  maritime  search  problem.  The  prediction  of 
drift  must  account  for  the  effects  of  both  sea  current  due  to  prevailing  circulation  and  predicted 
or  observed  surface  wind.  Any  object  floating  free  on  the  ocean  surface  is  transported  directly 
by  surface  current,  and  one  component  vector  of  drift  is  therefore  equal  to  the  predicted 
current  vector.  A  statistical  file  collected  from  ship  reports  over  many  years  has  been  assem- 
bled by  the  Coast  Guard  and  arranged  by  geographical  location  and  month  of  the  year.  The  file 
in  use  in  the  CASP  system  covers  most  of  the  North  Atlantic  and  North  Pacific  Oceans. 

As  mentioned  above,  wind  is  also  important  in  predicting  target  motion.  With  regard  to 
this  factor,  there  are  two  major  considerations.  The  first  is  the  drift  caused  by  the  wind  imping- 
ing on  the  drifting  object's  surface  area  above  water;  this  is  called  "leeway."  The  speed  and 
direction  of  leeway  is  different  for  different  objects,  and  is  usually  difficult  to  predict. 

The  second  wind  consideration  is  the  movement  of  the  surface  layer  of  the  ocean  itself; 
this  is  called  "local  wind  current."  It  is  one  of  the  most  complex  and  least  understood 
phenomena  in  the  entire  drift  process. 

The  primary  data  source  for  surface  winds  in  the  CASP  system  is  the  Navy's  Fleet 
Numerical  Weather  Central  in  Monterey,  California.  Every  twelve  hours  their  computers  gen- 
erate a  time  series  for  hemispheric  wind  circulation;  three  of  these  time  series  are  used  to  pro- 
duce certain  geographical  blocks  of  wind  data  which  are  transmitted  to  the  Coast  Guard  for  use 
by  CASP.   All  data  are  retained  in  the  system  for  two  to  three  months. 

The  process  of  applying  the  drift  motion  to  update  a  CASP  distribution  is  simple  enough. 
First,  a  set  of  total  drift  vector  probability  distributions  is  computed  for  various  geographical 
areas  based  upon  estimates  of  sea  current,  leeway,  and  local  wind  current.  Then  for  each  target 
location  replication,  a  random  vector  of  net  drift  is  drawn  from  the  appropriate  probability  dis- 
tribution and  used  to  move  the  target  forward  a  short  time.  The  procedure  is  repeated  until  the 
entire  update  time  is  taken  into  account. 

Updating  for  Negative  Search  Results 

Once  a  search  has  actually  been  conducted,  one  of  the  two  search  update  programs,  REC- 
TANGLE and  PATH  (depending  upon  the  type  of  search),  is  run  to  revise  the  target  location 
probabilities  to  account  for  unsuccessful  search.  The  effect  is  to  reduce  the  probabilities  within 
the  area  searched,  and  to  increase  them  outside. 


668  H.R.  RICHARDSON  AND  J.H.  DISCENZA 

Updating  the  target  location  probabilities  for  negative  search  results  is  carried  out  by  an 
application  of  Bayes'  theorem.  Recall  that  the  target  triple  file  contains  J  records  of  the  form 
(Xj,  Yj,  <t>j)  for  1  <  j  <  J,  where  the  pair  (XJt  Yj)  represents  target  position,  and  $, 
represents  the  probability  that  the  target  replication  would  not  have  been  detected  by  the  cumu- 
lative search  effort  under  consideration.  The  overall  cumulative  probability  of  detection  taking 
all  simulated  targets  into  account  is  called  search  effectiveness  probability  (SEP)  and  is  com- 
puted by  the  formula 

SEP  =  1  -  X  4),-// 

/=i 

Let  C  be  a  region  in  the  search  area,  and  let  fl,  denote  the  event,  "target  corresponds  to 
the  yth  replication  and  is  in  region  C"  The  posterior  probability  A(C)  that  the  target  is  located 
in  C  given  search  failure  is  computed  using  Bayes'  theorem  by 

j 
A(C)  =  Pr {Target  in  C\  Search  failure}  =  £  Pr[B,\  Search  failure} 

j 
=  £  /V  {Search  failure  |  B,\  Pr[B, }/Pr{ Search  failure} 

,/€r        y-i 
where  r  =  [j :  (Xh  Yj)  €  C)  denotes  the  set  of  indicies  corresponding  to  target  replications  in  C. 

Now  suppose  that  q}  denotes  the  probability  of  failing  to  detect  the  yth  target  replication 
during  a  particular  update  period.  Using  the  independence  assumption,  the  new  individual 
cumulative  failure  probability  4>,  is  computed  by 

4>;  =  q^'j, 

where  <$>',  denotes  the  cumulative  failure  probability  prior  to  the  last  increment  of  search. 

The  computation  of  the  conditional  failure  probability  qt  is  carried  out  in  CASP  by  use  of 
a  {M,  B,  o-)-detection  model  as  described  below.  Recall  (e.g.,  see  Koopman  [2])  that  the 
"lateral  range"  between  searcher  and  target  (both  with  constant  course  and  speed)  is  defined  as 
the  distance  at  closest  point  of  approach.  The  "lateral  range  function"  gives  single  sweep  cumu- 
lative detection  probability  for  a  specified  lateral  range  for  a  specified  period  of  time.  The 
integral  of  the  lateral  range  function  is  called  the  "sweep  width"  of  the  sensor. 

The  CASP  programs*  are  based  upon  the  assumption  that  the  lateral  range  function  for 
the  search  unit  is  rectangular  and  is  described  by  two  parameters,  M  and  B.  Here  M  denotes 
the  total  width  of  the  swept  path,  and  /3  denotes  ihe  probability  that  the  target  would  be 
detected  for  lateral  ranges  less  than  or  equal  to  Mil.  The  sweep  width  W  for  the  rectangular 
lateral  range  function  described  above  is  given  by 

W  =  BM. 

Navigational  uncertainties  ("pattern  error")  are  introduced  into  the  detection  model  by 
assuming  each  sweep  is  a  random  parallel  displacement  from  the  intended  sweep.    The  random 


*An  option  is  also  provided  to  use  an  "inverse  cube"  lateral  range  function  as  defined  in  [2]  together  with  search  pattern 


COAST  GUARD  COMPUTER-ASSISTED  SEARCH  (CASP)  669 

displacements  are  assumed  to  be  independent  identically  distributed  normal  random  variables 
with  zero  mean  and  standard  deviation.  This  model  was  introduced  by  R.  K.  Reber  (e.g.,  see 
Reber  [4])  and  used  extensively  in  certain  Navy  search  analyses. 

Rectangular  lateral  range  functions  are  a  useful  way  of  approximating  more  complex 
lateral  range  functions.  If  the  actual  lateral  range  function  has  sweep  width  M  and  is  nonzero 
over  an  interval  of  width  M,  then  one  may  define  /3  to  be  the  average  detection  probability  over 
the  effective  range  of  the  sensor,  i.e.,  /3  =  WjM.  Appendix  A  of  [4]  shows  that  replacement  of 
the  actual  lateral  range  function  by  a  rectangular  lateral  range  function  with  average  probability 
/3  usually  does  not  lead  to  significant  errors  in  the  computed  value  of  probability  of  detection 
for  parallel  path  search.  Cases  where  there  is  significant  disagreement  occur  when  the  lateral 
range  function  is  close  to  zero  over  a  large  part  of  its  support. 

Let  G„  denote  the  cumulative  normal  probability  distribution  function.  Let  (u,  v)  denote 
the  target's  position  in  a  coordinate  system  where  the  origin  is  at  the  midpoint  of  a  given 
sweep,  and  where  the  w-axis  is  parallel  to  the  sweep  and  the  v-axis  is  perpendicular  to  the 
sweep.  Then  for  fixed  M,  /3,  and  o-,  the  single  sweep  probability  p(u,  v)  of  detecting  the  target 
is  given  by 

m  I       \       Jr    I     _l    L\        r    \  L)l  \r    \     j.    M\        r    (  M 

(1)  p\u,\)  =  j8  GJ«  +  —   —  Ga  \u  — —     \G(T\y  +  —    -  G„  v  — — 

where  L  denotes  the  length  of  the  sweep. 

If  there  are  K  search  legs  to  be  considered,  and  if  {uf,  wf)  denotes  the  coordinates  of  the 
yth  simulated  target  position  relative  to  the  kx\\  search  leg,  then  the  failure  probability  <?,  is 
given  by 

(2)  qj=  n  [1 -/>(«/*.  v/,)]. 

The  application  of  these  formulas  in  programs  PATH  and  RECTANGLE  can  now  be  discussed. 

Path.  Program  PATH  is  used  to  represent  general  search  patterns  constructed  from 
straight  track  segments.  For  example,  PATH  can  be  used  to  compute  detection  probabilities  for 
a  circle  diameter  search  where  the  search  tracks  are  intended  to  cover  a  given  circle  by  making 
repeated  passes  through  its  center.    PATH  makes  direct  use  of  (1)  and  (2). 

Rectangle.  Program  RECTANGLE  has  been  designed  for  the  special  case  where  a  rectan- 
gle is  searched  using  parallel  sweeps.  RECTANGLE  reduces  the  computing  time  and  amount 
of  input  that  otherwise  would  be  required  using  program  PATH.  For  a  point  outside  the  desig- 
nated rectangle,  the  probability  of  detection  qi  is  assumed  to  be  0.  For  a  point  inside  the  desig- 
nated rectangle,  "edge"  effects  are  ignored  and  an  average  probability  of  detection  is  computed 
as  if  there  were  an  infinite  number  of  sweeps,  each  infinitely  long. 

The  following  line  of  reasoning  originated  with  R.  K.  Reber.  Reber  [4]  presents  results  in 
the  form  of  curves  and  tables,  and  these  have  been  adapted  to  program  RECTANGLE  by  use 
of  polynomial  approximations.  Let  S  denote  the  spacing  between  sweeps.  Since  the  sweeps  are 
assumed  to  be  parallel  and  of  infinite  extent,  the  coordinate  v/  expresses  the  lateral  range  for 
the  kth  sweep  and  the  yth  simulated  target  location  and  is  given  by 

v/  =  fij  +  kS 

for  —  oo  <  /c  <  oo  and  a  number  /x ,  such  that  |/x  / •  |  ^  5. 


670  H.R.  RICHARDSON  AND  J.H.  DISCENZA 

Now  for  arbitrary  /jl,  refer  to  (1)  and  (2)  and  define  g  by 

(3)g(fx,S)=    n    ll-p(u,H  +  kS)]=    f[  /3 \GAfjL  +  kS  +  ~    -  G, 

Note  that  since  the  sweeps  are  assumed  to  be  of  infinite  length,  one  has  u  =  °°  and  g  defined 
by  (3)  does  not  depend  upon  u.    The  function  g  is  periodic  in  its  first  argument  with  period  fi. 
Let  g(S)  denote  the  average  value  of  g(jx,  s)  with  respect  to  the  first  argument.   Then 
1 


Z{S)=  i  Jo   ^•S)d»- 


S 

The  function  g  has  been  tabulated  in  [4]  and  is  used  in  program  RECTANGLE  to 
represent  the  failure  probability  q}  =  g(S)  for  a  point  lying  within  the  designated  search  rectan- 
gle. RECTANGLE  and  PATH  agree  (as  they  should)  when  PATH  is  used  to  represent  a  paral- 
lel path  search. 

Search  Optimization 

Two  programs,  MAP  and  MULTI,  are  used  for  optimizing  the  allocation  of  search  effort. 
MAP  provides  a  quick  way  of  determining  the  search  cells  which  should  receive  effort  based 
upon  a  constraint  on  total  track  line  miles  available.  MULTI  determines  search  areas  for  multi- 
ple search  units  under  the  constraint  that  each  unit  must  be  assigned  a  uniform  coverage  of  a 
rectangle  and  that  the  rectangles  for  the  various  search  units  do  not  overlap. 

The  method  used  in  program  MAP  is  based  upon  use  of  an  exponential  detection  function 
(see  Stone  [8])  introduced  by  Koopman  [3]  and  does  not  impose  constraints  on  the  type  of 
search  pattern  employed.  The  primary  usefulness  of  this  program  is  to  provide  the  search 
planner  with  a  quick  method  for  defining  the  area  of  search  concentration.  The  following  para- 
graphs give  a  brief  sketch  of  the  methods  used  in  these  optimization  programs. 

Map.  Let  there  be  TV  search  cells,  and  for  1  <  n  ^  N  let  p„  and  a„  denote,  respectively, 
the  target  location  probability  and  the  area  associated  with  the  nth  cell.  The  probability  density 
for  target  location  in  the  nth  cell  is  given  by  d„  =  p„/an.  Suppose  that  total  search  effort  is 
measured  by  the  product  of  track  line  miles  and  sweep  width. 

Let  y  denote  an  allocation  of  search  effort  where  y(n)  denotes  the  amount  of  search 
effort  (measured  in  area  swept)  allocated  to  the  nth  cell.  Probability  of  detection  PD[y]  is  com- 
puted using  an  exponential  effectiveness  function,  i.e., 

PdW=  Z  PnU  -  exp(-y(n)/a„)]. 

The  objective  is  to  maximize  PD  subject  to  a  constraint  on  total  effort  available.  This  is  easily 
done  using  the  techniques  introduced  by  Koopman  [3];  easier  proofs  are  provided  in  Stone  [8] 
and  Wagner  [12]. 

It  can  be  shown  that  under  the  above  assumptions,  the  initial  increments  of  effort  should 
be  concentrated  in  the  highest  probability  density  cells,  and  that  there  should  be  a  succession  of 
expansions  to  cells  having  lower  target  location  probability  density. 

In  order  to  derive  the  formulas  used  in  program  MAP,  a  new  collection  of  equi-density 
search  regions  is  formed  made  up  of  the  unions  of  all  cells  having  equal  probability  density. 
Let 


COAST  GUARD  COMPUTER-ASSISTED  SEARCH  (CASP) 


671 


K  =  the  number  of  equi-density  regions 

dk  =  the  probability  density  for  region  k 

Ik    =  the  set  of  indices  corresponding  to  the  cells  comprising  region  k 

Ak  =  the  area  of  region  k. 

Using  the  above  notation 

/6/fc 

Let  Ek  denote  the  total  effort  which  must  be  expended  before  the  optimal  search  expands  into 
the  kth  region.  Assume  that  the  equi-density  regions  have  been  ordered  beginning  with  the 
region  having  the  highest  density.  Since  search  begins  in  the  highest  density  region,  we  have 
E\  =  0.    It  can  be  shown  that  in  general  for  k  ^  2 


(4) 


Ek  =  Ek_{  +  (ln^_,  -  In  dk)    £  Am. 

m=\ 


Figure  4  shows  output  from  program  MAP  illustrating  the  use  of  (4).  The  list  shows  the 
25  highest  probability  cells  specified  by  the  latitude  and  longitude  of  the  southeast  corner.  Each 
cell  is  15  minutes  wide,  and  the  numbers  in  the  last  column  correspond  to  the  values  Ek  given 
by  (4).  The  planning  advice  given  in  [10]  is  to  apply  search  effort  to  any  cell  for  which  the 
value  in  the  effort  column  is  less  than  the  total  effort  available. 


TOP  25 

PROBABILITY 

LOCATION 

(S.E.  CORNER) 

EFFORT 

1 

0.05133 

43-ON 

69-45W 

2 

0.04167 

42-45N 

69-30W 

35.0 

3 

0.04133 

43-0N 

70-OW 

36.3 

4 

0.04100 

43-0N 

69-30W 

40.3 

5 

0.03567 

43-15N 

69-30VV 

129.4 

6 

0.03467 

43-15N 

69-45W 

152.8 

7 

0.03333 

42-45N 

69-15W 

199.6 

8 

0.03267 

42-30N 

69-15W 

227.5 

9 

0.03267 

42-45N 

69-45W 

222.2 

10 

0.03267 

43-15N 

70-0W 

210.1 

11 

0.03200 

42-30N 

69-30VV 

264.1 

12 

0.02800 

43-ON 

69-15W 

491.5 

13 

0.02733 

42-45N 

70-0W 

547.2 

14 

0.02533 

43-ON 

70-15W 

701.3 

15 

0.02267 

42-30N 

69-0W 

976.5 

16 

0.02233 

43-15N 

69-15W 

983.1 

17 

0.02167 

42-30N 

69-45W 

1095.1 

18 

0.02133 

43-15N 

70-15W 

1104.4 

19 

0.02100 

42-45N 

69-OW 

1175.3 

20 

0.01867 

43-30N 

69-30W 

1505.8 

21 

0.01867 

43-30N 

69-45W 

1505.8 

22 

0.01800 

43-ON 

69-0W 

1659.9 

23 

0.01667 

42-30N 

68-45W 

1968.0 

24 

0.01600 

42-15N 

69-0W 

2137.7 

25 

0.01600 

42-15N 

69-15W 

2137.7 

Figure  4.  Optimal  allocation  of  effort  produced  by  Map 


672 


H.R.  RICHARDSON  AND  J.H.  DISCENZA 


Notice  that  the  numbers  in  the  effort  column  are  not  necessarily  increasing.  This  is 
because  the  list  is  ordered  according  to  containment  probability  rather  than  probability  density. 

Multi.  As  mentioned  above,  program  MAP  does  not  take  into  account  "simplicity"  con- 
straints which  are  considered  important  in  operational  planning.  Program  MULTI  was  designed 
to  overcome  this  drawback  in  cases  where  multiple  search  units  are  deployed  in  the  same  search 
area. 

The  first  simplicity  constraint  introduced  is  that  each  unit  will  be  assigned  to  uniformly 
search  a  rectangle.  Figure  5  shows  the  dimensions  of  the  optimal  rectangle  and  the  resulting 
probability  of  detection  under  the  assumption  that  the  target  location  probability  distribution  is 
normal.    In  order  to  use  this  figure,  one  first  computes  the  normalized  effort  E*  by  the  formula 

E-  -—*?—. 

crmaxo-rnJn 

where  R  is  the  sweep  rate  of  the  unit,  T  is  the  total  search  time,  and  o-max  and  crmln  are  the 
standard  deviations  of  the  normal  distribution  when  referred  to  principal  axes.  The  optimal 
search  rectangle  will  have  half  side  given  by  U*a-max  and  £/*crmin  where  the  size  factor  U*  is 
given  by  the  designated  curve  with  values  read  along  the  outer  vertical  scale. 


£ 

Tonxh 

b°f 

5 

—  1  0 

ootir 

«l  searc 

h  plan 

Opti 

n. 

Plan 

RT 

Probability  of  Detectio 

b 

—  .9 

E*  = 

(Inn 

er  Scale) 

Half 

sides  of 

rectangle 

I      4 

Q 

""""   8 

given 

by  U* 

?maxand  U*  omin 

s 

—  .7 

2 

Q 

o       3 

—    6 

l 

"*"  Optimal  Rectangle 

Plan 

| 

1 

5 

—     5 

2 

2        2 

—  .3 

—  2 

1 

1 

"*"  Optimal  Rectangle 
(Outer  Scale) 

Normalized  Effort  E' 

Figure  5.  Optimal  search  reciangle 


Figure  5  provides  curves  to  determine  the  probability  of  detection  for  the  optimal  rectan- 
gle plan  and  for  the  unconstrained  optimal  plan.  It  is  interesting  to  note  that  in  all  cases  the 
probability  of  detection  provided  by  the  optimal  rectangle  plan  is  at  least  95%  of  that  provided 


COAST  GUARD  COMPUTER-ASSISTED  SEARCH  (CASP)  673 

by  the  unconstrained  optimal  plan.  Thus,  under  the  assumption  stated,  uniform  search  of  the 
optimal  rectangle  can  be  recommended  without  hesitation  since,  in  most  cases,  the  simplicity  of 
the  rectangle  plan  is  more  important  than  the  small  improvement  in  effectiveness  obtained  by 
the  more  complicated  optimal  plan. 

MULTI  is  capable  of  allocating  the  effort  of  up  to  5  search  units  to  nonoverlapping  rectan- 
gles in  a  way  which  is  intended  to  maximize  overall  probability  of  detection.  The  first  step  in 
this  procedure  is  to  approximate  the  target  location  probability  distribution  by  the  weighted 
average  of  k  bivariate  normal  distribution  where  /  <  k  <  3.  This  is  done  by  locating  the  three 
highest  local  maxima  in  the  smoothed  cell  distribution  and  then  associating  each  simulated  tar- 
get position  with  the  nearest  cluster  point.  If  three  local  maxima  cannot  be  found,  then  the 
procedure  is  carried  out  with  one  or  two  local  maxima.  The  mean  and  covariance  matrix  of 
each  cluster  are  calculated  to  determine  the  parameters  of  the  approximating  normal  distribu- 
tion. 

The  program  next  considers  all  possible  assignments  of  search  units  to  one  of  the  three 
approximating  probability  distributions.  Since  there  are  a  maximum  of  five  units  and  three  dis- 
tributions, there  are  at  most  35  —  243  different  ways  of  assigning  units  to  distributions.  For 
each  assignment,  the  program  sums  up  the  total  effort  available  to  search  each  distribution  and 
then  computes  the  resulting  optimal  rectangle  and  associated  probability  of  detection.  If  Pk 
denotes  the  conditional  probability  of  detecting  the  target  with  optimal  rectangle  search  given 
that  the  target  has  the  Ath  distribution  (1  <  k  ^  k),  then  probability  of  detection  A  for  the 
allocation  is  given  by 

A  =  i  PkDk. 


The  program  prints  the  allocation  which  gives  the  maximum  probability  of  detection  and 
notes  whether  any  of  the  rectangles  overlap.  If  overlap  occurs,  then  the  next  ranking  allocation 
is  printed,  and  so  on.  This  continues  until  an  allocation  without  overlap  is  found  or  until  the 
top  five  allocations  have  been  listed  together  with  their  associated  detection  probabilities. 
Finally,  when  several  units  are  assigned  to  the  same  rectangle,  it  is  subdivided  in  a  way  which 
preserves  the  uniform  coverage. 

Recently  an  alternative  method  for  multiple  unit  allocation  has  been  developed  (see  Dis- 
cenza  [1])  based  upon  integer  programming  considerations. 

3.   CASP  CASE  EXAMPLE 

On  12  September  1976  the  sailing  vessel  S/V  Spirit  departed  Honolulu  enroute  San  Fran- 
cisco Bay.  The  owner,  who  was  awaiting  its  arrival  in  San  Francisco,  reported  concern  for  the 
vessel  to  the  Coast  Guard  on  14  October  1976  after  it  had  failed  to  arrive.  An  Urgent  Marine 
Information  Broadcast  (UMIB)  was  initiated  on  17  October.  The  following  day,  a  merchant 
vessel  the  M/V  Oriental  Financier  reported  recovering  a  life  raft  with  two  survivors  from  the 
S/V  Spirit  which  had  sunk  in  heavy  seas  in  mid-Pacific  on  the  morning  of  27  September.  Sur- 
vivors indicated  three  more  crewmembers  in  a  separate  raft  were  still  adrift.  This  information 
opened  an  extensive  six  day  air  and  surface  search  for  the  missing  raft  that  eventually  located 
the  raft  with  one  of  the  missing  persons  on  board. 


674  H.R.  RICHARDSON  AND  J  H    DISCENZA 

Each  day's  search  was  planned  utilizing  computer  SAR  programs.  Initial  distress  position 
information  was  gained  by  radio-telephone  debriefing  of  the  survivors  aboard  the  M/B  Oriental 
Financier  on  several  occasions.  The  search  began  19  October  based  on  a  SARP*  datum  for  a 
raft  without  a  drogue  from  an  initial  reported  position  of  36N  136W.  The  second  day's  search 
was  based  on  a  SARP  datum  for  a  position  160  nautical  miles  to  the  northeast  from  the  previ- 
ous position  (this  position  being  determined  from  further  debriefing  of  the  survivors  over 
radio-telephone).  The  third  through  the  six  days'  searches  were  planned  utilizing  CASP  output 
from  a  POSITION  scenario  consisting  of  an  ellipse  with  a  160  mile  major  axis  and  a  60  mile 
minor  axis.  The  CASP  program  was  updated  by  RECTANGLE  and  DRIFT  daily,  and  search 
areas  assigned  to  cover  the  highest  cells  which  could  be  reached  taking  into  account  search  unit 
speed  and  endurance. 

The  following  chronology  is  based  upon  the  official  USCG  report  and  describes  the  utiliza- 
tion of  CASP  in  the  search  planning.  This  case  is  a  good  illustration  of  the  many  uncertainties 
which  must  be  analyzed  during  a  search  and  the  way  both  negative  and  positive  information 
contribute  to  eventual  success. 

21  October  1976 

Search  planning  for  the  day's  operations  utilized  the  CASP  program  for  the  first  time. 
New  probable  distress  position  information  given  by  the  survivors  was  evaluated  and  the  CASP 
program  was  initiated  using  a  POSITION  scenario  with  center  length  160  miles  and  width  60 
miles  oriented  on  046°T,  with  the  southwest  end  at  position  36N  136W.  This  scenario  was  to 
be  used  for  the  rest  of  the  search.  A  search  plan  was  generated  for  the  21  October  search  cov- 
ering approximately  8  of  the  10  highest  CASP  cells  as  given  in  MAP.  Ten  units  were  desig- 
nated for  the  day's  efforts  and  consisted  of  3  Coast  Guard,  2  Navy,  and  4  Air  Force  aircraft  and 
the  USS  Cook. 

The  first  aircraft  which  arrived  on  scene  for  the  day's  search  reported  the  weather  in  the 
search  area  as  ceiling  varying  200-1500  feet  (scattered),  wind  from  330°  at  8  knots,  seas  4  feet, 
and  visibility  unlimited  except  in  occasional  rain  showers. 

At  3:06  PM  an  aircraft  located  what  appeared  to  be  the  life  raft  of  recovered  survivors  in 
position  35-38N  138-12W.  M/V  Oriental  Financier  had  been  unable  to  recover  this  raft  when 
the  survivors  were  rescued.   The  USS  Cook  investigated  and  reported  negative  results. 

Figure  6  shows  the  search  plan  for  21  October.  Note  that  the  target  was  eventually  found 
on  24  October  in  the  first  designated  area  C-l.  There  is,  of  course,  no  way  of  knowing  where 
the  target  was  on  the  21st. 

22  October  1976 

Planning  for  day's  search  was  done  using  updates  from  the  CASP  program.  Search  units, 
consisting  of  17  aircraft  (3  Coast  Guard,  6  Navy,  and  8  Air  Force)  and  the  USS  Cook,  were 
designated  areas  totaling  67,920  square  miles  for  the  day's  effort.  Areas  assigned  were  deter- 
mined from  the  MAP's  twelve  highest  cells.  High  altitude  photographic  reconnaissance  flight 
utilizing  U-2  aircraft  was  also  scheduled,  cloud  coverage  permitting,  to  cover  an  area  of  57,600 
square  miles. 


*A  computer  program  implementing  methods  described  in  the  National  SAR  Manual  and  a  precursor  to  CASP. 


COAST  GUARD  COMPUTER-ASSISTED  SEARCH  (CASP) 


CHARLIE  SEARCH  AREA 


NORTH 
PACIFIC 
OCEAN 


1st  Raft  Recovered 
36-15N  139-23W 
19  0130Z 


2nd  Raft  Recovered 

35-53N      138- 10W 

24    2137Z 


my 


^—Distress  P< 
^  Two 

IN     133- 


"SPIRIT"  Sank 
I"    JApprox.  Position 

36N    136W     27   1900Z 


Hawaiian  Islands 


** 


Tropic  of  Cancer 


2nd  Raft  Recovered 


"SPIRIT"  Sank 


Note: 

POD  is  the  estimated  conditional 

probability  of  detection  given 

the  target  is  in 

the  designated 

area. 

CHARLIE  SEARCH  PLAN 

AREA 

UNIT 

POD 

C-l 

NAVY  P-3 

70% 

C-2 

NAVY  P-3 

70% 

C-3{S)        AF  HC-130 

78% 

(N)        AF  HC-130 

72% 

C-4 

AF  HC-130 

70% 

C-5 

CG  HC-130 

78% 

C-6 

CG   HC-130 

55% 

C-7 

AF  HC-130 

78% 

C-8 

CG   HC-130 

77% 

C-9 

AF  HC-130 

70% 

C-ll 

CG  HC-130 

78% 

Search  plan  based  on  CASP  high 
probability  areas /distress  position  ellipse 


Figure  6.   Search  plan  for  21  October 


676  H.R.  RICHARDSON  AND  J.H.  DISCENZA 

The  first  aircraft  on  scene  for  the  day's  search  reported  the  weather  in  the  general  area  as 
ceiling  1800  feet  (broken),  winds  from  150°  at  6  knots,  seas  2  feet,  and  visibility  15  miles. 

Search  conducted  during  daylight  hours  utilized  15  aircraft,  the  USS  Cook,  and  a  U2  high 
altitude  reconnaissance  flight.  The  USS  Cook  was  unable  to  relocate  debris  sighted  during  pre- 
vious day's  search.  Two  Air  Force  aircraft  failed  to  arrive  on  scene  prior  to  darkness  and  were 
released.  Aircraft  on  scene  searched  88  percent  of  67,920  square  miles  assigned  and  obtained 
POD's  ranging  from  50  to  82  percent.  The  high  altitude  photographic  reconnaissance  flight  was 
conducted  from  an  altitude  of  approximately  50,000  feet. 

The  CGC  Campbell  arrived  on  scene  and  relieved  the  USS  Cook. 

23  October  1976 

The  Rescue  Coordination  Center  (RCC)  was  advised  by  the  Air  Force  that  development 
of  high  altitude  film  had  shown  an  "orange  dot"  in  position  35-16N  139-05W.  The  photo- 
graphed object  was  described  as  a  round  orange  object,  approximately  7  feet  in  diameter,  float- 
ing on  the  surface  of  the  water. 

Search  planning  was  done  using  updates  from  the  CASP  program.  Search  units,  consist- 
ing of  the  CGC  Campbell  and  8  aircraft  (2  Coast  Guard,  3  Navy,  and  3  Air  Force),  were 
assigned  areas  of  highest  CASP  cells.  The  object  photographed  by  reconnaissance  aircraft  was 
drifted  by  SARP  and  the  CGC  Campbell  and  1  aircraft  dedicated  to  locate  it. 

The  first  aircraft  on  scene  for  the  day's  search  reported  weather  in  the  search  area  as  ceil- 
ing 2000  feet,  wind  from  200°  at  12  knots,  seas  2  feet,  and  visibility  15  miles. 

Search  conducted  during  daylight  hours  utilized  8  aircraft  and  CGC  Campbell.  Search 
units  covered  97  percent  of  the  assigned  34,300  square  miles  with  POD's  ranging  from  50  to  92 
percent.  Several  sightings  of  assorted  flotsam  were  reported  but  none  linked  to  Spirit  or  rafts. 
The  object  photographed  by  the  high  altitude  reconnaissance  flight  on  22  October  was  not  relo- 
cated by  search  units. 

Figure  7  shows  the  search  plan  for  23  October.  Although  not  indicated  in  the  chart,  the 
position  where  the  target  was  found  on  the  24th  is  in  the  second  highest  probability  density  cell 
from  the  CASP  map. 

24  October  1976 

Search  planning  for  the  day's  operations  was  done  using  updates  from  the  CASP  program. 
Search  units  consisting  of  the  CGC  Campbell  and  5  aircraft  (2  Coast  Guard  and  3  Navy)  were 
assigned  areas  of  highest  CASP  probability  totaling  18,082  square  miles,  with  CGC  Campbell 
and  one  Coast  Guard  aircraft  designated  for  location  of  the  object  reported  by  the  reconnais- 
sance flight. 

The  position  of  the  reconnaissance  flight  sighting  of  22  October  was  drifted  utilizing 
SARP  and  the  new  position  passed  to  CGC  Campbell  for  search  purposes.  The  11:00  AM 
SARP  datum  was  computed  to  be  35-29. 4N  138-39. 2W  with  standard  first  search  radius  of  16.9 
miles.   The  search  plan  is  shown  in  Figure  8. 


COAST  GUARD  COMPUTER-ASSISTED  SEARCH  (CASP) 


677 


NORTH 
PACIFIC 
OCEAN 


1st  Raft  Recovered 
36-15N   139-23W 
19  0130Z 


\ 


-*f 


Distress  Position  Two 
37-54N  |  133- 36W 


2nd  Raft  Recovered 

35-53N   138- 10W  U. 

24  2137Z  ' 1 


-B 


"SPIRIT"  Sank 
Approx.  position 
36N   136W 
27  1900Z 


UNITED 
STATES 


ECHO  SEARCH  AREAS 


Hawaiian  Islands 


_  Xrpgic  o{_  Cancer _> 


:  Raft  Recovered 


"SPIRIT"  Sank 


ECHO  SEARCH  PLAN 


AREA  UNIT 

E-l  NAVY  P-3 

E-2  NAVY  P-3 

E-3  NAVY  P-3 

E-4  AF  HC-130 

E-5N/S  CG  HC-130 

E-6  AF  HC-130 

E-7  CG  HC-130 

E-8N/S  AF  HC-130 

E-9  CG  HC-130 


POD 

62% 

52% 

50% 

78% 

58/74% 

92% 

64% 

50/60% 

74% 


Search  plan  based  on  CASP  high  probability 
areas,  distress  position  ellipse,  and 
reconnaissance  sighting  (E-l). 


Note:  POD  is 
the  estimated 
conditional  probability 
of  detection  given 
the  target  is  in  the 
designated  area. 


138W 

134W 

Figure  7.   Search  plan  for  23  October 


678 


H.R.  RICHARDSON  AND  J.H.  DISCENZA 


NORTH 
PACIFIC 
OCEAN 


Distress  Position  T 
37-54N   133-36W 

1st  Raft  Recovered      L 

36-15N    139-2  3W  "t 

19  0130Z  \   |\ 


UNITED 
STATES 


x-eu 


2nd  Raft  Recovered 
35-53N   138-10W^- 
24  2137Z    ^ 


Hawaiian  Islands 


D 

b'SPIRIT"  Sank 
Approx.  position 
36N  136W 
27  1900Z 


FOXTROT   SEARCH  AREAS 


Tropic  of  Cance" 
120W 


FOXTROT   SEARCH  PLAN 
AREA  UNIT 


^Raft  Recovered 
35-53N   138- 10W 

F-3           NAVY  P-3 
F-4           NAVY  P-3 
F-5           CG  HC-130 

34N 

F-2 

32N 

F-4 

F-5 

Search  plan  based  on  CASP  high 
probability  area  and  reconnaissance 
sighting  (F-'5). 

135W 

Figure  8.   Search  plan  for  24  October 


COAST  GUARD  COMPUTER-ASSISTED  SEARCH  (CASP)  679 

The  first  aircraft  on  scene  for  the  day's  search  reported  weather  in  the  search  area  as  ceil- 
ing 1500  feet,  wind  from  000°  at  7  knots,  seas  3  feet,  and  visibility  10  miles. 

The  CGC  Campbell  reported  locating  a  rusty,  barnacle  encrusted  55  gallon  drum  in  posi- 
tion 35-27.2N  138-39.0W. 

At  12:05  PM  the  search  met  with  success!  A  Coast  Guard  HC-130H  reported  sighting  a 
raft  in  position  36-03N  138-00W  with  at  least  one  person  on  board.  The  CGC  Campbell  pro- 
ceded  enroute  to  investigate,  and  at  2:37  PM  CGC  Campbell  reported  on  scene  with  the  raft  in 
position  35-53N  138-10W.  A  small  boat  was  lowered  to  recover  the  survivor,  and  at  3:01  PM 
all  search  units  were  released  from  the  scene. 

4.   TRAINING 

CASP  training  began  with  an  operational  testing  phase  in  cooperation  with  the  New  York 
RCC.  This  operational  testing  was  useful  in  orienting  the  personnel  to  the  benefits  derived 
from  more  detailed  search  planning,  and  provided  an  idea  of  what  the  full  training  problem  was 
going  to  be  like. 

Coincident  with  this,  a  training  manual  [9]  and  a  completely  new  combined  operating 
handbook  [10]  were  developed  encompassing  all  of  the  operational  computer  services  available. 

At  the  time  of  official  implementation  in  February  1974,  a  special  four-day  class  was  con- 
ducted in  the  operation  of  the  CASP  system;  this  class  was  attended  by  one  representative  from 
each  Rescue  Coordination  Center.  It  was  intended  that  these  persons  would  learn  the  system 
thoroughly  and  return  to  their  respective  commands  and  teach  others.  This  plan  was  marginally 
successful,  and  worked  only  in  those  cases  where  an  extremely  capable  individual  was  selected 
for  attendance. 

During  the  next  six  months,  personnel  from  the  Operations  Analysis  Branch  visited  each 
East  Coast  RCC  for  one  week  apiece  in  order  to  provide  additional  training.  Subsequently,  the 
same  visit  schedule  was  repeated  on  the  West  Coast. 

Another  valuable  tool  for  training  has  been  telephone  consultation.  Fortunately,  all  mes- 
sages into  and  out  of  the  computer  are  monitored  at  New  York,  and  personnel  can  be  helped 
with  the  details  of  input  and  output  with  a  quick  telephone  call  on  the  spot. 

Finally,  the  National  Search  and  Rescue  School  has  made  CASP  training  a  regular  part  of 
its  curriculum.  The  school,  located  on  Governors  Island,  is  responsible  for  initial  training  of  all 
RCC  personnel  (among  many  others)  in  the  techniques  of  search  and  rescue.  The  present  SAR 
school  training  session  is  four  weeks  in  duration  with  the  fourth  week  devoted  to  computer 
search  planning  systems  training.   Over  half  of  this  time  is  devoted  directly  to  CASP. 

The  Coast  Guard  is  currently  in  the  process  of  separating  its  administrative  and  opera- 
tional systems  by  establishing  an  Operational  Computer  Center.  This  new  Center  will  give  res- 
cue coordinators  direct  access  to  CASP  through  on-line  terminals  and  will  improve  CASP's 
availability  and  reliability.   Interactive  program  control  will  make  the  modules  easier  to  use. 

The  application  of  CASP  in  operational  situations  has  been  quite  successful,  in  spite  of 
significant  encumberances  associated  with  computer  and  communications  services. 


680  H.R.  RICHARDSON  AND  J.H.  DISCENZA 

Continued  oceanographic  research  programs  will  expand  CASP's  applicability  to  important 
in-shore  regions.  Implementation  of  the  new  multi-unit  allocation  algorithm  [1]  is  expected  to 
simplify  the  search  area  assignment  problem.  These  additional  capabilities  coupled  with 
improved  computer  access  and  reliability  should  make  CASP  an  even  more  valuable  planning 
tool  in  the  future. 

ACKNOWLEDGMENTS 

The  development,  implementation,  training,  and  utilization  of  CASP  represents  the  con- 
tributions of  individuals  far  too  numerous  to  mention  by  name  in  this  paper.  Foremost  among 
these  are  the  officers  and  men  who  use  CASP  in  the  RCCs  and  without  whom  the  system 
would  be  useless.  The  contributions  of  the  following  individuals  to  the  support  and  develop- 
ment of  CASP  are  specifically  acknowledged:  C.  J.  Glass,  R.  C.  Powell,  G.  Seaman,  V. 
Banowitz,  F.  Mittricker,  R.  M.  Larrabee,  J.  White,  J.  H.  Hanna,  L.  D.  Stone,  D.  C.  Bossard,  B. 
D.  Wenocur,  E.  P.  Loane,  and  C.  A.  Persinger. 

REFERENCES 

[1]  Discenza,  J.H.,  "Optimal  Search  with  Multiple  Rectangular  Search  Areas,"  Doctoral  Thesis, 

Graduate  School  of  Business  Administration,  New  York  University  (1979). 
[2]  Koopman,  B.O.,  "The  Theory  of  Search,  Part  II,  Target  Detection,"  Operations  Research, 

4,  503-531  (1956). 
[3]  Koopman,  B.O.,  "The  Theory  of  Search,  Part  III,  The  Optimum  Distribution  of  Searching 

Effort,"  Operations  Research,  5,  613-626  (1957). 
[4]  Reber,  R.K.,  "A  Theoretical  Evaluation  of  Various  Search/ Salvage  Procedures  for  Use  with 

Narrow-Path   Locators,    Part    I,    Area   and   Channel   Searching,"    Bureau   of  Ships, 

Minesweeping  Branch  Technical  Report,  No.  117  (AD  881408)  (1956). 
[5]  Richardson,  H.R.,  Operations  Analysis,  February  (1967).    Chapter  V,  Part  2  of  Aircraft 

Salvage  Operation,  Mediterranean,  Report  to  the  Chief  of  Naval  Operations  prepared  by 

Ocean  Systems,  Inc.  for  the  Supervisor  of  Salvage  and  the  Deep  Submergence  Systems 

Project. 
[6]  Richardson,  H.R.  and  L.D.  Stone,  "Operations  Analysis  During  the  Underwater  Search  for 

Scorpion,"  Naval  Research  Logistics  Quarterly,  75,  141-157  (1971). 
[7]  Shreider,  Yu.  A.,  The  Monte  Carlo  Method  (Pergamon  Press,  1966). 
[8]  Stone,  L.D.,  Theory  of  Optimal  Search  (Academic  Press,  1975). 
[9]  U.  S.  Coast  Guard,  Commander,  Atlantic  Area,  CASP  Training  Course,  19-22  February 

(1974). 
[\0]  U.  S.  Coast  Guard,  Computerized  Search  and  Rescue  Systems  Handbook  (1974). 
[11]  U.  S.  Coast  Guard,  National  Search  and  Rescue  Manual  (1970). 
[12]  Wagner,  D.H.  "Nonlinear  Functional  Versions  of  the  Neyman-Pearson  Lemma,"  SI  AM 

Review,  11,  52-65  (1969). 


CONCENTRATED  FIRING  IN  MANY-VERSUS-MANY  DUELS 

A.  Zinger 


University  of  Quebec  at  Montreal 
Montreal,  Canada 


ABSTRACT 


A  simple  stochastic-duel  model,  based  on  alternate  firing,  is  proposed.  This 
model  is  shown  to  be  asymptotically  equivalent,  for  small  hit  probabilities,  to 
other  known  models,  such  as  simple  and  square  duels.  Alternate  firing  intro- 
duces an  interaction  between  opponents  and  allows  one  to  consider  multiple 
duels.  Conditions  under  which  concentrated  firing  is  better  or  worse  than 
parallel  firing  are  found  by  calculation  and  sometimes  by  simulation.  The  only 
parameters  considered  are  the  combat  group  sizes  (all  units  within  a  group  are 
assumed  identical),  the  hit  probabilities  and  the  number  of  hits  necessary  to 
destroy  an  opposing  unit. 


1.   INTRODUCTION 

Two  extremes  for  the  modeling  combat  attrition  are  given  by  the  so-called  Lanchester 
theory  of  combat,  which  treats  combat  attrition  at  a  macroscopic  level,  and  by  the  theory  of  sto- 
chastic duels,  which  treats  combat  attrition  at  a  microscopic  level  and  considers  individual  firers, 
target  acquisition,  the  firing  of  each  and  every  round,  etc.  (see  Ancker  [1,  pp.  388-389]  for 
further  details).  Actual  combat  operations  are,  of  course,  much  more  complex  than  their 
representation  by  such  relatively  simple  attrition  models  and  may  also  be  investigated  by  means 
of  much  more  detailed  Monte  Carlo  combat  simulations.  Unfortunately,  such  detailed  Monte 
Carlo  simulations  usually  fail  to  provide  any  direct  insights  into  the  dynamics  of  combat  without 
a  prohibitive  amount  of  computational  effort.  In  the  paper  at  hand,  we  will  consider  a  relatively 
simple  stochastic-duel  model  to  develop  some  important  insights  into  a  persisting  issue  of  mili- 
tary tactics  (namely,  what  are  the  conditions  under  which  concentration  of  fire  is  "beneficial"). 

In  his  now  classic  1914  paper,  F.W.  Lanchester  [10]  (see  also  [11])  used  a  simple  deter- 
ministic differential-equation  model  to  quantitatively  justify  the  principle  of  concentration,  i.e., 
a  commander  should  always  concentrate  as  many  men  and  means  of  battle  at  the  decisive  point. 
From  his  simple  macroscopic  model,  Lanchester  concluded  that  the  "advantage  shown  to  accrue 
from  fire  concentration  as  exemplified  by  the  n  square  law  is  overwhelming."  However,  this 
conclusion  depends  in  an  essential  way  on  the  macroscopic  differential-equation  attrition  model 
used  by  Lanchester  [10],  [11]  (see  Taylor  [14]  for  further  discussion)  and  need  not  hold  for 
microscopic  stochastic-duel  models  of  combat  attrition.  In  fact,  this  paper  shows  that  for  such 
microscopic  duel  models  it  is  not  always  "best"  to  concentrate  fire. 

Subsequently,  many  investigators  have  commented  on  the  benefits  to  be  gained  from  con- 
centrating fire.  For  example,  in  his  determination  of  the  probability  of  winning  for  a  stochastic 
analogue  of  Lanchester's  original  model,  Brown  [6]  stressed  the  fact  that  the  model  applied  to 

681 


682  A    /IN(,I  K 

cases  of  concentrated  firing  by  both  sides.  Other  investigators  of  deterministic  Lanchester-type 
models  from  the  macroscopic  combat-analysis  point  of  view  have  also  stressed  this  point  (e.g. 
see  Dolansky  [7],  Taylor  [13],  and  Taylor  and  Parry  [15]).  Recently,  Taylor  [14]  has  examined 
the  decision  to  initially  commit  forces  in  combat  between  two  homogeneous  forces  modeled  by 
very  general  deterministic  Lanchester-type  equations.  He  showed  that  it  is  not  always  "best"  to 
commit  as  much  as  possible  to  battle  initially  but  that  the  optimal  decision  for  the  initial  com- 
mitment of  forces  depends  on  a  number  of  factors,  the  key  of  which  is  how  the  trading  of 
casualties  depends  on  the  victor's  force  level  and  time. 

The  first  reference  to  problems  of  strategy  in  multiple  duels  is  found  in  Ancker  and  Willi- 
ams [2],  who  study  the  case  of  a  square  duel  (2  vs  2)  and  arrive  at  the  right  conclusion  that 
parallel  firing  is  better  than  concentrated  firing.  This  is  a  natural  conclusion  since  only  one  hit 
is  necessary  to  achieve  destruction,  and  in  concentrated  firing  there  is  a  certain  amount  of 
over-killing.  In  1967,  Ancker  [1]  makes  suggestions  for  future  research  concerning  mutliple 
duels  and  states  explicitly  that  the  difficulties  lie  in  the  strong  interaction  between  the  contes- 
tants. The  possibility  of  needing  more  than  one  hit  to  achieve  destruction  in  the  simple  duel 
situation  was  introduced  by  Bhashyam  [4]  in  1970. 

The  purpose  of  this  paper  is  to  combine  some  of  the  above  mentioned  concepts,  in  order 
to  gain  insight  concerning  a  problem  of  strategy  in  multiple  duels— should  one  concentrate 
one's  fire  or  not? 

2.   ASSUMPTIONS  AND  NOTATION 

Let  us  consider  two  forces  A  and  B  that  meet  each  other  in  combat.  A  consists  of  M  units 
and  B  of  N  units. 

The  following  assumptions  are  made: 

1.  Firing  is  alternating,  volley  after  volley,  i.e.,  A  fires  all  weapons  simultaneously,  then  B 
and  so  on  until  all  units  of  a  force  are  destroyed.  This  is  contrary  to  the  usual  assumption  of 
either  simultaneous  firing  or  random  firing  within  some  time  intervals  as  found  in  Robertson 
[12],  Williams  [17],  Helmbold  [8],  [9],  Thompson  [16],  Ancker  [3].  It  is  felt,  and  will  be 
shown  in  a  few  cases,  that  for  relatively  small  probabilities  of  hitting,  this  approach  gives  results 
comparable  to  Ancker  and  Williams  [2].  We  will  denote  by  Vt\j  the  probability  of  /  winning  if  j 
shoots  first  /',  j  =  A,B.   The  unconditional  probability  of  winning  will  be  denoted  by  VA  or  VB. 

2.  Hit  probabilities  are  constant  and  are  respectively  pA  and  pB,  with  q,  =  1  —  p,,  /'  =  A,B. 

3.  Each  unit  of  force  A  requires  KA  hits  to  be  destroyed.   Same  for  B  and  KB. 

4.  The  supply  of  ammunition  is  unlimited. 

5.  There  is  no  time  limit  to  score  a  hit. 

6.  In  a  multiple  duel  (more  than  1  vs  1)  the  units  of  A  concentrate  their  fire  on  a  single 
unit  of  B  while  the  units  of  B  each  fire  at  a  different  unit  of  A,  or  spread  their  fire  over  all 
available  units  of  A,  this  last  case  occurs  when  M  <  N.  B  has  to  allow  an  amount  of  concen- 
tration in  order  not  to  lose  some  shots.  Concentration  will  be  kept  at  a  minimum  to  preserve 
as  much  parallelism  as  possible.  For  example  if  M  =  3  and  N  =  7  the  pattern  of  fire  for  B  has 
to  be 


CONCENTRATED  FIRING  IN  MANY-VERSUS-M  ANY  DUELS 


7.  The  most  general  notation,  for  example,  VA\B(M,N,KA,KB,pA  pB)  will  be  avoided  if 
possible  and  replaced  by  an  appropriate  simpler  form. 

Before  proceeding,  a  general  remark  ought  to  be  made:  most  of  the  difficulties  come  from 
the  asymmetry  in  the  situation  and  from  the  interaction  between  the  opponents.  The  same 
model  has  to  express  concentration,  dispersion  and  partial  concentration  of  fire.  Moreover,  the 
probability  of  winning  depends  upon  the  whole  past  history  of  the  duel. 

3.   MULTIPLE  DUEL.   ONE  HIT  SUFFICIENT  TO  DESTROY 


Let  KA  - 


1  and  let  E(i,j)  be  the  state  of  group  A  with  i  units,  and  of  group  B  with 


If  A  fires  first,  the  next  state  is 

E(i,j)  with  probability  q'A  and 
E(i,j  —  1)  with  probability  1  —  q'A. 

When  B  fires,  let  us  first  consider  the  case  when  j  <  i.  Then, 

E(i,j)  becomes  E(i  -  k,j),  k  =  0,  . . .  ,  j  with  probability     ,  \pB  q 


E(i,j  —  1)  becomes  E(i  —  k,  j  —  1),  k  =  0, 


1  with  probability 


If  on  the  other  hand  j  >  /  some  regrouping  has  to  be  done. 

Let  j  =  ai  +  b  with  b  <  r,  a,  b  €  I+.   The  regrouping  which  spreads  the  fire  the  most  is 
given  by 

a  shots  are  fired  with  a  probability  of  success 

1  —  (1  —  pB)a  =  1  —  A0  at  each  of  /  —  b  targets 

a  +  1  shots  are  fired  with  a  probability  of  success 

1  -  (1  -  pB)a+l  —  1  —  Ax  at  each  of  b  targets. 

Define  r  =  min(ij).   Then  both  cases  j  <  /and  j  ^  /are  identical  if  one  defines  the  probabil- 
ity of  transition  from  state  E(i,j)  to  state  E(i  —  k,  j)  when  B  fires  as 

(  r  -  b)  kx     b  -  k\  k0     r  -  b  -  k0 

*  (l-^l)         A  (1   —   ^0>         A 

*o  Ax  AQ 


(3.1)    0(i,j,k,pB)=         X 

*o=0,l....r- 
*.=0.  1 


In  the  case  j  <  i,  a  =  0,  k0  =  0  and  k\  =  k. 


684  A    ZINGKR 

It  follows  that  if  A  starts  and  B  returns  fire  once,  the  intial  state  Ed,  j)  can  become 
Ed,  j)  with  probability  qA  0  (/,  j,  0,  pB)  =  qA  qfc 
Ed  -  k,  j)  with  probability  qA  0  (/',  j,  k,  pB),  k  =  1,  ....  r 
E(i  -  kj  -  1)  with  probability  (1  -  q'A)  0  (/,  j  -  1,  k,  pB),  k  =  0, 1,  . . .  ,  r' 
where  r'  =  min  (/,  j  —  1). 

If  B  starts,  the  initial  state  E(i,  j)  can  become 

E(i,  j)  with  probability  qA  0  (/',  j,  0,  pB)  =  qA  qB 
E(i  —  k,  j)  with  probability  qA~k  0(i,j,k,pB),  k  =  1,  . . .  ,  r 
E(i  -  k,j  -  1)  with  probability  (1  -  q'Ak)  0  d,  j,  k,  pB),  k  =  0, 1,  ....  r" 
where  r"=  min(/  —  1,  j). 

Let  VB\A(M,  AO  denote  the  probability  that  group  B  wins  with  initial  state  E(M,  N)  and 
A  starts  firing.  Then 

(3.2)  VBU  (M,  N)  =  q?  q$  VB{A  (M,  N) 

+  q%  £  0  (M,  N,  k,  pB)  VBU{M-k,  N) 
k=\ 

+  (1  -  Qa)  £  9  (M,  N  -  1,  k,  pB)  VBlA  (M-  k,  N-  1). 

A:=0 

This  corresponds  to  a  decomposition  into  all  the  mutually  exclusive  and  exhaustive  ways  for  B 
to  win  if  A  fires  once  and  then  B  returns  fire. 

In  a  similar  way 

(3.3)  VB]B(M,  N)  =  qjfqg  VBlB  (M,  N) 

+  Z  QA~k  9  (M,  N,  k,  pB)  Vb\b(M  -  k,  N) 
k=\ 

+  Z  (1  -  qH?~k)  9(M,  N,  k,  pB)  VB\B(M  -  k,  N  -  1). 

k=Q 

Since  we  have 

VBU  (M,  0)  =  VB\B(M,  0)  =  0        all  M 

and 

Vb\a®,  N)=  VBlB(Q,N)=  1       all  AT 

we  can  calculate  in  succession  all  required  probabilities.  For  example,  since  0  (1,1,1,  pB)  =  pB, 
one  finds  VB\A{\,  1)  =  qApBl  (1  -  qAqB).  Using  ^^(1,1)  and  0(1, 1,0,  pB)  =  qB, 
0(\,2,\,pB)  =  1  -  qh  one  finds  VB[A(l,2). 

Explicitly,  one  gets,  by  assuming  that  A  starts  half  the  time, 
1 


=  \pb4<m-x)I1  (i  +  mi  n  (i  -  ftci). 


CONCENTRATED  FIRING  IN  M ANY-VERSUS-M ANY  DUELS 


One  can  also  obtain  for  qA  =  qB  =  q 


v  (2   2)  =    l  +  4q  +  Aql  +  lq3  +  Aq*  +  3qS  +  q6 
B     '  2(\  +  q)2  (\  +  q2)  (\  +  q  +  q2) 


A  comparison  with  the  triangular  duel  and  the  first  square  duel  [2]  for  p  — »  0,  q  — •  1 


VB{2,\)  = 


plq(\  +  q2) 


q(\  +  q2) 


►  1/6 


and  VB(2,2)- 
in  [2]. 


2    (l_^2)(1_93)         2(1  +  q)(l  +  q  +  q2)     f-1 
»l/2  which  are  the  same  limits  as  the  one  obtained  from  Equation  29  and  37 


Table  1  gives  some  results  for  VB(M,  N,  pA,  pB). 

TABLE  1  -  (x  104) 


M 

N 

Pa        0.3 

0.3 

0.5 

0.5 

0.7 

0.5 

0.7 

Pb        0.3 

0.5 

0.3 

0.5 

0.5 

0.7 

0.7 

1 

1 

5000 

6538 

3462 

5000 

3824 

6176 

5000 

2 

2 

5166 

7307 

3100 

5317 

3850 

6873 

5447 

3 

3 

5678 

8227 

3405 

6418 

5081 

8343 

7386 

3 

5 

9634 

9982 

8869 

9913 

9805 

9998 

9994 

5 

3 

1292 

3806 

0368 

1780 

0997 

3907 

2832 

5 

5 

7258 

9614 

5118 

8940 

8359 

9920 

9848 

5 

7 

9831 

9999 

9422 

9994 

9986 

10000 

10000 

7 

5 

3418 

7843 

1626 

6060 

5075 

9090 

8629 

7 

7 

8850 

9978 

7538 

9919 

9853 

10000 

10000 

10 

10 

9900 

10000 

9708 

10000 

10000 

10000 

10000 

with  M.   We  conclude:  Parallel  firing  is  better. 

No  simple  relationship  exists  in  the  case  pA  ^  pB.    Neither  MpA  vs  NpB,  nor  M2pA  vs 
N2pB  are  sufficient  to  decide  if  VB  >  —. 


4.  SIMPLE  DUEL.   K  HITS  NECESSARY  TO  DESTROY 

Let  M  =  N  =  \  and  let  VB\A  (KA,KB)  denote  the  probability  that  B  wins  the  simple  duel 
if  A  starts  firing  and  KA  hits  are  necessary  to  destroy  A  and  KB  for  B. 

It  is  evident  that 

Vb\a(Ka,Kb)  =  pA  VB]B(KA,  KB-l)  +  qA  VBlB(KA,  KB) 


VBlB(KA,  KB)  =  pB  VBlA(KA  -  1,  KB)  +  qB  VBlA(KA,  KB). 


686  A   ZINGER 

This  gives 

(4.1)  (1  -  qA  qB)  VBU(KA,  KB)  -  pApB  VBlA(KA  -\,KB-  1) 

-  Pa  Qb  VbU{Ka,  KB-\)-  qApB  VB\A{KA  -  1,  KB)  =  0. 

In  order  to  solve  this  difference  equation,  following  Boole  [5] ,  let  us  define 
x  -  KA,  y  -  KB 
ux,y=  VBUKx  -  1,  y  -  1) 
Dxu  =  ux+Xy  and  Dyu  =  uxy+\. 

Substituting  these  into  Equation  (4.1)  we  get 

Kl  -  Qa  Qb)Dx  Dy  -  pApB-  pA  qB  Dx  -  qA  pB  Dy]u  -  0. 

Let  Dy  =  a. 

((1  ~  Qa  Qb)<*  ~  Pa  Qb)Dxu  -  pB(aqA  +  pA)u 

which  gives 

«  =  PbQ>a+  Qa  Dy)x[(l  -  qA  qB)Dy  -  pA  qB]~x  0  (y) 

where  0(y)  is  arbitrary.  Then, 

«  =  dt  Pa  QV  dA  (1  ~  Qa  Qb)~x  D~x  [l  -   j  gg^  dA      0(y). 

Since  D~x0(y)  =  0(y  -  x)  and 

,)*      -L  I  •*+./'  — l|  (     P^  Qtt     Y 


l-  QA<i 

we  get 


.  p*    1  ziur  ■   lL+'«r'«*  (i  -  ^  to)-y  %  -  /  - ;). 

1  -  «*  Qb  J   £o  j-o  I  'J  I         ^       i 


Taking  into  account  that 

PbQa 


VB\A(hD=- 


[~  QaQb 
a  good  choice  for  0(t)  is 

0{t)  =  1  if  t  >  0 
=  0  if  r  <  0. 

Defining  r  =  min  (A^,  KB  —  1)  the  solution  becomes 

,^-1(^)1^+7-1)  ^     Jfc.-I  _^ 

(4.2)     VB{A{KA,KB)  =  \?     £  /  /  J**  A     Ci  QbU-QaQb)     aJ 

with 

W^  0)  =  0  and  KflU(0,  KB)  =  1. 


CONCENTRATED  FIRING  IN  M  ANYVERSUS-MANY  DUELS 

One  can  verify  by  substitution  that  this  is  a  solution. 
One  can  evaluate  the  other  probabilities  of  winning  by 

va\b^a,  kb>  Pa.  Pb)  =  vb\a^b,  &a>  Pb.  Pa). 

VB\B(KA,  KB,  pA,  pB)  =  1  -  VA\B(KA,  KB,  pA,  pB), 

Va\a(Ka,  Kb,  pA,pB)  =  1  -  VBU(KA,  KB,  pA,pB). 

Table  2  gives  some  results  for  VB\A  (KA,  KB,  pA,  pB)  and  VB\B(KA,  KB,  pA  pB). 
TABLE  2  -  (x  104) 


687 


Pa  =  -3 

Pb=  -5 

Pa  = 

Pb='-5 

Pa  =  -5, 

Pb  =  -1 

Ka 

KB 

VB\A 

Vb\b 

Vb\a 

Vb\b 

Vb\a 

VB\B 

1 

1 

5385 

7692 

3333 

6667 

4118 

8235 

5 

3 

4257 

5010 

1139 

1728 

2576 

3579 

5 

5 

8201 

8630 

4512 

5488 

7414 

8381 

7 

5 

5955 

6541 

1674 

2266 

4159 

5278 

7 

7 

8695 

8981 

4599 

5401 

7981 

8669 

10 

10 

9160 

9330 

4671 

5329 

8545 

9002 

This  table  indicates  that  VB  =  1/2  if  KA  =  KB  and  pA  =  pB=  1/2,  VB  increases  towards  1 
if  KA  =  KB  and  pB  >  pA  and  |  VB\A  -  VB\B\  decreases  if  KA  and  KB  increase. 

An  interesting  comparison  is  to  be  made  with  the  results  given  by  Bhashyam  [4].  Under 
an  assumption  of  an  exponential  distribution  for  interfiring  times  he  finds  that  the  probability  of 
B  winning  is,  using  our  notation, 

P(B)=  1  -  /   Pa     (Kb,  Ka) 
Pa+Pb 
where  Ix  is  the  incomplete  Beta  function.   The  correspondance  in  the  notations  being  \p  for  pA , 
A  *p*  for  pB,  R  for  KB  and  R  *  for  KA . 

Table  3  shows  at  what  rate  a  model  with  alternate  firing  converges  towards  Bhashyam's 
model. 

Alternate  firing  gives  a  good  approximation  if  p  is  small.  In  fact,  consider  KA  and  KB 
fixed  and  pA  =  c  pB  with  pB  ~*  0. 


One  can  show  that 


lim  VB\A  - 


1 


(1  +  c)^     ~o 
and  this  limit  from  a  well  known  theorem  is 
1  -  I  c    (KB,KA). 


KA+j-\ 


1  +  c\ 


=  lim  VB\B 


TABLE  3  -  Rate  of  C 

onvergence  of  VB 

to  P(B) 

Pa 

Pb 

Ka 

KB 

vB 

P(B) 

0.4 

0.2 

5 

5 

0.1054 

0.2 

0.1 

0.1265 

0.02 

0.01 

0.1431 

0.002 

0.001 

0.1447 

0.1449 

0.1 

0.2 

5 

2 

0.3391 

0.01 

0.02 

0.3501 

0.001 

0.002 

0.3511 

0.3512 

0.1 

0.2 

10 

10 

0.9491 

0.01 

0.02 

0.9366 

0.9352 

5.   SQUARE  DUEL.   2  HITS  NECESSARY  TO  DESTROY 

Let  M  —  N  =  2  and  KA  =  KB  =  2.  One  can  represent  the  state  of  the  two  forces  by  (;b 
h>  J\>  Jt)  with  i\,  h>  J\»  h  =  0. 1.2,  representing  the  number  of  hits  necessary  to  destroy.  For 
example,  (1,  1;  0,  2)  means  that  A  has  2  units  that  can  be  destroyed  by  one  hit  each  and  B  has 
one  unit  that  has  been  destroyed  by  2  hits  and  one  unit  untouched. 

All  attempts  to  arrive  at  one  or  two  difference  equations  have  been  in  vain.  Two 
equivalent  approaches  have  been  used.  In  the  first,  taking  pA  =  pB  =  1/2,  and  defining  A,  as 
the  matrix  of  the  transitional  probabilities  corresponding  to  the  case  when  A  fires  first,  and  B 
the  corresponding  matrix  when  B  fires  first  one  obtains: 

VA  by  summing  all  the  probabilities  for  the  events  (/,  j\  0,  0)  in    lim    (AB)n  and  VB  by 

„->oo 

summing  all  the  probabilities  for  the  events  (0,  0;  i,  j)  in    lim    (BA)n. 


The  matrices  are  29  x  29.  The  possible  states  of  A  are  such  that  i\  ^  i2.  The  possible 
states  of  B  are  such  that  jy  <  j2  and  exclude  jx  =  j2  =  1  since  A  concentrates  its  fire  until  des- 
truction is  achieved. 

Assuming  the  ordering  i\  >  /'2,  two  variations  are  possible.    In  Case  1,  when  the  state  is 
(2,1;  0,  j)  with  j  =  1  or  2  and  B  fires,  B  chooses  at  random  among  the  two  units  of  A.   In  Case 
2,  B  fires  on  the  second  unit  of  A,  which  can  be  destroyed  by  one  shot.   We  find 
In  Case  1     VA  =  0.5586 
and  in  Case  2     VA  =  0.5396 
In  both  cases  concentrated  firing  is  better. 

The  other  approach  consists  in  writing  down  all  the  equations  that  define  the  battle.  For 
example, 

VAlA(2,2-\,2)  =  (1  -  flD  ^|fl(2,2;0,2)  +  q}  VAlB(2,2-\,2). 

The  difference  between  Case  1  and  2  is  seen  by  considering 

^|S(2,l;0,l)  =  0.5/7flF^M(l,l;0,l)+0.5pfl  ^ (2, 0;0, 1) 

+  qBVA]A(2,\-0,\) 


CONCENTRATED  FIRING  IN  MANY-VERSUS-MANY  DUELS  689 


VAlB(2, 1;0, 1)  -  pB  VAU(2, 0;0. 1)  +  qB  VaU{2,  I;0, 1). 

A  third  variation  is  possible  in  which  no  ordering  is  assumed  for  the  iKs.  Only  the  states  with 
ix  =  0  are  eliminated.  In  this  case,  B  fires  always  upon  the  last  unit  of  A  but  2  states  are  con- 
sidered 

VA]B(2, 1;0, 1)  =  pB  K^(2.0;0. 1)  +  qB  VAlA(2,  1,0, 1) 
and 

VMB{\.  2;0, 1)  =  pB  VA]A  (1, 1;0, 1)  +  qB  VaU  (2, 1;0, 1). 
In  this  case  VA  =  0.5553  for  pA  =  pB  =  0.5. 

The  total  system  consists  of  35  pairs  of  equations  and  is  solved  by  iterations. 

Table  4  gives  some  results  for  the  square  duel  in  this  last  case.  As  in  the  two  preceeding 
cases,  concentrated  firing  is  better. 

An  extension  of  this  last  case  is  considered  in  the  next  section. 

6.  MULTIPLE  FAIR  DUELS.   K  HITS  NECESSARY  TO  DESTROY 

Let  us  restrict  ourselves  to  the  case  of  a  fair  duel,  i.e.,  one  such  that  M  =  N  =  n, 
Pa=  Pb  =  P  and  KA  =  KB  =  K. 

All  nondestroyed  units  of  A  concentrate  their  fire  on  a  single  unit  of  B,  volley  after  volley 
until  destruction  is  achieved.  For  the  next  volley  they  concentrate  their  fire  on  the  next  undes- 
troyed  unit  of  B. 


There  are  nK  +  '. 

1  possible  states  for  B 

K,    K, 

..,  K 

K  -  1   K, 

..,  K 

1,    K, 

..,  K 

0,    K, 

...,  K 

On  the  other  hand  B  spreads  its  fire  over  all  units  of  A  and  all  states  are  possible,  eliminating 
only  the  destroyed  units. 

Since  there  are  K"~j  different  states  with  j  zeros  the  number  of  possible  states  for  A  is 

This  means  that  in  order  to  find  VA (K,  ...  ,  K;K,  ...  ,  K)  we  will  have  to  solve  a  linear  sys- 
tem consisting  of  {nK  +  1)  (Kn+1  -  l)/(K  -  1)  pairs  of  equations  of  the  form 

Va  \a  (state)  =  linear  combination  of  VA  \B  (outcome  of  A  firing) 

va\b  (state)  =  linear  combination  of  VA\A  (outcome  of  B  firing). 


TABLE  4  -  (x  104)  Square  Duel 


Number 

of 

Pa  =  Pb  =  P 

Va\a 

VA\B 

Va 

Iterations 

0.999 

9980 

40 

5010 

3 

0.99 

9809 

382 

5096 

3 

0.95 

9193 

1599 

5396 

4 

0.9 

8666 

2573 

5620 

5 

0.7 

7573 

3850 

5712 

8 

0.5 

6781 

4324 

5553 

13 

0.3 

6117 

4786 

5452 

25 

0.1 

5598 

5198 

5398 

80 

0.05 

5487 

5292 

5389 

157 

0.025 

5434 

5337 

5385 

303 

0.02 

5423 

5346 

5385 

373 

0.01 

5402 

5364 

5383 

844 

Unfortunately,  the  number  of  possible  states  increases  very  rapidly.    A  few  values  are 


Number  of  States 

n  =  2 

35 

3 

105 

4 

279 

5 

693 

6 

1651 

n  =  2 

91 

3 

400 

4 

1573. 

This,  however,  is  much  better  than  (K  +  l)2",  which  is  the  number  of  possible  states  without 
any  restrictions. 

Since  writing  down  the  necessary  equations  is  an  impossible  task,  a  computer  program  was 
written  to  build  the  equations  and  solve  them  by  iteration.  The  main  steps  are: 

(1)  define  the  necessary  states, 

(2)  define   VA  \A  =  0    for  all  states 

Va  \b  =  0    f°r  aN  states  if  B  is  not  destroyed 
va  \b  =  1    if  5  is  destroyed. 

These  will  be  the  initial  conditions. 


(3)  For  each  state  determine  the  number  of  effective  units  MA  and  NB.  If  A  fires,  the 
number  of  targets  is  T  =  1  and  the  degree  of  concentration  is  c  =  MA .  If  B  fires,  the  number 
of  targets  is  T=  min(MA,  NB).  If  MA  ^  NB,  the  degree  of  concentration  is  c  =  1  and  if 
NB  >  MA,  then  NB  =  a  MA  +  b  and  cx  =  a  for  Tx  =  MA-  b  units  and  c2  =  a  +  1  for  T2  =  b 
units. 


CONCENTRATED  FIRING  IN  MANY-VERSUS-M  ANY  DUELS  6V1 

(4)  Let  Qc(i,  j)  denote  the  probability  for  a  unit  to  go  from  state  K  =  /  to  state  K  =  j  if 
submitted  to  fire  of  concentration  c.  Then  the  matrix  Q2,  for  example,  has  the  form 


0 

1         2                  tf 

1 

2 
K 

l-q2 

<72 
2pq      q* 

\         x           \ 

V    X2W     V 

In  general,  for  /  =  1,  K  and  j  =  0, 1,  . . .  ,  AT 

(/  ^  J  Z''"7  <7c-'+y   for  y  ^  0 
1-  £&('.  7')    fory  =  0. 


aa  y)  = 


All  required  matrices  are  constructed. 

5)  For  each  state  the  equation  giving  VA  \A  is  constructed. 

Let  ;'  denote  the  state  of  the  target  unit. 

Let  j  denote  the  states  of  this  unit  after  A  has  fired,  the  rest  of  B  being  unaffected. 

Then, 

VaiaUm)-  J:QMaU,j)  Va\bUJ) 

the  corresponding  equation  for  VA\B  is  of  the  general  form 

VAlB(ihi2,    ....    iT-B)  =  £     n   Qce  (4  Je)\    Va\A<JIs,    ...,JTS\B). 


For  example, 

K4|5(l,2,0,0,0;l,2,2,2,2)=    £     02(U,)  Q3(2j2)  VAU OiJ2.0f0f0;lf  2, 2, 2, 2). 

y  1-0.1 


6)  When  all  possible  states  are  gone  through,  the  last  calculated  value  is 
VAlB(K,K,  ....  /s:;/r,  ...,  K). 

It  is  compared,  usually  within  10-6,  to  the  previously  calculated  value  and  the  process  is 
iterated  until  convergence  is  achieved. 

Table  5  gives  results  for  several  values  of  M  and  K.  The  dimension  of  the  linear  system 
is  twice  the  number  of  states.  The  probability  of  a  hit  is  taken  as  p  =  0.5.  Time  is  given  for 
some  cases.  The  computer  used  was  a  CDC6400. 

The  value  p  =  0.5  was  chosen  because  time  increases  very  fast  if  p  decreases,  as  is  seen 
from  Table  4. 


TABLE  5 

-  (K^  x  104),  Multiple  Fair  duel.  Exact  Results 

Number 

Number 

Ti 

M=  N 

K 

of 
Equations 

of 
Iterations 

Va 

(in  seconds) 

2 

2 

70 

13 

5553 

3 

182 

15 

5988 

4 

378 

17 

6364 

5 

682 

19 

6661 

3 

2 

210 

13 

5537 

3 

800 

15 

6211 

4 

2210 

17 

6822 

5 

4992 

19 

7289 

1700 

4 

2 

558 

13 

5152 

3 

3146 

15 

6132 

4 

11594 

17 

6872 

6141 

5 

2 

1386 

12 

4429 

3 

11648 

15 

5931 

8960 

Since  exact  calculations  of  VA  become  too  time  consuming,  some  results  were  obtained  by 
simulation.  Table  6  gives  some  results.  The  number  of  trials  was  2000  for  p  ^  0.5  and  6000 
for  p  =  0.5,  A  started  the  duel  in  half  the  cases. 

TABLE  6  -  (VA  x  103), 
Multiple  Fair  Duel.  Simulation  Results 


p 

0.1 

0.3 

0.5 

0.7 

0.9 

M 

K 

2 

2 

550 

542 

561 

569 

564 

4 

574 

562 

522 

485 

403 

6 

583 

490 

351 

148 

4 

8 

580 

382 

120 

2 

0 

10 

562 

249 

11 

0 

0 

2 

3 

600 

600 

611 

622 

575 

4 

652 

632 

599 

628 

642 

6 

672 

606 

564 

494 

270 

8 

705 

554 

444 

181 

2 

10 

725 

482 

233 

6 

0 

2 

4 

586 

616 

642 

691 

778 

4 

715 

694 

685 

666 

722 

6 

774 

700 

653 

672 

651 

8 

796 

684 

608 

548 

205 

10 

797 

643 

524 

204 

1 

2 

5 

630 

646 

674 

708 

705 

4 

754 

753 

760 

748 

556 

6 

812 

786 

725 

665 

852 

8 

838 

777 

668 

698 

601 

10 

878 

740 

639 

594 

134 

CONCENTRATED  FIRING  IN  MANYVERSUS-MANY  DUELS  693 

We  note  that  for  large  values  of  p  the  behaviour  of  VA  is  erratic.  This  is  due  to  the  deter- 
ministic issue  of  a  battle  for  p  =  1  as  a  consequence  of  alternative  firing.  For  example,  if 
M  =  6,  k  =  2  and  A  starts  firing,  the  sequence  of  states  is  B.  022222,  A:  211111,  B:  002222,  A: 
210000,  B:  000222,  B  wins. 

Two  independent  estimates  of  the  error  can  be  made;  one  by  comparing  the  results  of  the 
simulation  with  the  calculated  values  in  Table  5  for  p  =  0.5,  M  =  N  =  2  or  4  and  K  =  2,3,4, 
giving  5  =  0.0093,  the  other  estimate  is  given  by  assuming  a  binomial  distribution  with  6000 
trials  giving  s  =  0.0065.  To  be  on  the  safe  side  one  can  conclude  that  concentrated  firing  is 
better  if  the  simulation  gives  VA  ^  0.519  and  parallel  firing  is  better  if  the  simulation  gives 
VA  ^  0.481.  This  does  not  take  into  account  the  bias  introduced  by  alternate  firing  for  "large" 
values  of  p.  Since  the  sign  of  the  bias  is  evident,  one  can  adjust  one's  conclusions,  for  example 
for  M  =  10,  K  =  4  and  p  =  0.5  the  observed  value  0.524  is  pulled  down  and  almost  certainly  A 
wins  more  often  than  B.  On  the  other  hand  for  M  =  8,  K  =  3  and  /?  =  0.5  the  value  0.444  is 
certainly  pulled  down  and  one  can  hardly  conclude  that  B  wins  more  often. 

Table  7  summarizes  all  the  results  obtained. 

TABLE  7  —  Better  Strategy  of  Firing 


Concentrated 

Parallel 

Border  cases 

/>  =  0.1 

K  >  2 

K=  1 

/>  =  0.3 

K=  3 

2  ^  M  <  4 

tf=  2    M^l 

tf=2    M=5or6 

K  =  3 

2  <  M  ^  8 

#  =  3    m  =  9  or  10 

tf  =  4,5 

2  <  M  ^  at  least  10 

/>  =  0.5 

K  =  2 

2  <  M  ^  3 

K=2    M  >  5 

#  =  2    M=  4 

K=3 

2  <  M  ^  6 

/<:  =  3    M  ^  7 

K  =  4 

2  <  M  ^  10 

K=5 

2  <  M  <  at  least  10 

One  can  conclude  that  concentrated  firing  is  better  if  the  combination  of  group  size  and 
hit  probability  does  not  produce  a  high  degree  of  overkilling.  For  K  >  2  a  rough  rule  could  be 
concentrate  firing  if  pM  ^  K  (the  exception  is  p  =  0.5,  #  =  4  and  M  =  9  or  10). 

Up  to  this  time  we  have  compared  two  strategies:  parallel  firing  and  concentrated  firing. 
In  the  next  section  we  will  attempt  to  define  the  concept  of  partial  concentration. 

7.   MULTIPLE  FAIR  DUELS.   PARTIAL  CONCENTRATION  OF  FIRE 

Let  M  =  N  =  n,  pA  =  pB  =  p  and  KA  =  KB  =  K.  Let  cx  be  the  maximal  number  of  non- 
destroyed  units  of  A  that  are  allowed  to  concentrate  their  fire  on  a  single  unit  of  B,  volley  after 
volley  until  destruction  is  achieved. 


firing. 


If  cx  =  I,  A  uses  parallel  firing  in  the  same  manner  as  B.   If  cx  =  n,  A  uses  concentrated 


Under  partial  concentration  the  number  of  targets  for  A  is  given  by  the  integer  function 

f  n  +  cx  - 


and  the  number  of  possible  states  for  B  is 


(»-  ta  +  D*  "  +  - 

which  reduces  to  «AT  +  1  for  TA  =  1  and  (Ar"+1  —  1)/  (K  —  1)  for  7^  =  n.  The  number  of 
linear  equations  to  be  solved  becomes 


(n-  TA  +  1)  K  ' 


The  value  cx  =  \  (TA  =  n)  was  used  to  determine  the  precision  of  the  obtained  results, 
since  VA  =  0.5.  For  p  =  0.5  the  maximum  error  found  was  3  x  10~5  and  for  p  =  0.1  it  was 
1  x  10~4. 

Table  8  gives  some  calculated  results  for  p  =  0.5. 


TABLE  8 

—  (K^  x  104),  Partial  Concentration 

Number 

Number 

M=  ./V 

K 

cX 

of 
Equations 

of 
Iterations 

VA 

3 

2 

2 

330 

13 

5404 

3 

2 

1760 

15 

5950 

4 

2 

6290 

16 

6412 

4 

2 

2 

930 

12 

5639 

3 

930 

13 

5192 

3 

2 

7502 

14 

6304 

3 

7502 

15 

6127 

5 

2 

2 

3906 

12 

5541 

3 

2394 

12 

5311 

4 

2394 

12 

4523 

Comparing  the  results  of  Table  5  and  Table  8,  one  sees  that  partial  concentration  with 
cx  =  2  is  better  than  total  concentration  for  the  cases  M  =  4  and  k  =  2  or  3  and  any  partial 
concentration  is  better  for  the  case  M  =  5  and  k  =  2.  Further  investigations  are  needed. 

8.   SUMMARY 

The  proposed  model  is  an  idealization  of  combat  between  small  groups  of  individual 
identical  firers  and  is  very  far  from  the  very  complicated  process  of  real  combat.  However,  it 
has  provided,  through  the  use  of  alternate  firing  as  an  expression  for  the  interaction  between 
opponents,  some  important  insights  into  combat  dynamics  that  could  be  further  investigated 
with,  for  example,  a  high-resolution  Monte  Carlo  simulation.  It  has  been  shown  that  alternate 
firing  gives  the  same  results  for  small  hit  probabilities  as  some  previously  developed  models.  It 
has  also  been  shown  that  the  relationship  between  the  size,  the  hitting  capacity  and  the  resis- 
tance of  the  opponents  is  a  complex  one  and  that  concentrated  firing  is  better  than  alternate 
firing  if  the  amount  of  over-killing  is  not  too  high.  Moreover,  some  evidence  suggests  that  par- 
tial concentration  can  be  even  more  effective. 


CONCENTRATED  FIRING  IN  MANY-VERSUS-M ANY  DUELS  695 

ACKNOWLEDGMENTS 

This  research  was  supported  by  an  FIR  grant  from  Universite  du  Quebec  a  Montreal.  The 
author  wishes  to  thank  the  Service  de  lTnformatique  for  its  help  in  providing  computing  facili- 
ties and  also  the  referee  and  an  Associate  Editor  for  their  many  helpful  comments. 

REFERENCES 

[1]  Ancker,  C.J.,  Jr.,  "The  Status  of  Developments  in  the  Theory  of  Stochastic  Duels-II," 

Operations  Research  75,  388-406  (1967). 
[2]  Ancker,  C.J.,  Jr.  and  T.  Williams,  "Some  Discrete  Processes  in  the  Theory  of  Stochastic 

Duels,"  Operations  Research  13,  202-216  (1965). 
[3]  Ancker,  C.J.,  Jr.,  "Stochastic  Duels  with  Bursts,"  Naval  Research  Logistics  Quarterly  23, 

703-711  (1976). 
[4]  Bhashyam,  N.,  "Stochastic  Duels  with  Lethal  Dose,"  Naval  Research  Logistics  Quarterly 

17,  397-405  (1970). 
[5]  Boole,  G.,  A  Treatise  on  the  Calculus  of  Finite  Differences  (MacMillan  and  Co.,  London, 

1860). 
[6]  Brown,  R.H.,  "Theory  of  Combat:  The  Probability  of  Winning,"  Operations  Research  11, 

418-425  (1963). 
[7]  Dolansky,  L.,  "Present  State  of  the  Lanchester  Theory  of  Combat,"  Operations  Research 

12,  344-358  (1964). 
[8]   Helmbold,  R.L.,  "A  Universal  Attribution  Model,"   Operations  Research   14,  624-635 

(1966). 
[9]  Helmbold,  R.L.,  "Solution  of  a  General  Non- Adaptive  Many  versus  Many  Duel  Model," 
Operations  Research  16,  518-524  (1968). 
[10]  Lanchester,  F.W.,  "Aircraft  in  Warfare:  The  Dawn  of  the  Fourth  Arm-No.V,  The  Principle 
of  Concentration,"  Engineering  98,  422-423  (1914)  (reprinted  on  pp.  2138-2148  of  the 
World  of  Mathematics,  J.  Newman,  Editor  (Simon  and  Schuster,  New  York,  1956). 
[11]  Lanchester,  F.W.,  Aircraft  in  Warfare;  the  Dawn  of  the  Fourth  Arm,  (Constable  and  Co., 

London,  1916). 
[12]  Robertson,  J.I.,  "A  Method  of  Computing  Survival  Probabilities  of  Several  Targets  versus 

Several  Weapons,"  Operations  Research  4,  546-557  (1956). 
[13]   Taylor,   J.,  "Solving  Lanchester-Type   Equations  for   'Modern  Warfare'   with  Variable 

Coefficients,"  Operations  Research  22,  756-770  (1974). 
[14]  Taylor,  J.,  "Optimal  Commitment  of  Forces  in  Some  Lanchester-Type  Combat  Models," 

Operations  Research  27,  96-114  (1979). 
[15]  Taylor,  J.  and  S.  Parry,  "Force-Ratio  Considerations  for  some  Lanchester-Type  Models  of 

Warfare,"  Operations  Research  23,  522-533  (1975). 
[16]  Thompson,  D.E.,  "Stochastic  Duels  Involving  Reliability,  Naval  Research  Logistics  Quar- 
terly 19,  145-148  (1972). 
[17]   Williams,  T.,  "Stochastic  Duels-II,"   System  Development  Corporation  Document,  SP 
1017/003/00,  31-61  (1963). 


SPIKE  SWAPPING  IN  BASIS  REINVERSION* 


R.  V.  Helgason  and  J.  L.  Kennington 

Department  of  Operations  Research 

and 

Engineering  Management 

Southern  Methodist  University 

Dallas,  Texas 

ABSTRACT 

During  basis  reinversion  of  either  a  product  form  or  elimination  form  linear 
programming  system,  it  may  become  necessary  to  swap  spike  columns  to  effect 
the  reinversion  and  maintain  the  desired  sparsity  characteristics.  This  note 
shows  that  the  only  spikes  which  need  be  examined  when  an  interchange  is  re- 
quired are  those  not  yet  processed  in  the  current  external  bump. 


I.  INTRODUCTION 

An  important  component  of  a  large  scale  linear  programming  system  is  the  reinversion 
routine.  This  paper  addresses  an  important  ancillary  technique  for  implementing  a  reinversion 
routine  utilizing  the  pivot  agenda  algorithms  of  Hellerman  and  Rarick  [5,6].  Production  of  fac- 
tors during  reinversion  typically  involves  a  left-to-right  pivoting  process.  Unfortunately,  during 
the  left-to-right  process,  a  proposed  pivot  element  of  a  spike  column  may  be  zero,  in  which 
case  columns  are  interchanged  in  an  attempt  to  obtain  a  pivotable  column  while  maintaining 
desired  sparsity  characteristics.  In  this  paper  we  show  that  the  only  columns  which  need  be 
considered  for  the  interchange  with  a  nonpivotable  spike  are  other  spikes  lying  to  the  right 
within  the  same  external  bump. 

II.  PRODUCT  FORM  OF  THE  INVERSE 

Let  B  be  any  m  x  m  nonsingular  matrix.  One  of  the  most  common  factorizations  for  B~{ 
is  the  product  form  which  corresponds  to  the  method  for  solving  a  system  of  linear  equations 
known  as  Gauss- Jordan  reduction  (see  [3,  4]).  This  procedure  is  used  to  represent  B~x  (or  a 
row  and  column  permutation  of  B~l)  as  the  product  of  matrices  each  of  the  form 

/ 
Z  =  z        ,  «—  y'th  row 


'This  research  was  supported  in  part  by  the  Air  Force  Office  of  Scientific  Research  under  Contract  Number  AFOSR 
77-3151. 


698  R.V.  HELGASON  AND  J.L.  KENNINGTON 

where  z  is  an  m-component  column  vector,  and  j  is  called  the  pivot  row.    A  few  observations 
concerning  Z  are  obvious. 

PROPOSITION  1:   Z  is  nonsingular  if  and  only  if  z,  ^  0. 

PROPOSITION  2:  Let  3  be  any  m-component  vector  having  Bj  =  0.  Then  ZB  =  B. 

PROPOSITION  3:    Let  B  be  any  m-component  vector  having  Bj  ^  0,  and  let  ej  denote 
the  vector  having  yth  component  1  and  all  other  components  zero. 

\-Bk/Bj,  ifk*A 
Letz*=(     1//3,,  if*-;   )•  ThenZ^-e/. 

Let  B(i)  denote  the  rth  column  of  the  matrix  B.  Consider  the  following  algorithm. 
ALG  1:  Product  Form  Factorization 

0.  Initialization 

Interchange  columns  of  B,  if  necessary,  so  that  the  first  component  of  5(1)  is  nonzero. 
Set  / «-  1,0  «-  5(1),  and  go  to  3. 

1.  Update  Column 
Setfi^  Ej~x  ...EXBU). 

2.  Swap  Columns  If  Pivot  Element  Equals  Zero 

If  Bj  ^  0,  go  to  3;  otherwise,  there  is  some  column  B(j)  with  j  >  i  such  that  the  fth 
component  of  y  —  E'~l  . . .  ElB{j)  is  nonzero.  Interchange  B(J)  and  BU)  and  set/8  «—  y. 

3.  Obtain  New  Elementary  Matrix 

Set 

1//3,,  for  A:  =  i 
-Bk/Bj,  otherwise, 


E'+- 


4.    Test  for  Termination 

If  /  =  m,  terminate;  otherwise,  i «—  /  +  1  and  go  to  1.    At  the  termination  of  ALG  1, 
Em  . . .  E1  is  a  row  permutation  of  B~l. 

In  the  following  two  propositions  we  show  that  if  in  Step  2,  fi,  =  0,  then  the  proposed 
interchange  is  always  possible.  Consider  the  following: 

PROPOSITION  4:  For  /  ^j,EJ...  ElB(i)  =  e1. 


SPIKE  SWAPPING  IN  BASIS  REINVERSION 


699 


PROOF:  By  the  construction  of  E'  and  Proposition  3,  E' ...  E]B(i)  =  e'.  By  Proposition 
2,  EJ ...  Ei+le'—  e'.  So  EJ  . . .  ExB(i)  =  e'.  Using  Proposition  4  we  may  now  show  the  fol- 
lowing: 


PROPOSITION  5:    For  2  <  /  <  m,  let  B  =  E'~ 
j  >  i  such  that  [EHl  . . .  ExB(j)]i  *  0. 


.  ElB(i).    If  Bi  =  0,  there  is  some 


PROOF:  Suppose  [E'~l  ...  ElB(j)]i  =  0  for  all  j  >  i.  By  the  construction  of 
Ex,  ...  E'~l,  in  ALG  1,  and  Proposition  1,  each  factor  is  nonsingular.  Since  B  is  nonsingular, 
E'~x  ...  EXB  is  nonsingular.  By  Proposition  4,  E'~x  ...  EXB(J)  =  ej  for  1  ^  j  ^  /  —  1. 
Hence,  the  /th  row  of  E/_1  ...  EXB  is  all  zero,  a  contradiction. 

III.   BUMP  AND  SPIKE  STRUCTURE 


In  order  to  minimize  the  core  storage  required  to  represent  the  ETA  file,  i.e., 
El,  ...  ,  Em,  the  rows  and  columns  of  B  are  interchanged  in  an  attempt  to  place  B  in  lower  tri- 
angular form.  If  this  can  be  accomplished,  then  the  m  nonidentity  columns  of  E1,  ...  ,  Em, 
have  the  same  sparsity  structure  as  B.  Consider  the  following  proposition: 


Et 


PROPOSITION  6:    If  the  first  j  - 
...EXB(J)  =  B(J). 


1   components  of  B(j)  are  zero  for  j  >  2,  then 


PROOF:  This  follows  directly  from  successive  application  of  Proposition  2.  Therefore,  if 
B  is  lower  triangular,  the  factored  representation  of  B~x  may  be  stored  in  approximately  the 
same  amount  of  core  storage  as  B  itself.  In  practice  it  is  unneccessary  to  calculate  the  elements 
\/Bk  and  -8j/Bk  in  Step  3  of  ALG  1.  It  suffices  to  store  k  and  the  elements  of  Bt.  It  may 
prove  advantageous  to  store  \/Bk,  in  addition.  If  Proposition  6  applies  for  B(k),  then 
B  =  B(k)  and  the  only  additional  storage  required  is  for  the  index  k  (and  possibly  \/Bk). 
Clearly,  this  results  in  substantial  core  storage  savings  compared  to  storing  B~x  explicitly. 

If  B  cannot  be  placed  in  lower  triangular  form,  then  it  is  placed  in  the  form: 


Bx 

Hi 

B2 

111 

te 

2?3 

where  Bl  and  B3  are  lower  triangular  matrices  with  nonzeroes  on  their  diagonals.  We  assume 
that  if  B2  is  nonvacuous,  every  row  and  column  has  at  least  two  nonzero  entries,  so  that  no 
rearrangement  of  B2  can  expand  the  size  of  Bx  or  B2.  B2  is  called  the  bump  section,  the  merit 
section  or  the  heart  section.  We  further  require  the  heart  section  to  assume  the  following  form: 


B2- 


Fx 

Ijjg 

Gl 

^m 

F2 

^nn 

G2 

^^^^^^^^^^ 

mm 

HHt§ 

b§^ 

*■  i 

700  R.V.  HELGASON  AND  J.L    KENNINGTON 

where  Gk's  are  either  vacuous  or  lower  triangular  with  nonzeroes  on  the  diagonal.  The  only 
partitions  in  B  having  columns  with  nonzeroes  above  the  diagonal  are  the  /*'s  which  are  called 
external  bumps.  The  columns  extending  above  the  diagonal  are  called  spikes  or  spike  columns. 
An  external  bump  is  characterized  as  follows: 

(i)  the  last  column  of  an  external  bump  will  be  a  spike  with  a  nonzero  lying  in  the  top- 
most row  of  the  external  bump,  and 

(ii)  the  nonspike  columns  have  nonzero  diagonal  elements. 

The  algorithms  of  Hellerman  and  Rarick  [5,6]  produce  such  a  structure  for  any  nonsingular 
matrix,  and  we  shall  call  a  matrix  having  this  structure  an  HR  matrix.  It  should  be  noted  that  if 
one  applies  ALG  1  to  an  HR  matrix,  then  the  only  columns  which  may  require  an  interchange 
are  spike  columns.  We  now  prove  that  the  only  columns  which  need  be  considered  for  this  inter- 
change are  other  spikes  in  the  same  external  bump. 

Consider  the  following  result: 

PROPOSITION  7:  Let  B(i)  with  /  >  2  correspond  to  the  first  column  of  some  external 
bump,  /*  and  let  B(j)  be  a  spike  in  Fk.  Then  EHl  . . .  ElB(j)  =  B(J). 

PROOF:  Note  that  the  first  i  -  1  components  of  B(j)  are  zero.  Therefore,  by  successive 
application  of  Proposition  2,  the  result  is  proved. 

Note  that  Proposition  6  allows  one  to  eliminate  all  of  the  calculation  required  in  Step  1  of 
ALG  1  for  nonspike  columns  and  Proposition  7  allows  one  to  eliminate  some  of  this  calculation 
for  spikes.  We  now  address  the  issue  of  spike  swapping.  Consider  the  following  propositions: 

PROPOSITION  8:  Any  spike  B(J)  which  is  not  pivotable  cannot  be  interchanged  with  a 
spike  B(k),  k  >  j,  from  another  external  bump,  to  yield  a  pivotable  column. 

PROOF:  Since  B(k)  is  from  an  external  bump  lying  to  the  right  of  the  external  bump 
containing  B(j),  Bj(k)  =  0.  By  repeated  application  of  Proposition  2,  Ej~x...Ex 
B(k)  =  B(k).  Thus  B(j)  cannot  be  interchanged  with  B(k)  to  yield  a  pivotable  column. 

PROPOSITION  9:  Any  spike  B(J)  which  is  not  pivotable  cannot  be  interchanged  with  a 
nonspike  column  B(k),  k  >  y,  to  yield  a  pivotable  column. 

PROOF:  Let  B(k),  with  k  >  j  correspond  to  any  nonspike  column.  From  Proposition  6, 
Ej~x  ...  ExB(k)  =  B(k).  Since  the  yth  component  of  B(k)  is  zero,  B(j)  cannot  be  inter- 
changed with  B(k),  to  yield  a  pivotable  column.  We  now  present  the  main  result  of  this  note. 

PROPOSITION  10:  Any  spike  column  B(j),  which  is  not  pivotable  can  be  interchanged 
with  a  spike,  B(k),  with  k  >  j  within  the  same  external  bump,  to  yield  a  pivotable  column. 

PROOF:  If  B(J)  is  not  pivotable,  then  by  Proposition  5  there  exists  a  column  Bik)  with 
k  >  j  which  is  pivotable.  By  Proposition  8,  Bik)  cannot  be  a  spike  from  a  different  external 
bump.  By  Proposition  9,  B  (k)  cannot  be  a  nonspike.  Hence  B(k)  must  be  a  spike  from  the 
same  external  bump. 


SPIKE  SWAPPING  IN  BASIS  REINVERSION  701 

In  practice,  the  zero  check  in  step  2  is  replaced  by  a  tolerance  check.  Discussions  of  prac- 
tical tolerance  checks  may  be  found  in  Benichou  [1],  Clasen  [2],  Orchard-Hays  [7],  Saunders 
[8],  Tomlin  [9],  and  Wolfe  [10]. 

REFERENCES 

[1]  Benichou,  M.,  J.  Gauther,  G.  Hentges,  and  G.  Ribiere,  "The  Efficient  Solution  of  Large- 
Scale  Linear  Programming  Problems— Some  Algorithmic  Techniques  and  Computa- 
tional Results,"  Mathematical  Programming,  13,  280-322  (1977). 

[2]  Clasen,  R.J.,  "Techniques  for  Automatic  Tolerance  Control  in  Linear  Programming,"  Com- 
munications of  the  Association  for  Computing  Machinery,  9,  802-803  (1966). 

[3]  Forsythe,  G.E.  and  C.B.  Moler,  Computer  Solution  of  Linear  Algebraic  Systems,  (Prentice- 
Hall,  Englewood  Cliffs,  New  Jersey,  1967). 

[4]  Hadley,  G.  Linear  Algebra  (Addison  Wesley  Publishing  Co.,  Inc.,  Reading,  Massachusetts, 
1964). 

[5]  Hellerman,  E.  and  D.  Rarick,  "The  Partitioned  Preassigned  Pivot  Procedure  CP4),"  Sparse 
Matrices  and  Their  Applications,  D.  Rose  and  R.  Willoughby,  Editors,  (Plenum  Press, 
New  York,  New  York,  1972). 

[6]  Hellerman,  E.  and  D.  Rarick,  "Reinversion  with  the  Preassigned  Pivot  Procedure," 
Mathematical  Programming,  1,  195-216  (1971). 

[7]  Orchard-Hays,  W.,  Advanced  Linear  Programming  Computing  Techniques,  (McGraw-Hill, 
New  York,  New  York,  1968). 

[8]  Saunders,  M.A.,  "A  Fast,  Stable  Implementation  of  the  Simplex  Method  Using  Bartels- 
Golub  Updating,"  Sparse  Matrix  Computations,  213-226,  J.R.  Bunch  and  D.J.  Rose,  Edi- 
tors (Academic  Press,  New  York,  New  York,  1976). 

[9]  Tomlin,  J. A.,  "An  Accuracy  Test  for  Updating  Triangular  Factors,"  Mathematical  Program- 
ming Study  4,  M.L.  Balinski  and  E.  Hellerman,  Editors,  (North-Holland,  Amsterdam, 
1975). 
[10]  Wolfe,  P.,  "Error  in  the  Solution  of  Linear  Programming  Problems,"  Error  in  Digital  Com- 
putation, 2,  L.B.  Rail,  Editor  (John  Wiley  and  Sons,  Inc.,  New  York,  New  York  1965). 


AN  ALTERNATIVE  PROOF  OF 
THE  IFRA  PROPERTY  OF  SOME  SHOCK  MODELS* 

C.  Derman  and  D.  R.  Smith 

Columbia  University 
New  York,  New  York 


Let  Hit)  =  £ ,  {t)     P(k),  0  <  t  <  oo,  where  A  it)/t  is  nonde- 

_*=0  k-  _ 

creasing  in  t,  {Pik)xlk\  is  nonincreasing.  It  is  known  that  Hit)  =  1  -  H(t)  is 
an  increasing  failure  rate  on  the  average  (IFRA)  distribution.  A  proof  based 
on  the  IFRA  closure  theorem  is  given.  Hit)  is  the  distribution  of  life  for  sys- 
tems undergoing  shocks  occurring  according  to  a  Poisson  process  where  P(k)  is 
the  probability  that  the  system  survives  k  shocks.  The  proof  given  herein 
shows  there  is  an  underlying  connection  between  such  models  and  monotone 
systems  of  independent  components  that  explains  the  IFRA  life  distribution  oc- 
curring in  both  models. 


1.   INTRODUCTION 

In  Barlow  and  Proschan  [1,  p.  93]  a  fairly  general  damage  model  is  considered.  A  device 
is  subject  to  shocks  occurring  in  time  according  to  a  Poisson  process  with  rate  k.  The  damage 
caused  by  shocks  is  characterized  by  a  sequence  of  numbers  {Pik)},  where  P(k)  is  the  proba- 
bility that  the  device  will  survive  k  shocks.  The  Pik)'s  as  shown  in  [1]  can  arise  in  different 
models.  For  example,  the  damage  caused  by  the  rth  shock  can  be  assumed  to  be  a  nonnegative 
random  variable  Xh  where  X\,  X2  ...  are  independent  and  identically  distributed;  failure  of  the 

k 

device  occurs  at  the  /cth  shock  if  £  Xh  the  cumulative  damage,  exceeds  a  certain  thres- 

hold.  In  this  case  Pik)  =  /VJ£  Xj  ^  vL  where  y  is  the  threshold.  Ross  [2]  has  failure  occur- 
ring when  some  nondecreasing  symmetric  function  D(X\,  ....  X„)  first  exceeds  a  given  thres- 
hold; i.e.,  DiXx,  ...  ,  A"*)  is  a  generalization  of  £  Xr   Here,  P(k)  =  Pr  [D(XX,  . . .  Xk)  <  y\. 

t=\ 

Let  Hit)  denote  the  probability  that  the  device  survives  in  the  interval  [0,  t].  Then 

Hit)  -  £  J 
*=o 


In  Barlow  and  Proschan  [1]  (Theorem  3.6  p.  93)  it  is  proven  that  if  {Pik)]/k}  is  a  nonincreasing 
sequence  then  Hit)  =  1  -  Hit)  is  always  an  increasing  failure  rate  on  the  average  (IFRA) 


*Work  supported  in  part  by  the  Office  of  Naval  Research  under  Contract  N0014-75-0620  and  the  National  Science 
Foundation  under  Grant  No.   MCS-7725-146  with  Columbia  University. 


703 


704  C.  DERMAN  AND  DR.  SMITH 

distribution  function;  i.e., is  nondecreasing  in  t. 

Ross  [2],  generalizes  by  allowing  the  Poisson  process  of  successive  shocks  to  be  nonho- 
mogeneous  with  rate  function  \  (/)  such  that 

A  it)  _  Slx{s)ds 
t  t 

is  nondecreasing  in  /.   That  is,  the  same  assertion  can  be  made  when  Hit)  is  given  by 
™Ai 


_  OO  -A(t)    A   (f\k     _ 

(1)  Hit)  =  £- rj^-Pik),  0  ^  t  <  - 

*=0 


The  proof  given  in  [1]  is  based  on  total  positivity.  arguments.  Ross's  technique  for  prov- 
ing the  IFRA  result  is  obtained  by  making  use  of  recent  results  [3]  pertaining  to  what  he  calls 
increasing  failure  rate  average  stochastic  processes. 

Our  proof  below  shows  that  all  such  results  are  a  consequence  of  one  of  the  central 
theorems  of  reliability  theory,  the  IFRA  Closure  Theorem  ([1]  p.  83).  This  theorem  asserts 
that  a  monotone  system  composed  of  a  finite  number  of  independent  components,  each  of 
which  has  an  IFRA  life  distribution,  has  itself  an  IFRA  distribution. 

It  is  remarked  in  [1,  p.  91]  that  the  coherent  (or  monotone)  system  model  and  the  shock 
models  under  consideration  are  widely  diverse  models  for  which  the  IFRA  class  of  distribution 
furnishes  an  appropriate  description  of  life  length,  thus  reenforcing  the  importance  of  the  IFRA 
class  to  reliability  theory.  The  implication  of  our  proof  is  that  the  models  are  not  as  widely 
diverse  as  supposed. 

The  idea  of  the  proof  is  the  construction  of  a  monotone  system  (of  independent  com- 
ponents, each  of  which  has  the  same  IFRA  life  distribution)  whose  life  distribution  approxi- 
mates Hit).  The  proof  is  completed  by  allowing  the  number  of  components  in  the  system  to 
increase  in  an  appropriate  way  so  that  the  approximating  life  distributions  converge  to  Hit)', 
the  IFRA  property  being  preserved  in  the  limit. 

2.  APPROXIMATING  SYSTEMS  APPROACH 

For  each  m,  m  =  1,2  ...  let  Sm„,  n  =  1,2,  ...  be  a  monotone  system  of  n  independent 
components.   Let 

(1)  Pmnik)  =  Pr  {no  cut  set  is  formed  |  exactly  k  components  of  Sm „  are  failed} 

where  all  of  the  n  components  are  equally  likely  to  fail.    (A  cut  set  is  a  set  of  components  such 
that  if  all  components  of  the  set  fail,  the  system  does  not  function) .  Assume 

(2)  Pm,„ik)  =  0,  if  k  >  m  for  every  n, 

(3)  lim  Pm_n  ik)  =  Pmik),  for  every  k 

(4)  lim  Pmik)  =  Pik),  for  every  k. 
We  can  state 


IFRA  PROPERTY  OF  SOME  SHOCK  MODELS 


Hit)  =  1  -  Hit)  given  by  (1)  is  IFRA. 

PROOF:  Assume  every  component  in  Smn  is  independent  with  life  distribution 
Lit)  =  1  -  e~AU>/".  Then  every  component  has  an  IFRA  distribution.  Let  Qmnik,t)  denote 
the  probability  that  exactly  k  units  fail  within  [0,  t].   That  is 

M(      =*mki  =m.\*-* 

(5)  Qm,„  (k,0  =  L U  -  e     "    )    U    "    J      . 

Let  Hm  nit)  denote  the  probability  that  Smi„  works  for  at  least  /  units  of  time,  then 

(6)  Hmjlit)=  £  Qmj,ik,t)PmJk). 

k=0 

By  the  IFRA  Closure  Theorem,  Hm,„it)  is  IFRA. 

However, 

(7)  Hmit)  =  lim  Hmj,it) 

=  £  Hm  Qm,nik,t)  Pjk) 


=  1 

k=0 

i 


k\ 


k  =  0 

by   (2),  the  Poisson  limit  of  binomial  probabilities,  and   (3).    Since  the  IFRA_property  is 
preserved    in    the    limit,    Hmit)    is    IFRA.     That    is,    since    Hmit)  =  lim     Hmnit)    and 

-(log  Hm nit))/t\s  nondecreasing  in  t,  then  so  is  -(log  Hmit))/t.   However, 
-Ait)k  lim   Pmik) 

k=o  K- 

=  Hit). 

Since  again  the  IFRA  property  is  preserved  in  the  limit,  it  follows  that  Hit)  is  IFRA,  proving 
the  theorem. 


We  emphasize  that  the  IFRA  Closure  Theorem  is  invoked  only  to  show  that  that  Hmnit) 
is  IFRA.  The  condition  that  A  it)/ 1  is  nondecreasing  is  needed  so  that  all  components  of  Sm  „ 
have  an  IFRA  distribution. 

3.    APPLICATION  OF  THEOREM 

The  condition  that  [Pik)xlk)  is  a  nonincreasing  sequence  is  not  used  in  the  proof  nor  does 
it  appear  in  the  statement  of  Theorem  1.  That  the  condition  is  implicit  is  due  to  a  recent 
remarkable  result  of  Ross,  Shashahani  and  Weiss  [4]  that  [Pik)xlk)  is  necessarily  nonincreasing. 


706  C   DERMAN  AND  DR.  SMITH 

To  apply  Theorem  1  for  our  purpose  we  must  show 

THEOREM  2:  Let  {P(k)}  be  any  sequence  such  that  0  <  P(k)  ^  1  and  {P(k)Uk}  is 
nonincreasing.   Then  there  exist  the  monotone  systems  [Smn)  such  that  (2),  (3),  and  (4)  hold. 

PROOF:  Let  {P(k)}  be  any  sequence  with  the  hypothesized  properties.  Let  F  be  any 
increasing  continuous  distribution  function  over  [0,  °°)  and  {yk}  the  nonincreasing  sequence  of 
nonnegative  numbers  such  that 

F(yk)  =  P(k)Uk,    k  =  1,2,  ... 

For  each  m(m  <  n)  let  Smn  be  a  set  of  n  components,  /  =  1,  ...  ,  n.  The  cut  sets  are  con- 
structed in  the  following  way.  The  rth  component  has  an  associated  value  xh  /  =  1,  ...  ,  n 
where  the  values  are  assigned  so  that 

#{/U,  ^  x)  =  [n  Fix)],  0  <  x  ^  v,, 

=  n,  x  >  y\, 

where  #  means  "number  of"  and  [  ]  is  the  greatest  integer  designator.  Every  set  of  k  com- 
ponents is  a  cut  set  if  k  >  m\  if  k  ^  m  a  set  (i\ ik)  of  components  is  a  cut  set  if  and 

only  if 

max  (x/,  . . .  ,  Xj )  >  yk. 

Since  [yk]  is  nonincreasing,  Sm  „  is,  indeed,  a  monotone  set.   But  here, 

fe-l  [nF(yk)]  -  i 
Pm.n(k)  =  n ^3- ,      k  ^  m 


■■  0  ,     k  >  m. 


Thus, 


Pm(k)  =  lim  Pmj,(k) 


Fk(yk),    if  k  <  m 
0  ,    if  k  >  m 


lim  Pm(k)  =  Fk{yk) 

=  P{k)  ,    k  =  1,2,  ...    . 
This  proves  Theorem  2. 

Theorems  one  and  two  yield  the  slightly  more  general  version  of  Theorem  3.6  [1,  p.  93]. 

The  Ross  [2]  generalization  follows  by  defining  the  cut  sets  to  be  determined  by  a  nonde- 
creasing  symmetric  function  D(x]t  ....  xk)\  i.e.,  a  set  ih  ....  ik  of  components  is  a  cut  set  of 
Smj,  if  k  >  m  or,  if  k  ^  m,  when  D(xit  ...  ,  xk)  >  v,  a  given  threshold  value.  From  the 
construction  of  Theorem  2,  Theorem  1  and  the  result  referred  to  in  [4]  it  follows  that  the 
sequence  {P(k)}  of  this  model  satisfies  the  monotonicity  condition.    For  the  special  case  of 

k  _ 

D{X\,  ...  ,  Xk)  =  £  Xh  it  is  known  that  the  sequence  {P(k)Uk}  is  nonincreasing  (see  [1]  p. 

96). 


IFRA  PROPERTY  OF  SOME  SHOCK  MODELS  707 

REFERENCES 

[1]  Barlow,  R.  and  F.  Proschan,  Statistical  Theory  of  Reliability  and  Life  Testing,  Probability 
Models,  (Holt,  Rinehart  and  Winston,  New  York,  1975). 

[2]  Ross,  S.M.,  "Generalized  Poisson  Shock  Model,"  Technical  Report,  Department  of  Indus- 
trial Engineering  and  Operations  Research,  University  of  California,  Berkeley,  California 
(1978). 

[3]  Ross,  S.M.,  "Multivalued  State  Component  Systems,"  Annals  of  Probability  (to  appear). 

[4]  Ross,  S.M.,  M.  Shashahani  and  G.  Weiss,  "On  the  Number  of  Component  Failures  in  Sys- 
tems whose  Component  Lives  are  Exchangeable,"  Technical  Report,  Department  of  Indus- 
trial Engineering  and  Operations  Research,  University  of  California,  Berkeley,  California. 


NEWS  AND  MEMORANDA 

Defense  Systems  Management  College 
Military  Reservist  Utilization  Program 

Military  reservists  from  all  U.S.  Services  now  have  a  unique  opportunity  for  a  short  tour 
at  the  Defense  Systems  Management  College,  Ft.  Belvoir  Virginia.  By  volunteering  for  the 
Reservist  Utilization  Program,  an  individual  can  increase  proficiency  training,  maintain  currency 
in  DOD  Research,  Development  &  Acquisition  Policy,  contribute  to  the  development  and  for- 
mulation of  concepts  that  may  become  the  bases  of  future  DOD  policy  and  help  solve  critical 
problems  facing  the  acquisition  community. 

Once  accepted  for  the  program,  a  reservist  may  be  assigned  to  one  of  three  areas: 
research,  education  or  operations.  As  a  research  associate,  the  individual  researches  and 
analyzes  an  area  compatible  with  his  training  and  experience.  Many  reservists  in  this  category 
currently  assist  in  the  preparation  of  material  for  a  comprehensive  textbook  on  systems  acqusi- 
tion.  The  text  will  be  used  at  DSMC  by  the  faculty  and  students  as  well  as  by  the  systems 
acquisition  community.  As  an  academic  consultant,  a  reservist  provides  special  assistance  to 
the  College  faculty  by  reviewing  course  material  in  his  area  of  expertise  and  researching  and 
developing  training  materials.  In  the  operations/administration  category,  reservists  administer 
the  program  by  recruiting  other  reservists  for  the  program,  processing  these  reservists,  and 
maintaining  files  and  records. 

Because  of  the  complexity  and  broad  scope  of  the  systems  acquisition  business,  the  Reser- 
vist Utilization  Program  requires  a  large  number  of  reservists  from  many  diverse  career  fields. 
Some  examples  of  career  fields  used  include:  engineering,  procurement,  manufacturing,  legal, 
financial,  personnel,  administration  and  logistics.  Reservists  whose  reserve  duty  assignments 
are  not  in  these  types  of  career  fields,  but  who  have  civilian  experience  in  these  areas,  are  also 
urged  to  apply. 

Many  reservists  perform  their  annual  tours  with  the  Reservist  Utilization  Program  office. 
Others  perform  special  tours  of  active  duty  or  "mandays."  When  tour  dates  are  determined  and 
coordinated  with  your  organization  and  the  RUP  office,  submit  the  proper  forms  through  your 
reserve  organization  at  least  45  days  prior  to  the  tour  date  for  an  annual  tour  or  60  days  for  a 
special  tour. 

To  apply  for  active  duty  or  to  get  additional  information,  telephone  Professor  Fred  E. 
Rosell,  Jr.  at  commerical  (703)  664-5783  or  AUTOVON  354-5783.  Reservists  outside  of  Vir- 
ginia may  call  on  toll-free  number  (800)  336-3095  ext.  5783. 


NEWS  AND  MEMORANDA 


List  of  Referees 


The  Editors  of  the  Naval  Research  Logistics  Quarterly  are  grateful  to  the  following  indivi- 
duals for  assisting  in  the  review  of  articles  prior  to  publication. 


A.  Hax 

P.  Heidelberger 

D.  P.  Heyman 
A.  J.  Hoffman 
P.  Q.  Hwang 

E.  Ignall 
P.  Jacobs 

A.  J.  Kaplan 
U.  Karmarkar 
A.  R.  Kaylan 
J.  L.  Kennington 
P.  R.  Kleindorfer 
D.  Klingman 
J.  E.  Knepley 
K.  O.  Kortanek 
D.  Kreps 
W.  K.  Kruse 
G.  J.  Lieberman 
S.  A.  Lippman 
D.  Luenberger 
R.  L.  McGill 
W.  H.  Marlow 
C.  Marshall 
K.  T.  Marshall 
M.  Mazumder 
P.  McKeown 
K.  Mehrotra 

C.  B.  Millham 

D.  Montgomery 
R.  C.  Morey 

J.  G.  Morris 

J.  A.  Muckstadt 

S.  Nahmias 

M.  F.  Neuts 

I.  Olkin 

J.  Orlin 

S.  S.  Panwalkar 


J.  H.  Patterson 
M.  Posner 

D.  Reedy 

H.  R.  Richardson 

E.  E.  Rosinger 
S.  M.  Ross 

H.  M.  Salkin 
R.  L.  Scheaffer 

B.  Schmeiser 
P.  K.  Sen 

J.  Sethuraman 
M.  L.  Shooman 
M.  Shubik 

D.  O.  Siegmund 

E.  Silver 

N.  D.  Singpurwalla 
R.  Soland 
Henry  Solomon 
Herbert  Solomon 
R.  M.  Stark 
L.  D.  Stone 
W.  Szwarc 
H.  A.  Taha 
J.  G.  Taylor 
G.  Thompson 
W.  E.  Vesley 
H.  M.  Wagner 
A.  R.  Washburn 

C.  C.  White 
T.  M.  Whitin 
J.  D.  Wiest 

J.  W.  Wingate 
R.  T.  Wong 
M.  H.  Wright 
S.  Zacks 


CORRIGENDUM: 
STOCHASTIC  CONTROL  OF  QUEUEING  SYSTEMS 

Dr.  A.  Laurinavicius  of  the  Institute  of  Physical  and  Technical  Problems  of  Energetics, 
Academy  of  Sciences,  Lithuania,  USSR,  has  pointed  out  an  error  in  the  statement  of  Theorem 
1  of  this  paper  [1].  The  expression  for  the  generator  given  there  is  valid  only  for  x  >  0,  and  a 
different  expression  holds  for  x  =  0,  the  proof  for  this  case  being  similar.  Moreover,  the 
domain  of  the  generator  can  be  extended.   The  correct  statement  is  as  follows. 

THEOREM  1:  Let  the  function  f(t,x)  be  continuous  and  such  that  the  directional  deriva- 
tives 

(1)  Dp/Ux)  =   lim    /('  +  *.*-*)-/(/,*)•  (x  >  0) 

r  h-~  o+  h 

(2)  Bg /a o)  -  to  /<'  +  *-o)-/ao)  _  i±  /(t0) 

y  h~  o+  h  at 

where  P  =  (1,  —  1)  and  Q  =  (1,0),  exist,  be  continuous  from  one  side  and  bounded.    Then 
the  infinitesimal  generator  of  the  semigroup  {T,}  is  given  by 

(3)  Af(t.x)  =  Dp/(t,x)  -kfU,x)  +  X  f~f(t.x  +  v)B(dv)     for  x  >  0 

=  £>£  /(f,0)  -  X/U0)  +  X  JJ  f{t,\)B{d\)    for  x  =  0. 

As  a  consequence  of  this  error  the  example  of  Section  3  does  not  lead  to  the  stated  result. 
A  correct  example  is  provided  by  the  following.  Let  r(r),  the  revenue  per  unit  time,  and  c(t), 
the  operating  cost  per  unit  time,  be  given  by 

r(/)  =  r  for  0  <  t  <:  t0,    and  =  0  for  t  >  t0 

c(t)  =  C]  for  0  ^  t  <  t0,    and  =  c2  for  t  >  t0. 

The  profit  from  operating  the  system  up  to  a  time  Tis  given  by  f(T,WT),  where 

(4)  fU.x)  =  r  min(xf0)  _  C\h  ~  ?!  max  (0,f  +  x  -  f0). 
This  leads  to  the  following  correct  version  of  Theorem  3. 

THEOREM  3:  Let  W0  =  w  <  t0  and  assume  that 

(5)  Xc2  r°°     [1  -  B(\)]dv  <  r  <  Xc2/3 

where  /3  is  the  mean  service  time.   Then  the  optimal  time  is  given  by 

(6)  Ta  =  inf{/  >  0:  t  +  W,  >  a) 
where  a  is  the  unique  solution  of  the  equation 

(7)  Xc2  f~     [l-fi(v)]rfv=  r. 

•"o-a 

711 


712  CORRIGENDUM 

PROOF:   It  is  found  that  for  x  >  0 

C  IfUx  +  v)  -  f(t,x)]B(dv)  =  -c2   C         .  [1  -  £(v)]rfv 
where  (r0  -  f  -  x)+  =  max  (0,f0  -  t  -  x).   Also, 

Dp  /(f,x)  =  r  for  /  <  r0,  and  -  0  for  /  ^  t0  (x  >  0) 

Dq  fUO)  =  r  for  t  <  t0,  and  =  -c2  for  r  ^  f0. 
Therefore,  the  generator  in  this  case  is  given  by 

Af(t,x)  =  r  -Kc2  f  °°        x  [1  -  B(\)]dv  for  /  <  t0,  x  >  0 

J(t0-t-x)  + 

=  -x2/3      for  r  ^  f0.  x  >  0 

(8)  =  -  c2  -  XC2/3       for  f  ^  r0,  x  =  0. 

In  applying  Theorem  2  we  note  that  Af(t.x)  <  0  for  /  ^  f0>  *  ^  0,  so  it  suffices  to  consider 
Af(t,x)  for  /  <  /0>  x  ^  0.  We  can  write 

Af(t.x)  =  (f>(t  +  x)  for  /  <  /0,  x  >  0, 

where 

(9)  0(0  =  r  -  \c2  J°°        [1  -  B(v)]dv. 

(to   t) 

We  have 

0(0)  =  r  -  Ac2  J*°°  [1  -  B(v)]dv  >  r  -  Xc2  J"~w  [1  -  fl(v)]</v  >  0 

<t>(t0)  =  r  -\c2B  <  0 

on  account  of  (5).  Also,  <f>(t)  is  a  decreasing  function  of  t.  Therefore,  there  exists  a  unique 
value  a  such  that  <f>(t)  >  0  for  0  <  t  <  a  and  <f>(t)  <  0  for  a  <  t  ^  t0.  Since  <f>(t)  ^  0  for 
f  ^  /0,  we  have  <f>(t)  <  0  for  r  ^  a.  This  means  that  Af(t,x)  <  0  for  /  +  x  ^  o,  so  the  set 
/?  of  Theorem  2  is  given  by  R  =  [(t,x):  t  +  x  ^  a},  and  the  time  of  the  first  visit  to  R  is 
given  by  (6).  Since  the  process  t  +  W,  is  monotone  nondecreasing  with  probability  one,  the 
set  R  is  closed.  Moreover,  Ta  ^  a  with  probability  one  and  also  E(Ta)  <  <».  Thus,  the  condi- 
tions of  Theorem  2  are  satisfied,  and  Ta  is  optimal  at  fVQ  =  w,  as  was  required  to  be  proved. 

A  particular  case.  Let  B(x)  =  1  -  e'**  (x  ^  0,  0  <  fi  <  °°).   The  conditions  (5)  reduce 
to 


(10)  w   <    t0 log  l^yl   <    /„ 

and  the  Equation  (7)  gives 

(11)  a=t0-—  log    —  . 

H         [  fir  J 

On  account  of  (11)  we  have  a  >  w. 

REFERENCE 

[1]  Prabhu,  N.U.,  "Stochastic  Control  of  Queueing  Systems,"  Naval  Research  Logistics  Quar- 
terly 21,  411-418  (1974). 

N.U.  Prabhu 
Cornell  University 


INDEX  TO  VOLUME  27 

ALBRIGHT,  S.C.,  "Optimal  Maintenance-Repair  Policies  for  the  Machine  Repair  Problem,"  Vol.  27,  No.  1,  March 
1980,  pp.  17-27. 

ANDERSON,  M.Q.,  "Optimal  Admission  Pricing  Policies  for  M/Ek/\  Queues,"  Vol.  27,  No.  1,  March  1980,  pp.  57-64. 

BALCER,  Y.,  "Partially  Controlled  Demand  and  Inventory  Control:  An  Additive  Model,"  Vol.  27,  No.  2,  June  1980, 
pp.  273-280. 

BARD,  J.F.  and  J.E.  Falk,  "Computing  Equilibria  Via  Nonconvex  Programming,"  Vol.  27,  No.  2,  June  1980,  pp.  233- 
255. 

BAZARAA,  M.S.  and  H.D.  Sherali,  "Benders'  Partitioning  Scheme  Applied  to  a  New  Formulation  of  the  Quadratic  As- 
signment Problem,"  Vol.  27,  No.  1,  March  1980,  pp.  29-41. 

BEN-TAL,  A.,  L.  Kerzner  and  S.  Zlobec,  "Optimality  Conditions  for  Convex  Semi-Infinite  Programming  Problems," 
Vol.  27,  No.  3,  September  1980,  pp.  413-435. 

BERREBI,  M.  and  J.  Intrator,  "Auxiliary  Procedures  for  Solving  Long  Transportation  Problems,"  Vol.  27,  No.  3,  Sep- 
tember 1980,  pp.  447-452. 

BOOKBINDER,  J.H.  and  S.P.  Sethi,  "The  Dynamic  Transportation  Problem:  A  Survey,"  Vol.  27,  No.  1,  March  1980, 
pp.  65-87. 

CALAMAI,  P.  and  C.  Charalambous,  "Solving  Multifacility  Location  Problems  Involving  Euclidean  Distances,"  Vol. 
27,  No.  4,  December  1980,  pp.  609. 

CHANDRA,  S.  and  M.  Chandramohan,  "A  Note  of  Integer  Linear  Fractional  Programming,"  Vol.  27,  No.  1,  March 
1980,  pp.  171-174. 

CHANDRAMOHAN,  M.  and  S.  Chandra,  "A  Note  on  Integer  Linear  Fractional  Programming,"  Vol.  27,  No.  1,  March 
1980,  pp.  171-174. 

CHARALAMBOUS,  C.  and  P.  Calamai,  "Solving  Multifacility  Location  Problems  Involving  Euclidean  Distances,"  Vol. 
27,  No.  4,  December  1980,  pp.  609. 

CHAUDHRY,  M.L.,  D.F.  Holman  and  W.K.  Grassman,  "Some  Results  of  the  Queueing  System  E£/M/cV  Vol.  27, 
No.  2,  June  1980,  pp.  217-222. 

COHEN,  E.A.,  Jr.,  "Statistical  Analysis  of  a  Conventional  Fuze  Timer,"  Vol.  27,  No.  3,  September  1980,  pp.  375-395. 

COHEN,  L.  and  D.E.  Reedy,  "A  Note  on  the  Sensitivity  of  Navy  First  Term  Reenlistment  to  Bonuses,  Unemployment 
and  Relative  Wages,"  Vol.  27,  No.  3,  September  1980,  pp.  525-528. 

COHEN,  M.A.  and  W.P.  Pierskalla,  "A  Dynamic  Inventory  System  with  Recycling,"  Vol.  27,  No.  2,  June  1980,  pp. 
289-296. 

COOPER,  M.W.,  "The  Use  of  Dynamic  Programming  Methodology  for  the  Solution  of  a  Class  of  Nonlinear  Program- 
ming Problems,"  Vol.  27,  No.  1,  March  1980,  pp.  89-95. 

DERMAN,  C.  and  D.R.  Smith,  "An  Alternative  Proof  of  the  IFRA  Property  of  Some  Shock  Models,"  Vol.  27,  No.  4. 
December  1980,  pp.  703. 

DEUERMEYER,  B.L.,  "A  Single  Period  Model  for  a  Multiproduct  Perishable  Inventory  System  with  Economic  Substi- 
tution," Vol.  27,  No.  2,  June  1980,  pp.  177-185. 

DISCENZA,  J.H.  and  H.R.  Richardson,  "The  United  States  Coast  Guard  Computer-Assisted  Search  Planning  System 
(CASP),"  Vol.  27,  No.  4,  December  1980,  pp.  659. 

DISNEY,  R.L.,  DC.  McNickle  and  B.  Simon,  "The  M/G/l  Queue  with  Instantaneous  Bernoulli  Feedback."  Vol.  27, 
No.  4,  December  1980,  pp.  635. 

ELLNER,  P.M.  and  R.M.  Stark,  "On  the  Distribution  of  the  Optimal  Value  for  a  Class  of  Stochastic  Geometric  Pro- 
grams," Vol.  27,  No.  4,  December  1980,  pp.  549. 

ENGELBERG,  A.  and  J.  Intrator,  "Sensitivity  Analysis  as  a  Means  of  Reducing  the  Dimensionality  of  a  Certain  Class 
of  Transportation  Problems,"  Vol.  27,  No.  2,  June  1980,  pp.  297-313. 

FALK,  J.E.  and  J.F.  Bard,  "Computing  Equilibria  Via  Nonconvex  Programming,"  Vol.  27,  No.  2,  June  1980,  pp.  233- 
255. 

GAVER,  D.P.  and  P.A.  Jacobs,  "Storage  Problems  when  Demand  Is  'All  or  Nothing"'  Vol.  27,  No.  4,  December  1980, 
pp.  529. 

GLAZEBROOK,  K.D.,  "On  Single-Machine  Sequencing  with  Order  Constraints,"  Vol.  27,  No.  1,  March  1980,  pp.  123- 
130. 

GOLABI,  K.,  "An  Inventory  Model  with  Search  for  Best  Ordering  Price,"  Vol.  27,  No.  4,  December  1980,  pp.  645. 

GOLDEN,  B.L.  and  J.  R.  Yee,  "A  Note  on  Determining  Operating  Strategies  for  Probabilistic  Vehicle  Routing,"  Vol. 
27,  No.  1,  March  1980,  pp.  159-163. 

713 


714  index  to  Volume  27 

GRASSMAN,  W.K.,  D.F.  Holman  and  M.L.  Chaudhry,  "Some  Results  of  the  Queueing  System  E£/M/c*:  Vol.  27, 
No.  2,  June  1980,  pp.  217-222. 

GREENBERG,  I.,  "An  Approximation  for  the  Waiting  Time  Distribution  in  Single  Server  Queues,"  Vol.  27,  No.  2, 
June  1980,  pp.  223-230. 

HANSON,  M.A.  and  T.W.  Reiland,  "A  Class  of  Continuous  Nonlinear  Programming  Problems  with  Time-Delayed 
Constraints,"  Vol.  27,  No.  4,  December  1980,  pp.  573. 

HANSOTIA,  B.J.,  "Stochastic  Linear  Programs  with  Simple  Recourse:  The  Equivalent  Deterministic  Convex  Program 
for  the  Normal  Exponential,  Erland  Cases,"  Vol.  27,  No.  2,  June  1980,  pp.  257-272. 

HAYNES,  R.D.  and  W.E.  Thompson,  "On  the  Reliability,  Availability  and  Bayes  Confidence  Intervals  for  Multicom- 
ponent  Systems,"  Vol.  27,  No.  3,  September  1980,  pp.  345-358. 

HELGASON,  R.V.  and  J.L.  Kennington,  "Spike  Swapping  in  Basis  Reinversion,"  Vol.  27,  No.  4,  December  1980,  pp. 
697. 

HILDEBRANDT,  G.G.,  "The  U.S.  Versus  the  Soviet  Incentive  Models,"  Vol.  27,  No.  1,  March  1980,  pp.  97-108. 

HOLMAN,  D.F.,  W.K.  Grassman  and  M.L.  Chaudhry,  "Some  Results  of  the  Queueing  System  E£/M/e*"  Vol.  27,  No. 
2,  June  1980,  pp.  217-222. 

HSU,  C.L.,  L.  Shaw  and  S.G.,  Tyan,  "Optimal  Replacement  of  Parts  Having  Observable  Correlated  Stages  of  Deteriora- 
tion," Vol.  27,  No.  3,  September  1980,  pp.  359-373. 

INTRATOR,  J.  and  M.  Berrebi,  "Auxiliary  Procedures  for  Solving  Long  Transportation  Problems,"  Vol.  27,  No.  3,  Sep- 
tember 1980,  pp.  447-452. 

INTRATOR,  J.  and  A.  Engelberg,  "Sensitivity  Analysis  as  a  Means  of  Reducing  the  Dimensionality  of  a  Certain  Class 
of  Transportation  Problems,"  Vol.  27,  No.  2,  June  1980,  pp.  297-313. 

ISAACSON,  K.  and  C.B.  Millham,  "On  a  Class  of  Nash-Solvable  Bimatrix  Games  and  Some  Related  Nash  Subsets," 
Vol.  27,  No.  3,  September  1980,  pp.  407-412. 

JACOBS,  PA.  and  D.P.  Gaver,  "Storage  Problems  when  Demand  Is  'All  or  Nothing',"  Vol.  27,  No.  4,  December  1980, 
pp.  529. 

JOHNSON,  C.R.  and  E.P.  Loane,  "Evaluation  of  Force  Structures  under  Uncertainty,"  Vol.  27,  No.  3,  September  1980, 
pp.  511-519. 

KENNINGTON,  J.L.  and  R.V.  Helgason,  "Spike  Swapping  in  Basis  Reinversion,"  Vol.  27,  No.  4,  December  1980,  pp. 
697. 

KERZNER,  L.,  A.  Ben-Tal  and  S.  Zlobec,  "Optimality  Conditions  for  Convex  Semi-Infinite  Programming  Problems," 
Vol.  27,  No.  3,  September  1980,  pp.  413-435. 

KORTANEK,  K.O.  and  M.  Yamasaki,  "Equalities  in  Transportation  Problems  and  Characterizations  of  Optimal  Solu- 
tions," Vol.  27,  No.  4,  December  1980,  pp.  589. 

LAW,  A.M.,  "Statistical  Analysis  of  the  Output  Data  from  Terminating  Simulations,"  Vol.  27,  No.  1,  March  1980,  pp. 
131-143. 

LAWLESS,  J.F.  and  K.  Singhal,  "Analysis  of  Data  from  Life-Test  Experiments  under  an  Exponential  Model,"  Vol.  27, 
No.  2,  June  1980,  pp.  323-334. 

LEV,  B.  and  D.I.  Toof,  "The  Role  of  Internal  Storage  Capacity  in  Fixed  Cycle  Production  Systems,"  Vol.  27,  No.  3, 
September  1980,  pp.  477-487. 

LOANE,  E.P.  and  C.R.  Johnson,  "Evaluation  of  Force  Structures  under  Uncertainty,"  Vol.  27,  No.  3,  September  1980, 
pp.  499-510. 

LUSS,  H.,  "A  Network  Flow  Approach  for  Capacity  Expansion  Problems  with  Facility  Types,"  Vol.  27,  No.  4,  De- 
cember 1980,  pp.  597. 

McKEOWN  ,  P.G.,  "Solving  Incremental  Quantity  Discounted  Transportation  Problems  by  Vertex  Ranking,"  Vol.  27, 
No.  3,  September  1980,  pp.  437-445. 

McKEOWN,  P.G.  and  P.  Sinha,  "An  Easy  Solution  for  a  Special  Class  of  Fixed  Charge  Problems,"  Vol.  27,  No.  4,  De- 
cember 1980,  pp.  621. 

McNICKLE  DC,  R.L.  Disney  and  B.  Simon,  "The  M/G/l  Queue  with  Instantaneous  Bernoulli  Feedback,"  Vol.  27, 
No.  4,  December  1980,  pp.  635. 

MILLHAM,  C.B.  and  K.  Isaacson,  "On  a  Class  of  Nash-Solvable  Bimatrix  Games  and  Some  Related  Nash  Subsets," 
Vol.  27,  No.  3,  September  1980,  pp.  407-412. 

MORRIS,  J.G.  and  H.E.  Thompson,  "A  Note  on  the  'Value'  of  Bounds  on  EVPI  in  Stochastic  Programming,"  Vol.  27, 
No.  1,  March  1980,  pp.  165-169. 

OREN,  S.S.  and  S.A.  Smith,  "Reliability  Growth  of  Repairable  Systems,"  Vol.  27,  No.  4,  December  1980,  pp.  539. 

PIERSKALLA,  W.P.  and  J. A.  Voelker,  "Test  Selection  for  a  Mass  Screening  Program,"  Vol.  27,  No.  1,  March  1980,  pp. 
43-55. 

RAO,  R.C.  and  T.L.  Shaftel,  "Computational  Experience  on  an  Algorithm  for  the  Transportation  Problem  with  Non- 
linear Objective  Functions,"  Vol.  27,  No.  1,  March  1980,  pp.  145-157. 

REEDY,  D.E.  and  L.  Cohen,  "A  Note  on  the  Sensitivity  of  Navy  First  Term  Reenlistment  to  Bonuses,  Unemployment 
and  Relative  Wages,"  Vol.  27,  No.  3,  September  1980,  pp.  525-528. 

REILAND,  T.W.  and  M.A.  Hanson,  "A  Class  of  Continuous  Nonlinear  Programming  Problems  with  Time-Delayed 
Constraints,"  Vol.  27,  No.  4,  December  1980,  pp.  573. 

RICHARDSON,  H.R.  and  J.H.  Discenza,  "The  United  States  Coast  Guard  Computer-Assisted  Search  Planning  System 
(CASP),"  Vol.  27,  No.  4,  December  1980,  pp.  659. 

ROSENLUND,  S.I.,  "The  Random  Order  Service  G/M/m  Queue,"  Vol.  27,  No.  2,  June  1980.  pp.  207-215. 


INDEX  TO  VOLUME  27  715 

ROSENTHAL,  R.W  ,  "Congestion  Tolls:  Equilibrium  and  Optimality,"  Vol.  27,  No.  2,  June  1980,  pp.  231-232 

ROSS,  G.T.,  R.M.  Soland  and  A. A.  Zoltners,  "The  Bounded  Interval  Generalized  Assignment  Problem,"  Vol.  27,  No. 

4,  December  1980,  pp.  625. 
SETHI,  S.P.  and  J.H.  Bookbinder,  "The  Dynamic  Transportation  Problem:  A  Survey,"  Vol.  27,  No.  1,  March  1980,  pp. 

65-87. 
SHAFTEL,  T.L.  and  R.C.  RAO,  "Computational  Experience  on  an  Algorithm  for  the  Transportation  Problem  with 

Nonlinear  Objective  Functions,"  Vol.  27,  No.  1,  March  1980,  pp.  145-157. 
SHAPIRO,  R.D.,  "Scheduling  Coupled  Tasks,"  Vol.  27,  No.  3,  September  1980,  pp.  489-498. 
SHAW,    L.,    C-L.    Hsu   and    S.G.    Tyan,   "Optimal    Replacement   of  Parts    Having   Observable   Correlated   Stages   of 

Deterioration,"  Vol.  27,  No.  3,  September  1980,  pp.  359-373. 
SHEN,  R.F.C.,  "Estimating  the  Economic  Impact  of  the   1973  Navy  Base  Closing:  Models,  Tests,  and  an   Ex   Post 

Evaluation  of  the  Forecasting  Performance,"  Vol.  27,  No.  2,  June  1980,  pp.  335-344. 
SHERALI,  H.D.  and  M.S.  Bazaraa,  "Benders'  Partitioning  Scheme  Applied  to  a  New  Formulation  of  the  Quadratic  As- 
signment Problem,"  Vol.  27,  No.  1,  March  1980,  pp.  29-41. 
SHERALI,  H.D.  and  CM.  Shetty,  "On  the  Generation  of  Deep  Disjunctive  Cutting  Planes,"  Vol.  27,  No.  3,  September 

1980,  pp.  453-475. 
SHETTY,  CM.  and  H.D.  Sherali,  "On  the  Generation  of  Deep  Disjunctive  Cutting  Planes,"  Vol.  27,  No.  3,  September 

1980,  pp.  453-475. 
SIMON,  B.,  R.L.  Disney  and  D.C.  McNickle,  "The  M/G/l  Queue  with  Instantaneous  Bernoulli  Feedback,"  Vol.  27, 

No.  4,  December  1980,  pp.  635. 
SINGHAL,  K.  and  J.F.  Lawless,  "Analysis  of  Data  from  Life-Test  Experiments  under  an  Exponential  Model,"  Vol.  27, 

No.  2,  June  1980,  pp.  323-334. 
SINGPURWALLA,  N.D.,  "Analyzing  Availability  Using  Transfer  Function  Models  and  Cross  Spectral  Analysis,"  Vol. 

27,  No.  1,  March  1980,  pp.  1-16. 
SINHA,  P.  and  P.G.  McKEOWN,  "An  Easy  Solution  for  a  Special  Class  of  Fixed  Charge  Problems,"  Vol.  27,  No.  4, 

December  1980,  pp.  621. 
SMITH,  D.R.  and  C.  Derman,  "An  Alternative  Proof  of  the  IFRA  Property  of  Some  Shock  Model,"  Vol.  27,  No.  4, 

December  1980,  pp.  703. 
SMITH,  S.A.  and  S.S.  Oren,  "Reliability  Growth  of  Repairable  Systems,"  Vol.  27,  No.  4,  December  1980,  pp.  539. 
SOLAND,  R.M.,  G.T.  Ross  and  A. A.  Zoltners,  "The  Bounded  Interval  Generalized  Assignment  Problem,"  Vol.  27, 

No.  4,  December  1980,  pp.  625. 
STARK,  R.M.  and  P.M.  Ellner,  "On  the  Distribution  of  the  Optimal  Value  for  a  Class  of  Stochastic  Geometric  Pro- 
grams," Vol.  27,  No.  4,  December  1980,  pp.  549. 
TAYLOR,  J.G.,  "Theoretical  Analysis  of  Lanchester-Type  Combat  between  Two  Homogeneous  Forces  with  Supporting 

Fires,"  Vol.  27,  No.  1,  March  1980,  pp.  109-121. 
THOMPSON,  HE.  and  J.G.  Morris,  "A  Note  on  the  'Value'  of  Bounds  on  EVPI  in  Stochastic  Programming,"  Vol.  27, 

No.  1,  March  1980,  pp.  165-169. 
THOMPSON,  W.E.  and  R.D.  Haynes,  "On  the  Reliability,  Availability  and  Bayes  Confidence  Intervals  for  Multicom- 

ponent  Systems,"  Vol.  27,  No.  3,  September  1980,  pp.  345-358. 
TOOF,  D.I.  and  B.  Lev,  "The  Role  of  Internal  Storage  Capacity  in  Fixed  Cycle  Production  Systems,"  Vol.  27,  No.  3, 

September  1980,  pp.  477-487. 
TYAN,  S.G.,  L.  Shaw  and  C-L  Hsu,  "Optimal  Replacement  of  Parts  Having  Observable  Correlated  Stages  of  Deteriora- 
tion," Vol.  27,  No.  3,  September  1980,  pp.  359-373. 
VOELKER,  J. A.  and  W.P.  Pierskalla,  "Test  Selection  for  a  Mass  Screening  Program,"  Vol.  27,  No.  1,  March  1980,  pp. 

43-55. 
WASHBURN,  A.R.,  "On  a  Search  for  a  Moving  Target,"  Vol.  27,  No.  2,  June  1980,  pp.  315-322. 
WEISS,  L.,  "The  Asymptotic  Sufficiency  of  Sparse  Order  Statistics  in  Tests  of  Fit  with  Nuisance  Parameters,"  Vol.  27, 

No.  3,  September  1980,  pp.  397-406. 
WUSTEFELD,  A.  and  U.  Zimmermann,  "A  Single  Period  Model  for  a  Multiproduct  Perishable  Inventory  System  with 

Economic  Substitution,"  Vol.  27,  No.  2,  June  1980,  pp.  187-197. 
YAMASAK1,  M.  and  K.O.  Kortanek,  "Equalities  in  Transportation  Problems  and  Characterizations  of  Optimal  Solu- 
tions," Vol.  27,  No.  4,  December  1980,  pp.  589. 
YEE,  J.R.  and  B.L.  Golden,  "A  Note  on  Determining  Operating  Strategies  for  Probabilistic  Vehicle  Routing,"  Vol.  27, 

No.  1,  March  1980,  pp.  159-163. 
ZIMMERMANN,  U.  and  A.  Wustefeld,  "A  Single  Period  Model  for  a  Multiproduct  Perishable  Inventory  System  with 

Economic  Substitution,"  Vol.  27,  No.  2,  June  1980,  pp.  187-197. 
ZINGER,  A.,  "Concentrated  Firing  in  Many-Versus-Many  Duels,"  Vol.  27,  No.  4,  December  1980,  pp.  681. 
ZLOBEC,  S.,  L.  Kerzner  and  A.  Ben-tal,  "Optimality  Conditions  for  Convex  Semi-Infinite  Programming  Problems," 

Vol.  27,  No.  3,  September  1980,  pp.  413-435. 
ZOLTNERS,  A. A.,  R.M.  Soland  and  G.T,  Ross,  "The  Bounded  Interval  Generalized  Assignment  Model,"  Vol.  27,  No. 

4,  December  1980,  pp.  625. 
ZUCKERMAN,  D.,  "A  Note  on  the  Optimal  Replacement  Time  of  Damaged  Devices,"  Vol.  27,  No.  3,  September 

1980,  pp.  521-524. 


INFORMATION  FOR  CONTRIBUTORS 

The  NAVAL  RESEARCH  LOGISTICS  QUARTERLY  is  devoted  to  the  dissemination  of 
scientific  information  in  logistics  and  will  publish  research  and  expository  papers,  including  those 
in  certain  areas  of  mathematics,  statistics,  and  economics,  relevant  to  the  over-all  effort  to  improve 
the  efficiency  and  effectiveness  of  logistics  operations. 

Manuscripts  and  other  items  for  publication  should  be  sent  to  The  Managing  Editor,  NAVAL 
RESEARCH  LOGISTICS  QUARTERLY,  Office  of  Naval  Research,  Arlington,  Va.  22217. 
Each  manuscript  which  is  considered  to  be  suitable  material  tor  the  QUARTERLY  is  sent  to  one 
or  more  referees. 

Manuscripts  submitted  for  publication  should  be  typewritten,  double-spaced,  and  the  author 
should  retain  a  copy.  Refereeing  may  be  expedited  if  an  extra  copy  of  the  manuscript  is  submitted 
with  the  original. 

A  short  abstract  (not  over  400  words)  should  accompany  each  manuscript.  This  will  appear 
at  the  head  of  the  published  paper  in. the  QUARTERLY. 

There  is  no  authorization  for  compensation  to  authors  for  papers  which  have  been  accepted 
for  publication.  Authors  will  receive  250  reprints  of  their  published  papers. 

Readers  are  invited  to  submit  to  the  Managing  Editor  items  of  general  interest  in  the  field 
of  logistics,  for  possible  publication  in  the  NEWS  AND  MEMORANDA  or  NOTES  sections 
of  the  QUARTERLY. 


NAVAL  RESEARCH 

LOGISTICS 

QUARTERLY 


DECEMBER  1980 
VOL.  27,  NO.  4 

NAVSO  P-1278 


CONTENTS 
ARTICLES 

Storage  Problems  when  Demand  is  'All  or  Nothing' 

Reliability  Growth  of  Repairable  Systems 

On  the  Distribution  of  the  Optimal  Value  for  a 

Class  of  Stochastic  Geometric  Programs 
A  Class  of  Continuous  Nonlinear  Programming 

Problems  with  Time-Delayed  Constraints 
Equalities  in  Transportation  Problems  and 

Characterizations  of  Optimal  Solutions 
A  Network  Flow  Approach  for  Capacity  Expansion 

Problems  with  Two  Facility  Types 
Solving  Multifacility  Location  Problems 

Involving  Euclidean  Distances 
An  Easy  Solution  for  a  Special  Class 

of  Fixed  Charge  Problems 
The  Bounded  Interval  Generalized 

Assignment  Model 

The  M/G/l  Queue  with  Instantaneous 
Bernoulli  Feedback 

An  Inventory  Model  with  Search  for  Best  Ordering 
The  United  States  Coast  Guard  Computer-Assisted 

Search  Planning  System  (CASP) 
Concentrated  Firing  in  Many-Versus-Many  Duels 
Spike  Swapping  in  Basis  Reinversion 

An  Alternative  Proof  of  the  IFRA  Property 

of  Some  Shock  Models 
News  and  Memoranda 
Corrigendum 
Index  to  Volume  27 


Page 

D.  P.  GAVER    529 
P.  A.  JACOBS 
S.  A  SMITH    539 
S.  S.  OREN 
P.  M.  ELLNER    549 
R.  M.  STARK 
T.  W.  REILAND    573 
M.  A.  HANSON 
K.  O.  KORTANEK    589 
M.  YAMASAKI 

H.  LUSS    597 

P.  CALAMAI    609 
C.  CHARALAMBOUS 

P.  G.  MCKEOWN    621 
P.  SINHA 
G.  T.  ROSS    625 
R.  M.  SOLAND 
A.  A.  ZOLTNERS 

R.  L.  DISNEY    635 
D.  C.  MCNICKLE 
B.  SIMON 
Price  K.  GOLABI    645 

H.  R.  RICHARDSON    659 
J.  H.  DISCENZA 

A.  ZINGER    681 
R.  V.  HELGASON    697 
J.  L.  KENNINGTON 

C.  DERMAN    703 
D.  R.  SMITH 


OFFICE  OF  NAVAL  RESEARCH 
Arlington,  Va.  22217