Skip to main content

Full text of "Journal of the Washington Academy of Sciences"

See other formats


HARVARD  UNIVERSITY 


Ernst  Mayr  Library 
of  the  Museum  of 
Comparative  Zoology 


MCZ 

UBR/VRy 

MAY  27  m 

harvard 

UNIVERSITY 


— 

233<a 


MCZ 

LIBRARY 

JUM  1  9  2006 


Journal  of  the 


HARVARD 

UNIVERSITY 


WASHINGTON 
ACADEMY  OF  SCIENCES 


Volume  92 
Number  1 
Spring  2006 


Contents 

The  Editor  Comments  .  . .  . .  . . . .  i 

Instructions  for  Authors  . . . . . . . . ii 

Raymond  Prince,  Large  Scale  Public-Private  Projects  and  Their  Macroeconomic  Impact  ...  1 

Stephen  Weil,  Learning  from  the  Eurotunnel  to  Benefit  Moonbase  Development  . .  19 

Gary  G.  Nelson,  Organizational  Evolution,  Life-Cycle  Program  Design  ....  ...  25 

Dragon  Tevdovsky,  Irina  Naoumova,  &  Stuart  Umpleby,  A  Method  for  Designing 
Improvements  in  Organizations,  Products  and  Services  .  . . 45 

Affiliated  Institutions.  . . . . . . . . . . . . . 62 

The  Philosophical  Society  of  Washington,  Selected  Minutes . 63 

Marine  Technology  Society  News . . . 79 

Affiliated  Societies . Inside  back  cover 


ISSN  0043-0439 


Issued  Quarterly  at  Washington  DC 


l^asfjtngton  Stabemp  of  ^ttcntefi 

Founded  in  1898 


Board  of  Managers 
Elected  Officers 

President 

William  Boyer 
President  Elect 

Alain  Towaide 
Treasurer 

Harvey  Freeman 
Secretary 

James  Cole 

Vice  President,  Administration 
Rex  Klopfenstein 
Vice  President,  Membership 

Thomas  Meylan 

Vice  President,  Junior  Academy 

Paul  L.  Hazan 

Vice  President,  Affiliated  Societies 
Mark  Holland 
Members  at  Large 

Cyrus  (Bob)  Creveling 
Donna  Dean 
Frank  Haig,  S .J. 

Jodi  Weseman 
Vary  Coates 
Peg  Kay 

Past  President:  F. Douglas  Witherspoon 

AFFILIATED  SOCIETY  DELEGATES: 
Shown  on  back  cover 


Editor  of  the  Journal 

Vary  T.  Coates 
Associate  Editors: 

Alain  Touwaide 
Sethanne  Howard 

Academy  Office 

Washington  Academy  of  Sciences 
Room  63 1 

1200  New  York  Ave 
Washington,  DC  20005 
Phone:  202/326-8975 
email:  was@washacadsci.org 


The  Journal  of  the  Washington  Academy  of 
Sciences 

The  Journal  is  the  official  organ  of  the  Academy. 
It  publishes  articles  on  science  policy,  the  history  of 
science,  critical  reviews,  original  science  research, 
proceedings  of  scholarly  meetings  of  its  Affiliated 
Societies,  and  other  items  of  interest  to  its  members. 
It  is  published  quarterly.  The  last  issue  of  the  year 
contains  a  directory  of  the  current  membership  of 
the  Academy. 

Subscription  Rates 

Members,  fellows,  and  life  members  in  good 
standing  receive  the  Journal  free  of  charge. 
Subscriptions  are  available  on  a  calendar  year  basis, 
payable  in  advance.  Payment  must  be  made  in  U.S. 
currency  at  the  following  rates. 

US  and  Canada .  $25.00 

Other  Countries  .  30.00 

Single  Copies  (when  available)  .  10.00 

Claims  for  Missing  Issues 

Claims  must  be  received  within  65  days  of  mailing. 
Claims  will  not  be  allowed  if  non-delivery  was  the 
result  of  failure  to  notify  the  Academy  of  a  change 
of  address. 

Notification  of  Change  of  Address 

Address  changes  should  be  sent  promptly  to  the 
Academy  Office.  Notification  should  contain  both 
old  and  new  addresses  and  zip  codes. 

POSTMASTER: 

Send  address  changes  to  WAS,  Rm.631, 

1200  York  Ave.  NW 
Washington,  DC.  20005 

Journal  of  the  Washington  Academy  of  Sciences 
(ISSN  0043-0439) 

Published  by  the  Washington  Academy  of  Sciences 
202/326-8975 

website:  www.washacadsci.org 


THE  EDITOR  COMMENTS: 


1 »  ttoe 


WE  are  pleased  to  announce  a  new  Associate  Editor,  Dr.  Sethanne 
Howard,  an  astronomer  with  much  experience  in  editing  scientific 
publications  for  the  National  Naval  Observatory. 

“Large  scale  public  technology”  might  serve  as  the  theme  for  this  issue 
of  the  Journal.  The  first  two  papers,  originally  presented  at  the  Academy’s 
2005  MoonBase  Conference,  discuss  the  development  and  funding  of  very 
large  scale  projects  such  as  the  Channel  Tunnel  or  a  manned  facility  on 
the  moon.  A  third  paper  probes  issues  in  systems  engineering  and  the 
acquisition  of  complex  systems.  Finally,  a  paper  by  authors  from  three 
countries  presents  and  demonstrates  a  method  for  identifying  priorities  for 
improving  an  organization,  institution,  project,  or  service. 

The  activities  of  two  of  the  Academy’s  affiliated  scientific  societies  are 
featured  in  this  issue.  The  Minutes  from  selected  meetings  of  the 
Philosophical  Society  are  presented,  each  of  which  features  a  succinct  and 
pithy  summary  of  a  lecture  by  a  distinguished  local  scientist.  Second, 
Mark  Holland  lays  out  the  activities,  accomplishments,  and  agenda  of  the 
Marine  Technology  Society.  All  of  Affiliated  Societies  and  Institutions  are 
urged  to  send  similar  information  for  coming  issues  of  the  Journal. 


^SiTy 


Spring  2006 


II 


INSTRUCTIONS  FOR  AUTHORS 


THE  JOURNAL  of  the  Washington  Academy  of  Sciences  is  a 
peer-reviewed  journal.  Exceptions  are  made  for  papers  requested 
by  the  editors  or  positively  approved  for  presentation  or 
publication  by  one  of  our  affiliated  scientific  societies. 

We  welcome  disciplinary  and  interdisciplinary  scientific  research 
reports  and  papers  on  technology  development  and  innovation, 
science  policy,  technology  assessment,  and  history  of  science  and 
technology.  Book  reviews  are  also  welcome. 

Contributors  of  papers  are  requested  to  follow  these  guidelines 
carefully. 

•  Papers  should  be  submitted  as  e-mail  attachments  to  the  chief 
editor,  vcoatesfohnac.com.  along  with  full  contact  information  for 
the  primary  or  corresponding  author. 

•  Papers  should  be  presented  in  Word;  do  not  send  PDF  files. 

•  Papers  should  be  6000  words  or  fewer.  If  more  than  6  graphics  are 
included  the  number  of  words  allowed  will  be  reduced 
accordingly. 

•  Graphics  must  be  in  black  and  white  only.  They  must  be  easily 
resized  and  relocated.  It  is  best  to  put  graphics,  including  tables,  at 
the  end  of  the  paper  or  in  a  separate  document,  with  their  preferred 
location  in  the  text  clearly  indicated. 

•  References  should  be  in  the  form  of  endnotes,  and  may  be  in  any 
style  considered  standard  in  the  discipline(s)  represented  by  the 
paper. 


Spring  2006 


LARGE  SCALE  PRIVATE-PUBLIC  PROJECTS 
AND  THEIR  MACROECONOMIC  IMPACT* 


i 


Raymond  Prince 

U.  S.  Department  of  Energy 


Abstract 


This  paper  examines  the  factors  common  to  three  periods  of  the 
elevated,  sustained  productivity  growth  for  which  adequate  data  exists: 
the  late  1800s  from  the  end  of  the  Civil  War  to  1890;  the  decade 
between  the  end  of  World  War  I  and  the  onset  of  the  Great  Depression; 
and  the  two  decades  after  World  War  II.  The  question  posed  is  whether 
the  public  sector  should  be  a  source  of  support  for  technologies  that 
have  a  potential  for  a  broad  economic  impact.  Noting  that  these 
productivity  booms  followed  major  wars  in  which  US  emerged  the 
clear  victor,  the  paper  concludes  that  a  potential  role  of  government  is 
to  pursue  peacetime  policies  that  bring  into  play  some  of  the  factors 
present  in  a  postwar  economy.  Such  activities  could  include  the  support 
of  applied  technology  research;  reducing  the  risk  borne  by  private 
sector  through,  for  example,  caps  on  liability;  and  fiscal  policies,  such 
as  depreciation  rates,  that  impact  the  capital  turnover  rate. 


Introduction 

The  U.S.  economy  has  been  enjoying  substantially  faster 
productivity  growth  for  the  past  eight  years  than  it  did  over  the  preceding 
two  decades.  From  1995  to  2003,  labor  productivity  -  measured  as  output 
per  worker  -  rose  at  an  average  annual  rate  of  about  3  percent,  up  from 
around  1.5  percent  between  1973  and  1995.  A  rise  in  the  rate  of 
productivity  growth  over  an  extended  period  of  time  can  create  a  societal 
benefit  in  the  form  of  a  significantly  higher  average  standard  of  living. 
For  an  economy  with  a  labor  force  that  is  a  constant  percentage  of  the 
population,  a  1.5  percent  productivity  growth  can  allow  a  doubling  of  the 


This  paper  was  presented  at  the  MoonBase  Conference,  held  in  Washington,  D.C., 
October  2005,  co-sponsored  by  the  Academy,  the  Italian  National  Institute  of 
Astrophysics,  and  High  Frontier,  Inc. 


Spring  2006 


2 


standard  of  living  every  47  years.  At  a  3  percent  growth,  the  doubling  time 
is  virtually  halved. 

Past  periods  of  rapid  productivity  growth  are  worth  examining, 
therefore,  to  answer  questions  that  arise  when  considering  future 
technologies  with  a  potential  for  a  broad  based  impact  including  cost 
reductions  and  enhanced  capabilities  in  communications,  transportation 
and  energy  industries.  Such  past  periods  witnessed  a  marked  increase  in 
investment  in  capital  that  embodied  the  new  technologies.  Paramount 
among  questions  about  these  periods  is,  therefore,  the  role  of  government 
in  fostering  such  periods  of  rapid  growth.  More  specifically,  should  the 
public  sector  be  a  source  of  support  of  these  technologies  that  have  a 
potential  for  a  broad  impact,  for  example,  by  funding  of  research, 
investment  in  infrastructure,  or  guarantees  that  reduce  the  risk  borne  by 
the  private  sector. 

In  this  paper  I  examine  1  e  factors  common  to  three  earlier  periods 
of  the  elevated,  sustained  productivity  growth  for  which  adequate  data 
exists:  the  late  1800s  from  roughly  the  end  of  the  Civil  War  to  around 
1890;  the  decade  or  so  between  the  end  of  World  War  I  and  the  onset  of 
the  Great  Depression;  and  the  two  decades  after  World  War  II.  I  do  not 
attempt  to  examine  our  latest  period  because,  in  the  opinion  of  this  author, 
not  enough  time  has  lapsed  to  fully  assess  the  causes  and  effects  of  the 
productivity  growth  associated  with  the  dissemination  of  computers,  the 
internet,  wireless  communications  and  related  technologies  during  the  last 
ten  years. 

The  majority  view  of  economists  regarding  the  factors  behind 
elevated  productivity  growth  was  recently  expressed  by  Ferguson  and 
Washer  (2004).  [See  also  Chandler  (1977),  Baskin  and  Miranti  (1997), 
White  (2000),  and  Bemanke  (2005).] 

Productivity  booms  seem  to  involve  four  key  ingredients: 
technological  innovation;  the  willingness  and  ability  of  owners  and 
corporate  managers  to  reengineer  the  internal  organization  of  their  firms  to 
take  maximum  advantage  of  those  innovations;  financial  sector 
innovations  tailored  to  the  forms  of  business  organization  predominating 
at  the  time;  and  a  skilled  and  flexible  workforce. 

Government  policies  have  only  a  limited  role  in  these  periods  of 
elevated  productivity  growth.  The  larger  share  of  the  credit  goes  to  the 
private  sector,  as  private  agents  are  generally  responsible  for  creating  and 
exploiting  the  technologies  that  drove  these  previous  productivity  booms. 


Washington  Academy  of  Sciences 


3 


Nonetheless,  governments  can  play  an  important  subordinate  role  by 
promoting  an  economic,  financial  and  legal  environment  that  is  conducive 
to  innovation  and  to  the  diffusion  of  new  technologies. 

While  not  disagreeing  with  this  conclusion,  it  is  the  purpose  of  this 
paper  to  shed  some  additional  light  on  the  influence  of  the  public  sector  in 
productivity  spurts  and  to  suggestion  some  additional  policies  that  may  be 
appropriate. 

The  Historical  Record 

For  the  total  period  from  1873  to  2003,  labor  productivity  rose  at 
an  average  rate  of  2.2  percent  per  year,  with  both  technological  change 
and  capital  deepening  contributing  importantly  to  overall  productivity 
growth.  Periods  of  robust  growth  were  interspersed  with  periods  of  more 
modest  productivity  gains.  The  first  episode  of  strong  productivity  growth 
was  the  post  Civil  War  period  from  1873  to  1890.  During  this  period, 
labor  productivity  rose  2.6  percent  per  year,  a  rate  thought  to  be 
considerably  higher  than  the  average  growth  experienced  over  the  first 
100  years  of  the  United  States. 

From  1 890  to  1917,  the  growth  rate  of  labor  productivity  slowed  to 
an  average  pace  of  only  1.5  percent  per  year.  The  U.S.  economy  then 
enjoyed  a  relatively  brief  spurt  in  labor  productivity  growth  after  World 
War  I  until  about  1927,  with  labor  productivity  rising  3.8  percent  per  year. 
Productivity  growth  was  markedly  slower  during  the  period  that  included 
the  Great  Depression  and  World  War  II  (from  1927  to  1948),  largely 
because  of  a  lack  of  capital  deepening.  However,  the  rate  of  technological 
innovation  -  as  measured  by  the  growth  of  multifactor  productivity  - 
continued  at  a  solid,  if  somewhat  slower,  pace  than  earlier  in  the  century. 

From  1948  to  1973,  a  period  sometimes  referred  to  as  the  golden 
age  of  productivity  growth,  labor  productivity  rose  at  an  annual  rate  of 
close  to  3  percent.  During  this  period,  productivity  accelerated  across  a 
broad  range  of  industries,  and  both  capital  deepening  and  a  high  rate  of 
technological  innovation  contributed  to  the  strong  pace  of  growth.  During 
the  productivity  slowdown  of  the  1970s  and  1980s,  labor  productivity 
growth  slowed  to  an  average  pace  of  1 .4  percent  per  year. 


Spring  2006 


4 


The  Post  Civil  War  Productivity  Boom 

Diffusion  of  Technology > 

The  productivity  boom  after  the  Civil  War,  for  instance,  appears  to 
have  had  its  genesis  in  a  set  of  technological  improvements  that  increased 
the  flexibility  of  production  and  reduced  transportation  costs,  which 
allowed  firms  to  take  advantage  of  economies  of  scale  in  production  and 
distribution.  The  widespread  introduction  of  steam  engines  and  machinery 
powered  by  coal  enabled  firms  to  move  away  from  sources  of  water  power 
and  closer  to  areas  where  inputs  of  labor  and  raw  materials  were  more 
readily  available. 

The  expansion  of  railroad  transportation  also  helped  raise 
productivity  growth  in  the  second  half  of  the  nineteenth  century.  Improved 
methods  of  steel  production-notably,  the  Bessemer  process  and,  later, 
Siemens's  open  hearth  method-enabled  railroads  to  lay  steel  track  that  was 
longer-lasting  than  iron  track.  The  growth  of  telegraphy  enabled  railroad 
companies  to  coordinate  the  movement  of  trains  over  a  wider  area. 

Although  the  magnitude  of  the  railroad’s  contribution  to 
productivity  growth  during  this  period  is  the  subject  of  considerable 
debate  (David  1969,  Fishlow  2000,  Fogel  1979),  the  expansion  of  the 
railroads  clearly  drove  transportation  costs  sharply  lower  and  resulted  in 
significant  increases  in  the  geographic  size  of  product  markets.  In  1830, 
the  transportation  of  goods  from  New  York  to  Chicago  occurred  mainly 
by  canal  and  required  three  weeks  even  during  the  warmer  months  of  the 
year.  By  1870  the  same  goods  could  be  transported  between  these  two 
cities  in  three  days  by  railroad  at  any  time  of  the  year  (Paullin  1932). 
Subsequently,  freight  rates  fell  from  2.25  cents  per  ton-mile  in  1860  to 
less  than  1  cent  per  ton-mile  by  1 890.  As  a  result,  the  quantity  of  goods 
transported  by  rail  increased  sharply,  from  about  12  billion  ton-miles  in 
1870  to  80  billion  ton-miles  in  1890  (Fishlow  1966). 

The  advances  in  transportation  were  complemented  by  improved 
communications,  largely  as  a  result  of  the  expansion  of  the  telegraph. 
Initially,  sending  a  telegram  was  relatively  expensive,  with  rates  between 
New  York  and  San  Francisco  averaging  $7.45  for  ten  words  or  fewer  in 
the  late  1860s.  By  the  late  1880s,  rates  for  the  same  message  had  fallen  to 
as  little  as  $1.00.  As  a  result,  the  number  of  telegraph  messages  handled 


Washington  Academy  of  Sciences 


5 


by  Western  Union  rose  from  fewer  than  6  million  in  1867  to  nearly  56 
million  in  1890  (U.S.  Census  Bureau,  1997,  Series  R48  and  R74). 


Changes  in  Business  Organization 

Before  the  Civil  War,  most  businesses  were  either  sole 
proprietorships  or  partnerships  serving  local  markets.  As  the  spread  of 
railroads  lowered  transportation  costs  and  increased  the  size  and  number 
of  potential  markets,  the  greater  availability  of  steam  power  enabled 
manufacturers  to  set  up  factories  to  take  advantage  of  economies  of  scale 
in  production.  As  a  result,  the  size  of  firms  rose  substantially  in  many 
industries.  In  the  cotton  industry,  for  example,  the  median  firm  size 
(measured  as  the  annual  value  of  gross  production  in  1860  dollars)  rose 
from  $31,000  in  1850  to  nearly  $100,000  in  1870;  similarly,  in  the  iron 
industry,  the  median  firm  size  rose  from  $24,000  in  1850  to  more  than 
$200,000  in  1870  (Atack  1986). 

With  the  telegraph  making  rapid  communication  over  great 
distances  more  feasible,  firms  were  able  to  monitor  activities  from  a 
central  administrative  office.  However,  to  make  effective  use  of  the 
opportunities  presented  by  better  communications,  firms  often  set  up 
hierarchical  management  systems  to  control  the  production  process  and  to 
coordinate  the  flow  of  goods  across  the  distribution  system.  The  more 
informed  decision  making  associated  with  this  administrative  structure 
enabled  firms  to  match  production  to  orders,  shorten  delivery  times  and 
reduce  inventory  holdings. 

Changes  in  Finance 

Before  the  Civil  War,  most  non-financial  business  investment  was 
financed  internally  with  retained  earnings,  with  capital  provided  by  family 
or  friends  or  through  partnerships  formed  with  other  proprietors.  The  chief 
exceptions  were  the  canals  and  railroads,  which  were  issuing  stocks  and 
bonds  in  the  1850s  (Chandler  1977).  The  main  sources  of  funding  in  the 
decades  after  the  Civil  War  were  debt  and  preferred  stock.  (Railroad 
companies  were  an  exception  to  this  pattern  -  they  sold  sizable  amounts  of 
common  stock  to  investors  seeking  large  capital  gains  after  the  completion 
of  new  construction  projects  (Fishlow  2000).)  Debt  often  took  the  form  of 
secured  loans,  in  large  part  because  investors  were  concerned  about  the 
informational  asymmetries  they  faced  in  evaluating  the  bankruptcy  risk  of 


Spring  2006 


6 


particular  firms.  Indeed,  the  total  value  of  bank  loans  rose  from  less  than 
$1  billion  in  1870  to  more  than  $4  billion  in  the  early  1890s,  during  a 
period  when  the  aggregate  price  level  was  falling  (U.S.  Census  Bureau, 
1997,  Series  X581). 

Changes  in  the  Labor  Market 

During  the  productivity  boom  in  the  late  nineteenth  century, 
technological  change  had  two  disparate  effects  on  the  demand  for  labor. 
First,  the  shift  in  manufacturing  production  from  artisanal  shops  in  the 
mid- 1800s  to  factories  after  the  Civil  War  and  the  subsequent  rapid 
growth  in  the  capital  stock  led  to  a  substantial  increase  in  the  demand  for 
unskilled  labor  (Engerman  and  Sokoloff  2000).  Although  this  effect 
reduced  the  average  skill  level  of  the  manufacturing  workforce,  the 
availability  of  a  large  pool  of  unskilled  labor  enabled  firms  to  take 
advantage  of  the  potential  organizational  efficiencies  and  economies  of 
scale  associated  with  the  new  technologies,  thus  raising  productivity  for 
the  economy  as  a  whole. 

Second,  increases  in  firm  size  and  the  growth  of  businesses  in  the 
distribution  sector  increased  the  demand  for  workers  who  could  perform 
clerical  and  managerial  tasks.  For  example,  the  share  of  employed  men 
who  worked  in  white-collar  occupations  rose  from  less  than  5  percent  in 
1850  to  nearly  18  percent  by  1900  (Margo  2000).  Such  workers  tended  to 
have  more  formal  education  than  the  average  individual  although  the  level 
of  competency  needed  for  these  jobs  required,  at  most,  a  high  school 
education  (Chandler  1977). 


The  Post  World  War  1  Productivity  Boom 

Diffusion  of  Technology 

In  the  second  productivity  boom  in  the  years  after  World  War  I, 
the  chief  technological  innovation  was  most  likely  the  spread  of 
electrification  to  the  factory  floor  (David  1990,  Mowery  and  Rosenberg 
2000).  For  example,  the  amount  of  mechanical  energy  derived  from 
electric  motors  rose  from  475,000  horsepower  in  1899  to  nearly  34  million 
horsepower  in  1929,  and  the  fraction  of  overall  factory  horsepower 
produced  with  electricity  rose  from  less  than  5  percent  to  more  than  80 
percent  over  that  period  (U.S.  Census  Bureau,  1997,  Series  P70).  As  a 
result,  manufacturing  plants  could  be  organized  in  a  way  that  maximized 
the  efficient  movement  of  materials,  rather  than  the  efficient  transmission 


Washington  Academy  of  Sciences 


7 


of  power.  In  this  regard,  electric  motors  facilitated  the  spread  of 
continuous  processing  techniques  and  assembly  lines.  By  one  estimate, 
productivity  growth  in  the  manufacturing  sector  as  a  whole  increased 
about  5.5  percent  per  year  between  1919  and  1929  (Kendrick  1961). 

Other  technological  innovations  also  contributed  to  productivity 
growth  during  this  period.  Notable  among  them  were  the  telephone,  the 
internal  combustion  engine,  and  a  variety  of  technological  advances  in 
machine  tools.  In  addition,  the  early  1900s  were  characterized  by  the  first 
wave  of  office  automation  equipment,  including  the  portable  typewriter 
and  adding  and  duplicating  machines.  These  machines  improved  the 
efficiency  of  a  wide  range  of  management  and  accounting  tasks.  In  real 
terms,  business  investment  in  office  equipment  increased  from  about  $50 
million  (in  1929  dollars)  in  1899  to  nearly  $500  million  in  1929,  with  a 
particularly  large  jump  evident  in  the  1920s  (Cortada  1993). 

Changes  in  Business  Organization 

The  second  major  productivity  boom,  in  the  years  after  World  War 
I,  required  changes  in  business  organization  that  permitted  firms  to  take 
advantage  of  advances  in  production  processes  in  the  early  1900s.  A 
change  in  the  optimal  size  of  the  firm  occurred  that  involved  both  the 
economies  of  scale  associated  with  the  increasingly  complex  production 
techniques  and  also  large  organizations  embracing  economies  of  scope. 
The  diffusion  of  the  electric  motor  throughout  the  factory  floor  increased 
the  use  of  continuous-process  methods  and  the  assembly  line  and,  thus, 
accelerated  the  trend  toward  mass  production.  In  addition,  as  early  as  the 
1880s,  manufacturers  had  begun  to  integrate  forward  into  distribution;  one 
noteworthy  example  was  the  meatpacking  industry,  in  which  firms 
purchased  refrigerated  rail  cars  that  allowed  the  shipment  of  beef  from 
centralized  slaughterhouses  to  branch  houses  that  served  local  markets. 
The  advances  in  mass  production  techniques  and  the  increasing 
complexity  of  many  manufactured  products  led  firms  in  other  industries  to 
integrate  forward  not  only  into  distribution  but  also  into  retailing;  this 
vertical  integration  reduced  transactions  costs  even  more  and  further 
increased  the  optimal  size  of  firms. 

As  a  result,  marketing,  advertising,  and  accounting  departments 
increased  in  size  and  importance  within  the  typical  corporation.  Also,  with 
their  executives  now  more  sensitive  to  market  share  and  their  cost 
advantage  over  their  competitors,  large  corporations  began  to  develop 


Spring  2006 


8 


applied  research  departments  aimed  at  providing  the  firm  with  a 
technological  edge. 


Changes  in  Finance 

Corporate  finance  in  the  years  after  World  War  I  was  characterized 
by  an  increase  in  the  importance  of  equity  markets.  At  the  New  York 
Stock  Exchange  alone,  the  volume  of  stock  sales  rose  from  186  million 
shares  in  1917  to  more  than  1  billion  shares  in  1929  (U.S.  Census  Bureau, 
1997,  Series  X531),  the  value  of  preferred  and  common  stock  issuance 
increased  from  $455  million  to  $6.8  billion  over  the  same  period  (U.S. 
Census  Bureau,  1997,  Series  X5 14-5 15),  and  the  number  of  individuals 
holding  stock  jumped  from  500,000  in  1900  to  10  million  by  1930 
(Hawkins  1963). 

The  public’s  interest  in  common  stock  increased  for  several 
reasons.  First,  expanding  middle  and  upper  classes  wanted  to  take  part  in 
the  economic  gains  associated  with  the  introduction  of  new  technologies. 
Second,  about  the  same  time,  the  informational  problems  that  had 
constrained  interest  in  common  stock  through  the  early  1900s  were 
declining.  Starting  in  the  late  1800s,  there  was  a  proliferation  of 
newsletters  that  reported  on  developments  in  the  railroad  industry,  and 
similar  publications  soon  sprang  up  to  provide  information  on  other  traded 
securities.  These  newsletters  evolved  into  ratings  agencies  covering  a  wide 
range  of  individual  corporations,  with  Moody’s  issuing  the  first  bond 
ratings  in  1909.  Third,  more  public  companies  recognized  a  need  to 
address  investors’  concerns  about  risk  and  began  to  issue  regular  audited 
financial  statements  (Miranti  2001).  Fourth,  the  marketing  of  securities  to 
the  household  sector  became  more  aggressive  in  the  1920s,  led  by 
investment  trusts  which  offered  investors  a  means  of  diversifying 
individual  portfolios-and  retail  brokerage  firms. 

Changes  in  the  Labor  Force 

The  productivity  boom  of  the  early  twentieth  century  was 
accompanied  by  a  significant  rise  in  the  demand  for  higher-skilled  labor. 
The  need  for  white-collar  workers  continued  to  increase  with  the  further 
growth  in  corporate  size  and  the  new  focus  on  activities  not  directly 
related  to  the  manufacture  of  goods.  The  greater  complexity  of  the  newly 
installed  capital  equipment  increased  the  demand  for  workers  who  could 


Washington  Academy  of  Sciences 


9 


read  manuals  and  blueprints,  perform  mathematical  calculations,  and  had 
some  basic  knowledge  of  science  (Goldin  and  Katz  1998).  In  response, 
enrollment  rates  in  secondary  schools  increased  sharply,  and  the  high 
school  graduation  rate  rose  to  more  than  25  percent  by  the  late  1920s 
(U.S.  Census  Bureau,  1997,  Series  H599).  Chandler  (1977)  also  notes  the 
inception  of  the  modem  business  school  during  this  period,  with  classes 
on  commerce,  accounting,  marketing,  law  and  finance. 


The  Post  World  War  II  Productivity  Boom 

Diffusion  of  Technology 

The  productivity  gains  of  the  1950s  and  1960s  had  their  roots  in  a 
wide  range  of  technologies  first  developed  during  the  1930s  (Field  2003, 
Kleinknecht  1987,  Schmookler  1966,  and  Mensch  1979).  Examples  of 
important  innovations  during  this  decade  include  research  advances  in 
polymer  chemistry  that  led  to  the  invention  of  Plexiglas,  Teflon  and 
Nylon;  significant  advances  in  civil  engineering;  and  the  introduction  of 
the  DC-3  aircraft  in  1936. 

Research  aimed  at  enhancing  U.S.  military  capabilities  during 
World  War  II  also  led  to  new  technologies  that  had  important  spillovers  to 
commercial  applications  after  the  war  (Mowery  and  Rosenberg  2000).  For 
example,  although  the  major  research  advances  in  synthetic 
polymerization  chemistry  (most  notably,  the  introduction  of  catalytic 
cracking  in  the  processing  of  crude  oil)  were  made  in  the  1920s  and 
1930s,  the  synthetic  rubber  program  launched  during  the  war  resulted  in 
techniques  that  led  to  the  mass  production  of  the  first  synthetic  polymer 
from  petroleum-based  feedstocks.  Similarly,  production  of  polyethylene,  a 
petrochemical  based  plastic  discovered  in  the  1930s,  jumped  in  the  1940s 
because  of  its  widespread  use  in  military  equipment.  The  military’s  need 
for  large  stocks  of  penicillin  led  to  a  production  process  for  it  that  turned 
out  to  have  applicability  to  a  wide  range  of  pharmaceuticals,  while 
wartime  advances  in  microelectronics  subsequently  contributed 
significantly  to  the  development  of  new  commercial  electronic  products. 
Overall,  between  1947  and  1970,  production  in  the  rubber  and  plastic 
products  industry  rose  nearly  7  percent  per  year,  and  the  output  of  the 
chemical  products  industry  rose  more  than  8  percent  per  year  (Board  of 
Governors  of  the  Federal  Reserve  System,  Indexes  of  Industrial 
Production).  In  comparison,  over  the  same  period,  production  in  the 
manufacturing  sector  as  a  whole  rose  about  4  percent  per  year. 


Spring  2006 


10 


Another  notable  contributor  to  productivity  growth  during  this 
period  is  the  invention  of  the  transistor  in  1947.  Commercial  applications 
of  the  transistor,  initially  in  solid  state  consumer  electronic  products,  were 
stimulated  by  improvements  in  the  fabrication  process  (in  1954)  and  by 
the  introduction  of  the  integrated  circuit  (in  1958).  With  the  rise  in 
demand,  semiconductor  production  jumped  markedly,  rising  nearly  20 
percent  per  year  during  the  1960s  (Board  of  Governors  of  the  Federal 
Reserve  System,  Indexes  of  Industrial  Production). 

In  transportation,  the  1950s  and  1960s  saw  major  productivity 
improvements  in  all  three  major  segments:  air,  rail  and  trucking. 
Contributing  importantly  to  those  productivity  gains  were  the  replacement 
of  steam  locomotives  with  diesel  locomotives  and  innovations  that 
increased  the  capacity  of  the  rolling  stock  (Mansfield  1965).  The  use  of 
the  jet  engine  in  commercial  aircraft  -  most  notably,  the  introduction  of 
the  Boeing  707  in  1958  -  was  important.  Gordon  (1992)  estimates  place 
the  growth  of  productivity  in  the  commercial  airline  industry  at  more  than 
7  percent  per  year  during  the  1960s,  well  above  the  rate  of  labor 
productivity  growth  for  the  economy  as  a  whole. 

Finally,  productivity  gains  in  trucking  -  estimated  by  Gordon  at 
about  3.5  percent  per  year  in  the  1950s  and  1960s  -  were  fueled 
importantly  by  substantial  investment  in  road  improvements,  most  notably 
the  federally  funded  expansion  of  the  U.S.  highway  system  (Keeler  and 
Ying,  1988). 

Changes  in  Business  Organization 

During  the  third  productivity  boom,  following  World  War  II,  firms 
responded  to  the  myriad  of  new  products  by  increasingly  splitting  their 
firm's  activities  into  separate  divisions,  each  with  its  own  manufacturing 
and  marketing  departments.  For  domestic  production,  this  multidivisional 
approach  was  well  suited  to  the  manufacturing  of  diverse  product  lines  by 
a  single  company  (Baskin  and  Miranti  1997).  This  structure  also  turned 
out  to  be  an  effective  method  of  handling  corporate  operations  in  different 
geographic  areas,  as  seen  by  the  rise  of  multinational  corporations  during 
this  period.  To  handle  these  long-distance  operations  more  easily, 
corporations  often  set  up  foreign  subsidiaries  that  could  adapt  quickly  to 
changing  circumstances  in  the  host  country's  marketplace. 


Washington  Academy  of  Sciences 


11 


Changes  in  Finance 

The  third  productivity  boom,  in  the  years  after  World  War  II,  was 
accompanied  by  another  rapid  increase  in  bond  and  equity  issuance  with 
the  ratio  of  external  financing  to  overall  capital  spending  rising  from  an 
average  of  around  30  percent  in  the  late  1940s  to  more  than  40  percent  in 
the  early  1970s  (Board  of  Governors  of  the  Federal  Reserve  System,  Flow 
of  Funds  Accounts). 

Two  specific  developments  in  financial  markets  during  this  period 
bear  mentioning. 

First,  the  late  1950s  and  1960s  saw  the  rise  of  the  Eurodollar 
market  -  a  market  for  U.S.  dollar  deposits  and  loans  outside  the  United 
States  that  became  a  useful  source  of  short-term  financing  - 
complementary  to  the  commercial  paper  market  -  for  large  corporations 
seeking  alternatives  to  more  costly  domestic  commercial  bank  loans 
(Johnston  1982,  Kindleberger  1993).  Baskin  and  Miranti  (1997)  estimate 
that  this  market  increased  from  about  $9  billion  in  1964  to  $247  billion  by 
1976. 


Second,  the  1950s  and  1960s  were  characterized  by  a  sharp  rise  in 
the  importance  of  large  institutional  investors  -  especially  pension  funds  - 
in  the  stock  and  bond  markets.  This  rise,  coupled  with  the  growth  of 
mutual  funds  and  brokerage  houses,  enabled  smaller  investors  (either 
explicitly  or  implicitly)  to  invest  more  easily  in  stocks  and  bonds  and  to 
diversify  their  portfolios. 

Changes  in  the  Labor  Force 

The  productivity  boom  of  the  1950s  and  1960s  showed  a  similar 
pattern.  The  new  technologies  and  skilled  labor  again  were  complements 
in  production,  so  that  the  availability  of  skilled  labor  in  this  episode 
helped  to  maintain  the  returns  to  technological  innovation.  As  in  the  early 
1900s,  the  greater  cognitive  skills  possessed  by  more  educated  workers 
were  especially  effective  in  implementing  the  new  technologies  (Nelson, 
Peck  and  Kalachek  1967),  and  in  this  instance,  the  demand  for  workers  in 
professional  and  technical  occupations  increased  sharply,  with  especially 
rapid  growth  for  engineers  and  technicians  (U.S.  Census  Bureau,  1997, 
Series  D233-D682).  With  the  occupations  in  highest  demand  now 
requiring  a  college  education,  the  percentage  of  18-  to  24-year-olds 


Spring  2006 


12 


enrolled  in  college  rose  from  about  14  percent  in  1950  to  roughly  32 
percent  in  1970  (U.S.  Census  Bureau,  1997,  Series  H701). 


Lessons  from  the  Past 

The  Characteristics  of  a  Sustained  Productivity  Boom 
Productivity  booms  seem  to  involve  four  key  ingredients: 

•  technological  innovation; 

•  the  willingness  and  ability  of  owners  and  corporate  managers  to 
reengineer  the  internal  organization  of  their  firms  to  take 
maximum  advantage  of  those  innovations; 

•  financial  sector  innovations  tailored  to  the  forms  of  business 
organization  predominating  at  the  time;  and 

•  a  skilled  and  flexible  workforce. 

There  are  undoubtedly  many  valuable  lessons  from  these 
similarities,  but  we  will  touch  on  a  few  that  seem  particularly  important. 

First,  many  of  the  technological  innovations  associated  with  past 
productivity  booms  were  “general  purpose  technologies”  with  widespread 
applicability.  Such  technologies  often  operate  through  various  channels  - 
through  improvements  in  energy,  transportation  or  communications,  for 
example  -  raising  productivity  not  only  in  production  but  also  in 
distribution  and  business  practices. 

Second,  in  many  cases  -  railroads  and  computers  being  notable 
examples  -  the  productivity  improvements  were  initially  most  pronounced 
in  the  production  of  the  capital  equipment  embodying  the  new 
technologies. 

Third,  the  development  of  these  new  technologies  often  had 
important  intersectoral  linkages  to  other  industries  (Fishlow  2000; 
Mowery  and  Rosenberg  2000).  In  the  nineteenth  century,  for  example,  the 
construction  of  railroads  had  backward  linkages  to  the  coal,  iron  and  steel, 
and  machinery  industries  and  forward  linkages  to  the  distribution  sector. 
Likewise,  in  the  twentieth  century,  the  innovations  in  electricity, 
chemistry  and  the  development  of  the  internal  combustion  engine  led  both 
to  widespread  productivity  improvements  in  mature  industries  (like  steel 
and  railroads)  and  the  creation  of  new  industries  (like  plastics  and 
commercial  air  transportation)  (Meyer  2003). 


Washington  Academy  of  Sciences 


13 


A  fourth  lesson  from  past  productivity  booms  is  that  investors 
must  be  willing  to  hold  securities  if  firms  are  to  raise  the  capital  they  need 
to  take  advantage  of  the  productivity  potential  of  new  technologies. 

Fifth,  efforts  by  policymakers  to  provide  broad  access  to  education 
has  also  helped  to  stimulate  economic  growth  by  improving  the  ability  of 
the  workforce  to  adapt  to  technological  change. 

Sixth,  sound  macroeconomic  policies  have  also  been  essential  in 
promoting  long-run  economic  growth.  Several  empirical  observations 
suggest  such  a  link  between  the  level  of  business  fixed  investment  -  and 
thus  the  diffusion  of  new  technologies  through  renewal  of  the  capital  stock 
-  and  an  economic  environment  characterized  by  sustainable  economic 
growth  and  low  inflation  (Fischer  1993,  Rudebusch  and  Wilcox  1994). 

The  Role  of  the  Public  Sector  Reconsidered 

The  importance  of  general  purpose  technologies  raises  the  question 
of  whether  governments  should  attempt  to  stimulate  the  development  of 
these  technologies.  To  be  sure,  government  intervention  has,  at  times, 
made  valuable  contributions  to  technological  progress.  First,  state  and 
federal  governments  have  been  an  important  source  of  funding  for  basic 
research.  Second,  the  legal  system  provides  incentives  for  innovation 
through  the  protection  of  intellectual  property  rights  by  allowing  the 
inventors  of  new  technologies  to  reap  the  benefits  of  their  innovations, 
while,  hopefully,  encouraging  the  timely  diffusion  of  new  technologies 
(Engerman  and  Sokoloff  2000). 

Third,  in  some  cases,  government  has  supported  certain  new 
technologies  more  directly.  In  the  1850s  and  after  the  Civil  War,  for 
example,  federal  land  grants  and  state  and  local  aid  were  a  source  of 
financing  for  railway  construction.  Military  support  for  chemical  research 
that  focused  on  developing  new  materials  during  World  War  II 
contributed  to  subsequent  productivity  gains  in  the  private  sector.  After 
World  War  II,  new  trade  agreements  and  efforts  to  revitalize  Europe  and 
Japan  allowed  American  firms  to  make  significant  inroads  into  foreign 
markets.  Also,  the  Federal  government  funded  the  building  of  the 
interstate  highway  system  during  the  1 950s  and  1 960s. 

Without  downplaying  the  role  of  government  in  encouraging 
invention,  however,  the  prevailing  view  of  economists  is  that  the 
government  can  arguably  contribute  most  effectively  to  technological 
change  by  promoting  an  economic,  financial,  and  legal  environment  that  is 


Spring  2006 


14 


conducive  to  innovation  and  to  the  diffusion  of  new  technologies  -  and 
then  allowing  businesses  the  flexibility  to  reorganize  their  operations  in 
ways  that  permit  them  to  take  maximum  advantage  of  new  technologies. 
It  is  often  pointed  out,  for  example,  that  even  for  the  nineteenth-century 
railroads,  external  financing  came  mainly  from  private  domestic  or  foreign 
sources;  the  proportion  of  government-funded  investment  by  railroad 
companies  was  less  than  10  percent  after  the  Civil  War  (Fishlow  2000). 

At  the  same  time,  it  is  impossible  to  notice  that  each  of  the 
historical  productivity  booms  followed  the  end  of  a  major  war.  Those  who 
regard  this  as  largely  correlation,  rather  than  causation,  often  point  out  that 
it  is  difficult  to  see  what  activities  of  the  government  during  the  Civil  War 
and  World  War  I  could  have  made  a  major  contribution  to  the  subsequent 
period  of  elevated  productivity  growth.  Moreover,  it  is  usually  noted  that 
the  post- 1995  productivity  boom  did  not  follow  the  end  of  a  war  and  that 
other  major  wars  were  not  followed  by  productivity  booms.  We  have 
already  noted  that  we  prefer  to  defer  judgment  on  the  last  ten  years  for  a 
while  longer.  Regarding  the  last  point,  we  will  only  note  that  other  past 
productivity  booms  followed  major  wars  in  which  the  United  States 
emerged  the  clear  victor. 

In  addition,  it  is  possible  to  point  to  certain  factors  present  in  a 
postwar  economy  and  attributable  to  public  sector  activity  (conduct  of  a 
war)  that  can  make  an  important  contribution  to  productivity  growth. 
Among  these  are  a  pent-up  demand  for  consumer  products  resulting  from 
the  reallocation  of  resources  to  wartime  production  (World  War  II); 
control  over  new  resources  and  technologies  acquired  from  vanquished 
nations  (World  War  I);  a  breakdown  of  factors  previously  restraining 
growth  (Civil  War  -  see  Olson  1982);  war-related  advances  in  applied 
technology  (World  War  II);  and  a  rise  in  capital  turnover  rates. 

Regarding  this  last  point,  capital  turnover  rates  are  important 
because  so  much  of  new  technology  is  embodied  in  new  equipment. 
Typically,  during  a  war  maintenance  of  equipment  falls  below  pre-War 
schedules  due  to  manpower  shortages.  By  the  end  of  an  extended  war,  a 
significant  part  of  the  capital  stock  may  be  at,  or  nea  r,  the  end  of  its  useful 
life.  The  shift  of  resources  away  from  war  productions  creates,  therefore, 
an  opportunity  to  invest  in  new  capital  that  embodies  the  latest 
technology,  thus  providing  the  impetus  for  a  rise  in  productivity  growth. 

Of  course,  few  would  advocate  a  major  war  in  order  to  benefit 
from  a  subsequent  productivity  boom.  More  relevant  is  the  question  of 


Washington  Academy  of  Sciences 


15 


whether  it  is  possible  through  peacetime  policies  to  bring  into  play  some 
of  the  factors  present  in  a  postwar  economy.  For  some  factors  -  pent-up 
demand,  acquisition  of  new  resources  from  conquered  nations,  and  even 
the  breakdown  of  restraining  factors  -  the  answer  is  probably  no.  But  it 
may  be  possible  for  the  public  sector  to  set  the  stage  for  accelerated 
productivity  growth  during  peacetime  by  increasing  its  support  of  applied 
technology  research;  reducing  the  risk  borne  by  the  private  sector  through, 
for  example,  caps  on  liability;  and  through  fiscal  policies,  such  as 
depreciation  rates,  that  impact  the  capital  turnover  rate. 

We  have  for  some  time  now  existed  in  a  world  where  rates  of 
return  are  low  by  historical  standards.  In  such  a  world,  favorable  tax 
treatment  and  enhanced  public  sector  funding  of  applied  research  may  be 
appropriate.  Such  policies,  however,  would  have  to  be  predicated  on  the 
availability  of  a  set  of  technologies  whose  impact  is  broad  enough  to  have 
a  measurable  effect  at  the  macro-economic  level. 


REFERENCES 

Atack,  Jeremy.  1986.  “Finn  Size  and  Industrial  Structure  in  the  United  States  during  the 
Nineteenth  Century.”  Journal  of  Economic  History.  June,  46:2,  pp.  463  -  475. 

Balke,  Nathan  S.  and  Robert  J.  Gordon.  1989.  “The  Estimation  of  Prewar  Gross  National 
Product:  Methodology  and  New  Evidence.”  Journal  of  Political  Economy. 
February,  97:1,  pp.  38  -  92. 

Bemanke,  Ben.  2005  “Economic  Opportunity.”  Talk  before  the  National  Economists 
Club. 

Board  of  Governors  of  the  Federal  Reserve  System.  Various  years.  Flow  of  Funds 
Accounts.  Washington,  D.C.:  Federal  Reserve  Board. 

Board  of  Governors  of  the  Federal  Reserve  System.  Various  years.  Indexes  of  Industrial 
Production.  Washington,  D.C.:  Federal  Reserve  Board. 

Chandler,  Alfred  D.  1977.  The  Visible  Hand:  The  Managerial  Revolution  in  American 
Business.  Cambridge,  Mass.:  Belknap  Press  of  Harvard  University  Press. 

Cortada,  James  W.  1993.  Before  the  Computer:  IBM,  NCR,  Burroughs,  &  Remington 
Rand  &  the  Industry  They  Created,  1865-1956.  Princeton,  NJ.:  Princeton 
University  Press. 

David,  Paul  A.  1969.  “Transportation  Economics  and  Economic  Growth:  Professor  Fogel 
On  and  Off  the  Rails.”  Economic  History  Review.  December,  22:3,  pp.  506 
25. 


Spring  2006 


16 


David,  Paul  A.  1990.  “The  Dynamo  and  the  Computer:  An  Historical  Perspective  on  the 
Modem  Productivity  Paradox.”  American  Economic  Review.  May,  80:2,  pp.  355 
-61. 

Engerman,  Stanley  and  Kenneth  L.  Sokoloff.  2000.  “Technology  and  Industrialization 
1790-1914,”  in  Cambridge  Economic  History  of  the  United  States,  Volume  2. 
Stanley  Engerman  and  Robert  Gallman,  eds.  Cambridge:  Cambridge  University 
Press,  pp.  367  -  401. 

Ferguson,  Roger  W.  and  William  L.  Wascher.  2004.  “Lessons  from  the  Past  Productivity 
Booms.”  Journal  of  Economic  Perspectives.  Spring,  18:2,  pp.  3  -  28. 

Field,  Alexander  J.  2003.  “The  Most  Technologically  Progressive  Decade  of  the 
Century.”  American  Economic  Review.  September,  93:4,  pp.  1399  1413. 

Fischer,  Stanley.  1993.  “The  Role  of  Macroeconomic  Factors  in  Growth.”  Journal  of 
Monetary  Economics.  December,  32:3,  pp.  485  -  512. 

Fishlow,  Albert.  1966.  “Productivity  and  Technological  Change  in  the  Railroad  Sector, 
1840-1910,”  in  Output,  Employment,  and  Productivity  in  the  United  States  after 
1800.  Dorothy  S.  Brady,  ed.  New  York:  National  Bureau  of  Economic 
Research,  pp.  583  -  646. 

Fishlow,  Albert.  2000.  “Internal  Transportation  in  the  Nineteenth  and  Early  Twentieth 
Centuries,”  in  Cambridge  Economic  History  of  the  United  States,  Volume  2. 
Stanley  Engerman  and  Robert  Gallman,  eds.  Cambridge:  Cambridge  University 
Press,  pp.  543  -  642. 

Fogel,  Robert  W.  1979.  “Notes  on  the  Social  Saving  Controversy.”  Journal  of  Economic 
History.  March,  39:1,  pp.  1  50. 

Gordon,  Robert  J.  1992.  “Productivity  in  the  Transportation  Sector”,  in  Output 
Measurement  in  the  Service  Sectors.  National  Bureau  of  Economic  Research 
Studies  in  Income  and  Wealth.  Zvi  Griliches,  ed.  Chicago:  University  of 
Chicago  Press,  pp.  371  -  422. 

Gordon,  Robert  J.  2000.  “Does  the  ‘New  Economy’  Measure  Up  to  the  Great  Inventions 
of  the  Past?”  Journal  of  Economic  Perspectives.  Fall,  14:4,  pp.  49  -  74. 

Griliches,  Zvi.  1988.  “Productivity  Puzzles  and  R&D:  Another  Nonexplanation.”  Journal 
of  Economic  Perspectives.  Fall,  2:4,  pp.  9  -  21. 

Hawkins,  David  F.  1963.  “The  Development  of  Modem  Financial  Reporting  Practices 
among  American  Manufacturing  Companies.”  Business  Histoiy  Review.  Winter, 
37:1,  pp.  135  -68. 

Johnston,  R.  B.  1982.  The  Economics  of  the  Euro-Market.  New  York:  St.  Martin's  Press. 

Keeler,  Theodore  E.  and  John  S.  Ying.  1988.  “Measuring  the  Benefits  of  a  Large  Public 
Investment:  The  Case  of  the  U.S.  Federal-Aid  Highway  System.”  Journal  of 
Public  Economics.  June,  36:1,  pp.  69  -  85. 

Kendrick,  John  W.  1961.  Productivity >  Trends  in  the  United  States.  Princeton,  NJ.: 
Princeton  University  Press. 

Kindleberger,  Charles  P.  1993.  A  Financial  History  of  Western  Europe.  New  York: 
Oxford  University  Press. 

Kleinknecht,  Alfred.  1987.  Innovation  Patterns  in  Crisis  and  Prosperity:  Schumpeter's 
Long  Cycle  Reconsidered.  New  York:  St.  Martin's  Press. 

Mansfield,  Edwin.  1965.  “Innovation  and  Technical  Change  in  the  Railroad  Industry,”  in 
Transportation  Economics.  New  York:  National  Bureau  of  Economic  Research, 
pp.  169-97. 


Washington  Academy  of  Sciences 


17 


1 


Margo,  Robert  A.  2000.  “The  Labor  Force  in  the  Nineteenth  Century,”  in  Cambridge 
Economic  History  of  the  United  States,  Volume  2.  Stanley  Engerman  and  Robert 
Gallman,  eds.  Cambridge:  Cambridge  University  Press,  pp.  207  43. 

Mensch,  Gerhard.  1979.  Stalemate  in  Technology:  Innovations  Overcome  the 
Depression.  Cambridge,  Mass.:  Ballinger. 

Meyer,  Peter  B.  2003.  “Episodes  of  Collective  Invention.”  Bureau  of  Labor  Statistics 
Working  paper,  Washington,  D.C.,  August. 

Miranti,  Paul  J.  Jr.  2001.  “U.S.  Financial  Reporting  Standardization:  1840-2000.”  World 
Development  Report  2002:  Institutions  for  Markets.  Washington,  D.C.:  World 
Bank. 

Mowery,  David  and  Nathan  Rosenberg.  2000.  “Twentieth-Century  Technological 
Change,”  in  Cambridge  Economic  History  of  the  United  States,  Volume  3. 
Stanley  Engerman  and  Robert'  Gallman  eds.  Cambridge:  Cambridge  University 
Press,  pp.  803  925. 

Nelson,  Richard  R.,  Merton  J.  Peck  and  Edward  D.  Kalachek.  1967.  Technology, 
Economic  Growth,  and  Public  Policy.  Washington,  D.C.:  The  Brookings 
Institution. 

Paullin,  Charles  O.  1932.  Atlas  of  the  Historical  Geography  of  the  United  States. 
Washington,  D.C.:  Carnegie  Institute  and  American  Geographical  Society. 

Rudebusch,  Glenn  D.  and  David  W.  Wilcox.  1994.  “Productivity  and  Inflation:  Evidence 
and  Interpretations.”  Unpublished  manuscript.  Board  of  Governors  of  the 
Federal  Reserve  System. 

Schmookler,  Jacob.  1966.  Invention  and  Economic  Growth.  Cambridge,  Mass.:  Harvard 
University  Press. 

U.S.  Census  Bureau.  1997.  Historical  Statistics  of  the  United  States  on  CD-ROM: 
Colonial  Times  to  the  Present.  Susan  Carter,  Scott  Gartner,  Michael  Haines, 
Alan  Olmstead,  Richard  Sutch  and  Gavin  Wright,  eds.  New  York,  N.Y: 
Cambridge  University  Press. 

White,  Eugene  N.  2000.  “Banking  and  Finance  in  the  Twentieth  Century,”  in  Cambridge 
Economic  Histoiy  of  the  United  States,  Volume  3.  Stanley  Engerman  and  Robert 
Gallman,  eds.  Cambridge:  Cambridge  University  Press,  pp.  743  -  802. 


Spring  2006 


Page  intentionally  left  blank 


Washington  Academy  of  Sciences 


19 


LEARNING  FROM  THE  EUROTUNNEL* 

TO  BENEFIT  MOONBASE  DEVELOPMENT 

Stephen  Weil 
TI  Partners,  London 


Abstract 

A  manned  facility  on  the  Moon  would  likely  be  comparable  in  scale  in 
its  early  days  to  the  infrastructure  developed  in  the  early  1 990s  for  the 
Eurotunnel  connecting  France  and  the  United  Kingdom.  That  consisted 
of  three  railway  tunnels  under  the  Channel  and  the  marshalling  yards 
and  associated  service  facilities  at  the  ends  of  the  tunnel.  Many  of  the 
issues,  especially  funding  issues,  raised  by  a  Moonbase  development 
are  similar  to  issues  raised  by  the  Channel  tunnel.  Review  of  decisions 
made  by  Eurotunnel  developers  could  help  Moonbase  developers  avoid 
some  mistakes. 


A  MANNED  FACILITY  CONSTRUCTED  ON  THE  MOON,  which  WC  will 
call  “the  Moonbase,”  would  at  the  outset  presumably  consist  of  sizeable 
infrastructure  to  support  a  transport  hub,  including  shuttle-docking, 
launch,  and  repair  facilities;  a  medical  facility;  a  rest  and  recreation 
facility;  and  communications  and  security  facilities. 

Quite  quickly,  it  is  expected,  the  Moonbase  would  develop  a 
hospitality  facility  for  visitors;  research  laboratories  for  R&D;  an 
academic  facility;  several  observatories;  and  the  infrastructure  necessary 
to  support  a  growing  mineral  extraction  industry  and  related  construction 
facilities. 

The  scale  of  the  Moon-side  facilities  is  likely  to  be  comparable,  even 
in  the  earliest  days,  to  the  infrastructure  created  in  the  early  1990s  on  the 
French  side  of  the  railway  tunnels  under  the  Channel  connecting  France 
with  the  United  Kingdom.  The  French  facilities  include  major  railway 
marshalling  yards,  a  huge  (by  European  standards)  retail  complex,  and 
office  buildings,  linked  to  the  European  motorway  network. 


This  paper  was  presented  at  the  MoonBase  Conference,  held  in  Washington,  D.C., 
October  2005,  co-sponsored  by  the  Academy,  the  Italian  National  Institute  of 
Astrophysics,  and  High  Frontier,  Inc. 


Spring  2006 


20 


On  account  of  the  scale  of  the  facilities  required  for  a  manned 
observatory  on  the  Moon,  it  is  useful  to  look  at  the  methods  of,  and 
lessons  learnt  from,  the  funding  of  the  railway  tunnels  connecting  France 
with  Great  Britain  as  a  useful  precedent  of  how  an  international,  public- 
private  project  of  such  great  strategic  significance  might  come  together. 
Many  of  the  issues  raised  by  the  Moonbase  -  issues  of  international 
relations,  concessions  to  private  sector  suppliers,  structure  of  the  supplier 
contracts,  commitments  from  users,  project  funding  -  were  addressed  and 
answered  (if  not  always  wisely,  as  we  shall  see)  by  Eurotunnel,  the 
Channel  Tunnel  operator. 

Some  readers  may  have  taken  the  train  between  London  and  Paris,  or 
between  London  and  Brussels;  for  those  not  familiar  with  the  Channel 
Tunnel,  the  Tunnel  infrastructure  has  two  railway  tunnels  plus  one  service 
tunnel  under  the  Channel,  with  real  estate  at  either  end  of  the  Tunnel  for 
retail,  office  and  distribution.  The  Tunnel  is  a  railroad-only  link,  allowing 
express  trains  to  link  directly  London  with  Paris  and  Brussels;  cars  and 
freight  are  carried  on  special  shuttle  vehicles. 

The  Channel  Tunnel  idea  shares  with  the  Moonbase  some  distinctive 
challenges: 

•  The  primary  role  of  both  projects  is  to  act  as  an  international 
gateway:  the  Channel  Tunnel  is  a  gateway  between  Britain  (50 
million  population  in  1987)  and  continental  Europe  (290  million 
population  in  1987),  while  the  Moonbase  will  provide  access 
between  the  Earth  (6  billion  population  in  2005)  and  the  Solar 
System.  As  a  result,  some  form  of  international  treaty,  and 
implicitly  political  backing  at  the  highest  levels,  is  needed  to  give 
the  project  authority  and  credibility. 

•  A  second  shared  feature  is  that,  due  to  the  nature  and  novelty  of 
the  physical  challenges,  the  up-front  sunk  costs  are  very  high,  and 
subject  to  significant  large  variations.  With  the  Tunnel,  the  cost 
was  projected  at  the  start  of  construction  to  be  $10.8  billion  US; 
actually,  the  cost  turned  out  to  be  $18  billion  US,  close  to  $30 
billion  US  in  current  terms. 

Today’s  Channel  Tunnel  was  three  years  in  formation,  from  the 
signing  of  an  international  treaty  by  France  and  the  United  Kingdom  in 
late  1984,  laying  the  legal  framework,  to  the  raising  of  most  of  the  funding 
at  the  end  of  1987.  There  had  been  a  prior  bid,  ten  years  earlier,  between 
1973  and  1975,  by  the  French  and  British  Governments  to  authorize  the 


Washington  Academy  of  Sciences 


21 


building  of  a  tunnel.  This  first  attempt  failed  when  both  governments  fell. 
The  second  attempt  won  out,  with  strong  leaders  in  power  in  both 
countries.  President  Francois  Mitterand  in  France  and  Prime  Minister 
Margaret  Thatcher  in  the  United  Kingdom.  Maybe  the  first  important 
lesson  for  the  Moonbase  is  this  imperative  for  strong  governments  to  give 
it  force. 

The  two  governments  succeeded  in  selling  the  idea  of  the  Tunnel  to 
their  voters.  They  did  this  in  part  by  announcing  that  there  was  to  be  no 
public  funding  for  the  construction  and  operation  of  the  Channel  link;  the 
project  was  to  be  realized  by  private  promoters.  This  took  away  the  sting 
of  much  of  the  opposition  to  the  project  as  a  cost  to  the  tax-payer.  By 
insisting  on  no  state  funding.  President  Mitterand  and  Prime  Minister 
Thatcher  switched  the  focus  of  the  debate  from  tax  and  cost  to  opportunity 
and  challenge.  This  is  the  second  lesson  for  realizing  the  Moonbase 

The  two  governments  were  careful  to  select  a  private  promoter 
through  an  “open”  bidding  process,  so  no  one  could  accuse  the 
governments  of  favoritism.  Three  bidders  emerged,  each  offering 
different  technical  solutions:  the  winner  was  the  Eurotunnel  consortium. 

Technically,  the  Tunnel  is  a  major  engineering  triumph.  Since  its 
1994  opening,  apart  from  one  fire  in  which  fortunately  no  one  died,  it  has 
had  no  major  structural  or  technical  problems. 

Financially,  the  Eurotunnel  company  has  been  a  fiasco  for  its  backers 
and  funders. 

Eurotunnel’s  core  private  sector  consortium  emerged  out  of  a  wide 
spread  of  construction  and  finance  industry  interests  from  France  and  the 
United  Kingdom.  Though  the  Project’s  credibility  demanded  the  inclusion 
of  major  corporations,  it  would  have  been  better  to  have  widened  the  core 
investor  group  to  include  users.  For  the  Moonbase,  we  might  think  of 
energy,  minerals  explorations  and  chemicals  corporations,  and  satellite 
tracking  businesses.  The  absence  of  users  in  the  Eurotunnel  consortium 
was  to  prove  a  fundamental  weakness  in  the  design  and  execution  of  the 
project. 

With  hindsight,  the  make-up  of  the  core  consortium,  narrowed  to 
suppliers,  was  the  first  error  for  a  project  on  this  scale  of  complexity  and 
novelty. 

The  core  members  subscribed  collectively  for  around  5  percent  of  the 
total  initial  funding  requirement.  The  balance  of  the  equity,  around  10 


Spring  2006 


22 


percent  of  the  total  initial  funding  requirement,  came  through  a  public 
equity  issue.  Raising  money  from  the  public  for  a  project  of  this 
magnitude  and  uncertainty,  before  building  work  was  completed  and  a 
track  record  established,  was  the  second  mistake  made  by  Eurotunnel. 
(Eurotunnel’s  then  managers  would  probably  have  argued  that,  given  the 
way  the  capital  markets  looked  at  that  time,  only  a  public  equity  offering 
was  viable.)  Anyway,  the  outcome  was  wild  speculation,  and  a  volatile 
shareholder  base  which  lacked  stability. 

A  premature  approach  to  the  public  market  was,  then,  the  second 
mistake  made  by  Eurotunnel  in  its  finance  structure. 

The  rest  of  the  funding,  around  85  percent,  was  bank  debt,  partly 
secured  by  contracts  from  the  railways  agreeing  to  pay  a  minimum  usage 
charge  for  the  first  12  years  of  operation.  Even  so,  the  greater  part  of  the 
bank  debt  servicing  hinged  on  revenue  streams,  projected  by  consultants, 
of  the  number  of  vehicles  and  freight  traffic  using  the  Tunnel,  and  the 
prices  their  owners  would  be  prepared  to  pay. 

As  it  turned  out,  the  projections  were  wrong  by  a  factor  of  one  half, 
primarily  because  no  one  foresaw  the  impact  of  budget  airlines  and 
Internet  bookings,  which  encouraged  more  traffic  to  cross  the  English 
Channel  by  air,  not  through  the  Tunnel. 

Too  much  reliance  on  consultants’  projections  for  a  new  business, 
rather  than  firm  user  contracts,  was  the  third  key  weakness  in  Eurotunnel’s 
financial  structure. 

What  can  the  Moonbase  learn  from  Eurotunnel’s  successes  (broad 
political  support,  technical  triumph)  and  its  mishaps  (losses  for  investors 
and  banks)?  The  decisions  taken  prior  to  start  of  construction  are  critical. 
We  can  identify  three  distinctive  stages  prior  to  the  start  of  building: 

Stage  1 

A  number  of  governments  agree  to  invite  commercial  concessions  for 
a  Moonbase,  with  a  clear  statement  that  there  will  be  no  state  funding  for 
the  Moonbase,  setting  out  an  outline  structure  of  a  transparent  concession 
bidding  process,  run  by  a  combination  of  public  officials  and  private 
sector  specialists,  and  monitored  by  an  independent,  internationally- 
recruited  Ethics  Committee,  in  turn  reporting  to  a  committee  of 
representatives  of  the  governments. 


Washington  Academy  of  Sciences 


1 


23 


Stage  2 

The  transparent  concession  bidding  process  takes  place,  subject  to 
overview  by  an  independent  Ethics  Committee  (similar  to  the  Ethics 
Committees  in  a  drug  company),  leading  to  the  selection  of  a  successful 
bidder. 

Stage  3 

The  contractor  negotiates  contracts  with  users  such  as: 

•  NASA  (R&R  for  astronauts,  hospitals,  monitoring, 
tracking.  Earth  observation), 

•  Energy  exploration, 

•  Minerals  exploration,  and 

•  Academic  research. 

The  contracts  are  contingent  upon  supply  of  facilities  by  a  given  date. 
In  a  parallel  process  the  contractor  puts  in  place  funding  contingent  on 
user  contracts;  and  insurance  to  cover  the  risk  of  technical  failure. 

Throughout  the  various  initial  stages  and  subsequently,  the  Ethics 
Committee  supervises  the  activity  of  the  contractor  to  check  for  any 
damage  to  the  environment  or  to  human  beings.  The  contractor  should  not 
be  allowed  by  the  financial  authorities  to  offer  its  shares  to  the  public 
before  an  operating  track  record  is  established,  while  firm  user  contracts 
should  provide  a  solid  base  for  the  debt  funding  of  the  program. 

If  structured  correctly  at  the  outset,  the  Moonbase  provides  an 
opportunity  for  a  major  cooperative  effort  among  governments, 
universities,  and  business  for  the  benefit  of  all. 


Spring  2006 


Page  intentionally  left  blank 


Washington  Academy  of  Sciences 


25 


ORGANIZATIONAL  EVOLUTION,  LIFE-CYCLE 
PROGRAM  DESIGN: 

ESSENTIAL  ISSUES  IN  SYSTEMS  ENGINEERING  AND  ACQUISITION  OF 

COMPLEX  SYSTEMS  * 

Gary  G.  Nelson 

Homeland  Security  Institute 


Abstract 

A  society  of  puiposeful  agents — humans  and  now  their  artificial 
symbionts — self-organizes  into  partitions  of  activity  (nations, 
corporations,  agencies,  services)  and  a  scale  hierarchy  of  governance. 
This  complex-adaptive  society  is  properly  evolutionary.  Such  a  society 
designs  “projects”  that  create  purposeful  artifacts  with  a  developmental 
life  cycle.  As  the  projects  become  more  and  more  ambitious,  the  real 
and  conceptual  boundary  between  the  evolutionary  and  the 
developmental — between  the  project  and  the  society  blurs.  Over  just 
the  last  seven  decades,  this  growth  in  “designed  complexity”  has 
spawned  the  design  of  formal  processes  of  systems  engineering, 
concurrent  with  formal  processes  of  acquisition  (allocation  of  and 
accountability  for  significant  social  resources).  The  most  interesting 
examples  of  such  projects  are  the  information  systems  for  decision 
support  that  become  the  social  linkage  of  indefinite  extent,  and  the  very 
means  of  collective  design  and  acquisition  of  projects.  So,  we  have  two 
essential  problems:  1)  The  blurring  of  designing  subject  and  designed 
object,  and  2)  a  recursive  relation  between  the  object  and  the  designing 
process.  Empirically  (from  experience  with  major  federal  programs), 
there  is  a  vast  conceptual  confusion  between  evolution  and 
development  (including  the  design  phase).  This  is  reflected  in 
persistent  problems  with  the  formal  engineering  and  acquisition 
processes.  Some  principles  of  complex  systems,  with  special  reference 
to  decision  support  systems,  are  articulated  to  identify  and  ameliorate 
the  very  practical  problems  encountered  in  the  life-cycle  development 
of  complex  designs  by  complex  organizations. 


There  is  nothing  uncommon  about  the  scene:  an  organization 
undertakes  a  project  of  some  complexity  in  order  to  fulfill  the  purpose  of 
the  organization.  To  be  concrete,  suppose  the  organization  is  a  federal 


*  Presented  at  CapitalScience  06,  March  25-26,  2006,  in  the  Washington  Evolutionary 
Systems  Society’s  Symposium  on  the  Emergences  of  Designs. 


Spring  2006 


26 


agency  and  the  project  is  a  multi-billion  dollar  automation 
/surveillance/communication  system  that  enables  the  agency  to  perform  its 
operational  mission  (e.g.,  homeland  security,  air  traffic  management, 
transportation  operation).  There  are  many  such  concurrent  projects  going 
on  at  any  time.  Several  end  up  in  the  newspapers  as  immense  failures.  I 
was  intimately  involved  in  one — the  Advanced  Automation  System 
(AAS)  for  air  traffic  management — in  its  early  design,  program 
management  and  acquisition  phases.  But  there  are  other  more  prosaic 
projects,  like  highway  or  rail  transit  construction,  that  collectively  alter  the 
physical  connectivity  of  society.  Both  informational  connectivity  through 
“cybernetic”  systems  and  geographical  connectivity  through  transport 
projects  represent  an  extraordinary  endeavor:  designed  systems  that  are 
inherent  in  the  emergence  of  the  designing  society  itself. 

Starting  in  the  1970s,  it  took  me  that  decade  to  realize  the  vast  gap 
between  reach  and  grasp  in  transport  planning:  the  projects  were  made  to 
look  small,  through  the  stated  regulatory  and  judicially-confirmed  doctrine 
of  “logical  termini.”  This  is  effectively  a  decomposition  of  the  network 
into  its  links,  one  at  a  time.  And  yet  all  the  purpose  and  rhetoric  about  the 
projects — individually  or  collectively — are  extrapolated  to  vast  social 
extents  to  cite  benefits  to  the  economy,  society,  and  even  the  environment 
(ironic  in  face  of  the  fact  that  most  tangible  environmental  impacts  are 
local  and  negative).  These  observations  soon  connected  with  the  work 
inspired  by  Ilya  Prigogine  and  conducted  by  Peter  Allen  [1978,  1981, 
1987]  under  the  sponsorship  of  Bob  Crosby  (founder  of  the  Washington 
Evolutionary  Science  Society)  who  was  then  at  the  U.S.  Department  of 
Transportation.  That  work  was  about  the  networks  that  self-organize  as 
our  geography,  polity,  and  economy.  I  was  introduced  to  that  work  by  my 
graduate  school  advisor,  Pitu  Mirchandani  in  1981.  The  connections  I 
started  making  between  what  I  had  observed  about  regional  transport 
planning  and  the  clear  principles  of  complex  systems  [Nelson/Allen, 
2001]  included: 

1.  Society  presumed  to  do  “long  range”  planning  about  regional 
development  (including  transport),  but  all  the  evidence  pointed  to 
this  being  nugatory  (neither  legally  effective  nor  factually  accurate 
in  prediction). 

2.  The  long  range  planning  was  a  legal  and  predictive  cover  for  a  host 
of  projects  (mostly  highway  expansions),  that  in  the  disaggregate 
had  only  the  most  tenuous  relation  to  the  regional  models. 


Washington  Academy  of  Sciences 


27 


3.  The  very  idea  of  planning  was  defeated  by  our  social  ideology  of 
localistic  laissez-faire. 

Compared  to  our  presumptions  about  large-scale  planning  of  any  kind, 
our  real  and  intuitive  models  of  society  are  much  better  fitted  to  the 
complex  system  that  society  is.  Both  our  economic-market  and  political- 
democratic  doctrines  emphasize  emergence  from  localistic  interaction. 
The  residual  debate  is  about  the  role  of  a  scale  hierarchy  of  governance  in 
what  otherwise  would  be  simply  a  one-scale  anarchy  of  interaction.  But 
the  scale  hierarchy  (groups  that  conceive  and  act  upon  aggregate  social 
objects,  or  sets  of  common  rules  of  interaction)  is  an  inherent  structure  of 
self-organizing  (i.e.,  complex-adaptive  or  evolutionary)  systems.  That 
much  I  realized  after  reading  Stan  Salthe  [1985,  1993]  with  whom  I  since 
have  engaged  in  many  fruitful  dialogs. 

In  the  1980s  I  joined  MITRE  and  became  engaged  in  the  AAS,  at 
that  time  the  most  ambitious  example  of  a  cybernetic  system  yet 
undertaken.  This  was  in  direct  line  of  succession  from  what  I  consider  to 
be  the  first  system-integration  effort  (the  British  air  defense  system  in 
World  War  II)  through  the  Semi-Automatic  Ground  Environment  (SAGE) 
system  from  which  MITRE  emerged  (1958)  and  the  original  digitized  air 
traffic  management  system,  called  National  Airspace  System  (NAS)  Stage 
A  (1973).  It  happens  that  air  traffic  management  is  a  fine  example  of  the 
tri-scaled  structure  posed  by  Salthe.  Being  involved  with  the  AAS  benefit- 
cost/risk  analysis  and  some  of  its  operational  planning  [Nelson,  1992]  I 
began  to  see  how  information  and  risk-decision  making  were  really  the 
key  evolutionary-development  analogues  between  social  activity,  the  true 
biological  structure,  dynamics  of  evolution,  and  development. 

More  recently,  I  have  been  concerned  with  facilitating  groups  in 
complex  system  engineering  processes  that  I  center  on  the  concept  of 
operations  (CONOPS)  portion,  for  reasons  that  will  be  elaborated.  I  have 
also  witnessed  what  I  consider  to  be  the  third  phase  of  development  of  the 
system  engineering  process,  between  the  early  1960s  and  now.  Since  the 
1980s,  the  acquisition  process — taken  as  the  budgeting  and  expenditure  of 
public  funds  for  the  complex  system  projects — has  also  become  much 
more  formalized,  culminating  in  the  OMB  Exhibit-300  requirements  for 
federal  projects  and  including  “reference  models”  with  appeal  to  complex 
system  theory.  The  following  observations  emerge: 

A.  Complexity,  in  projects  or  processes,  tends  to  be  laid  out  in  linear, 
sequential  ways  that  lose  important  attributes  of  the  complex 


Spring  2006 


28 


system  and  process.  This  is,  of  course,  responsive  to  projects  that, 
however  complex,  are  purposefully  designed  and  have  finite  life 
cycles  because  the  purpose,  their  environment,  or  their 
functionality  changes. 

B.  There  is  much  talk  about  “evolutionary  development”  that 
conceptually  tries  to  bridge  the  ongoing  project  environment 
(including  the  designers  as  well  as  interfacing  systems)  and  the 
finite  designed-project.  A  small  fraction  of  people  who  espouse  the 
concept  know  what  they  are  talking  about  (as  witnessed  by  making 
“to  evolve”  a  transitive  verb). 

C.  The  problems  with  the  acquisition  process,  much  more  than  the 
system  engineering  process,  are  evident,  acute  and  expensive. 
They  could  be  addressed  by  the  necessary  fusion  of  evolution  and 
development  as  evo-devo  (to  adopt  another  biological  metaphor). 

The  theme  of  “design”  versus  evolution  is  current.  Strictly,  design  of  a 
system  is  just  one  phase  of  its  life  cycle  development  and  not  part  of 
acquisition.  The  thrust  of  this  paper  is  that  the  fundamental  dichotomy  is 
in  the  original  biological  concepts  of  evolution  (referring  to  the  processes 
of  genetic  change)  and  development  (referring  to  the  life  cycL  of  the 
individual  organism,  or  proteome  or  phenotype  tried  by  the  genor  The 
important  principles  concern  how  the  purposeful  organization  (at  multiple 
scales)  persists  robustly  through  the  mechanism  of  instantiating  a  number 
of  trials  of  interaction  in  the  common  environment. 

An  Image  of  Evo-Devo 

Evo-Devo  is  a  network.  In  a  geographical  metaphor,  it  is  the 
emergent  “fitness  landscape”  that  incorporates  the  dynamic  competition 
between  transient  niches  that  persist  long  enough  to  be  seen  as  real  and 
reproducing  phenotypes.  The  genome  is,  of  course,  physically  instantiated 
with  the  phenotypes.  The  conceptual  trick  leading  to  the  scale  hierarchy  is 
that  the  genome  also  has  a  persistence  and  emergent  identity  above  the 
physical  instantiations.  At  that  point,  our  metaphors  become  muddled:  the 
genome  is  a  design  for  the  phenotype,  but  does  not  purposefully  design  the 
phenotype.  Anything  real  and  physical  is  recursive  between  parts  of 
organisms,  organisms,  and  the  rest  of  the  physical  environment  that 
organisms  shape. 

Let  us  shift  what  we  can  from  the  biology  to  a  particular  set  of 
organisms  (us)  who  are  purposeful  agents  (decision  makers).  But  in  this 


Washington  Academy  of  Sciences 


29 


information  age,  the  “us”  will  be  extended  to  artificial  agents.  Here  is  an 
axiomatic  statement  of  that  evo-devo  network: 

1.  Agents  are  linked  by  information. 

2.  The  information  is  created  pairwise  by  agents  making  decisions, 
and  each  agent  is  a  node  where  information  is  received  and 
created. 

3.  Every  decision  is  informed  by  received  information  (this  excludes 
decisions  totally  indifferent  to  the  environment).  This  means  that 
every  decision  is  a  risk  decision:  there  is  uncertainty  relative  to  the 
value,  or  “utility”  of  the  decision  to  an  agent,  and  the  uncertainty  is 
resolved  in  whole  or  part  by  received  information. 

4.  The  emergent  network  of  agent  nodes  and  information  - 
communication  links  is  a  congested  network.  It  is  the  environment 
of  the  agents,  including  all  their  information  inputs  and  outputs, 
and  so  partially  the  result  of  the  agents  themselves. 

5.  The  network  has  an  associated  physics  of  mass/energy,  but  this  is 
either  irrelevant  to  the  logical  agents  and  information,  or  interfaces 
to  the  agent  network  by  sensors  and  actuators  via  risk  decisions. 

This  set  of  axioms  was  reduced  after  long  consideration  about 
what  is  essential  to  complex  organizations  and  their  projects.  Yes, 
physical  outcomes  matter  and  are  the  ultimate,  objective  measures  of 
“fitness.”  But  all  our  processes  and  how  we  act  to  design,  develop,  and 
acquire  are  concerned  with  the  logical  network  of  agents,  as  are  design- 
objects  that  are  decision  support  systems  (including  communication  and 
information  systems,  whose  purpose  is  to  inform  risk  decisions). 

These  axioms  may  seem  a  long  way  from  practical  matters  of  how 
evolutionary  organizations  (sub  networks  as  described)  develop  projects. 
But  the  axioms  imply  structures,  dynamics  and  deductions  (principles) 
that  get  directly  to  the  problems  of  complex  systems  trying  to  develop 
complex  systems.  A  few  relevant  expansions  of  the  axioms  will  be  made. 

Risk  Decisions 

There  are  at  least  three  qualitatively  different  definitions  of  risk. 
The  axioms  consist  with  an  assertion  that  risk  is  the  uncertainty  around  a 
payoff  to  an  agent  making  a  decision.  And  although  there  have  been 
decades  of  debate  about  even  this  notion  of  risk,  it  is  largely  consistent 
with  the  view  taken  by  game  theory  and  economics. 


Spring  2006 


30 


We  can  link  the  network  of  agents  with  learning  concepts,  because 
that  also  is  essential  to  evo-devo:  agents  or  their  organizations 
demonstrate  fitness  by  “learning”  to  promote  some  local  measure  of  value 
(e.g.,  fertility,  wealth,  happiness,  power).  Those  familiar  with  the  concept 
of  “utility”  will  realize  that  the  measure  can  be  internalized  and 
completely  subjective,  with  no  real  explanatory  power.  It  is  important  that 
groups  of  agents  tend  to  demand  objective  value  measures  regarding  the 
use  of  joint  resources  (including  information).  But  such  measures  internal 
to  a  group  can  be  just  as  metaphysical  as  utility.  Benefit-cost  analysis  for 
public  projects  is  a  case  in  point,  where  the  best  we  can  do  is  translate 
supposed  distributed  outcomes  (physical  or  not)  into  dollars  that  is  just  an 
abstract  numeraire,  and  furthermore  outside  of  real  markets  that  are  the 
only  source  of  “value”  for  such  a  numeraire. 

Game  theory  is  the  best  context  for  discussing  the  network  of 
interacting  agents.  Each  decision  is  a  commitment  to  a  “move.”  In 
complex  organizations,  there  are  usually  extended  path(s)  through  agents 
(sequential  and  parallel  decision  chains)  to  get  to  a  physical  product  (and 
with  or  without  relevant  physical  data  to  begin  with!).  Therefore,  a 
commitment  or  move  may  be  taken  as  either  the  sending  of  information  or 
(rarely)  the  creation  of  a  real  outcome-producing  physical  state  of  the 
environment.  Each  agent  considers  the  agent-local  value  of  potential 
decisions.  Because  the  chains  are  long  to  ultimate  outcomes,  they  are 
subject  to  other  decisions  (i.e.,  information  not  yet  available),  hence 
uncertain.  Uncertainty  is  just  the  absence  of  valuable  information.  This 
puts  the  game  in  a  Bayesian  framework:  the  value  of  incoming 
information  is  always  relative  to  how  it  affects  a  hypothesis  (predicted 
value)  of  an  agent.  This  is  how  and  why  information  is  created  pairwise 
among  agents.  Information  accrues  value  only  in  the  transaction  between 
an  agent  in  one  “hypothesis  state”  and  another. 

But  this  also  gets  back  to  the  basic  Shannon/Hartley/Szilard 
definition  of  information.  Information  is  defined  by  its  resolution  of 
uncertainty.  But  when  we  analyze  what  this  means,  it  is  about  an  outcome 
on  a  channel  with  two  agents  (source  and  receiver).  A  physical  bit  means 
nothing  by  itself.  The  logical  bit  referred  to  here  has  to  inform  by  being 
part  of  a  message  whose  meaning  is  defined  by  the  payoff  to  the  receiving 
(and  decision  making)  agent.  But  it  equally  requires  the  participation  of 
the  sending  agent,  for  whom  the  transmission  of  the  bit  is,  in  fact,  a 
decision. 


Washington  Academy  of  Sciences 


31 


The  approach  here  clarifies  the  difference  between  physical 
information  (unpredictable  channel-state  change)  and  logical  information 
(or  valuable  “meaning”  to  an  agent).  That  difference  is  essential  to  the 
network  of  agents,  how  it  structures  and  how  it  changes  (adapts,  or  fails 
to).  Via  game  theory  we  can  start  getting  an  “economic”  explanation  of 
the  network.  But  essentially,  it  is  always  about  the  allocation  of  scarce 
resources.  In  discussing  projects,  the  allocation  is  about  budgets  to 
projects  and  the  agents  are  people  with  authority  to  contribute  resources 
that  are  either  theirs  personally  or  “public.”  In  the  latter  case,  agent  value 
about  power  and  position  (their  ex  officio  identity)  are  extremely 
important  to  what  happens.  In  order  to  get  to  the  essence  of  development 
and  acquisition  processes,  we  need  to  see  the  agents  as  the  congested 
network  contending  over  abstract  payoffs  via  information  (although 
fisticuffs  cannot  be  excluded  entirely). 

This  model  of  a  congested  network  of  risk  decisions  has  a  very 
important  generalization:  It  does  not  matter  whether  the  agents  are  within 
a  “collaborative”  organizational  identity  or  not.  It  works  across  all 
contentions  over  some  common  resource.  It  works  for  antagonistic 
economic  desires,  and  it  works  for  terrorists  versus  securers.  The  network 
is  marked  by  what  does  emerge  as  the  common  context  of  interaction,  and 
that  is  what  is  local  to  organizational  clusters.  But  the  problems  faced  by 
the  organizations  cross  organizations.  Since  so  much  of  the  systems 
approach  was  developed  for  weapons,  this  is  the  general  model  we  need. 
Both  the  process  of  development  and  the  objects  of  development  are  about 
resource  contention  games.  Improvements  in  play  are  about  the 
information  to  support  risk  decisions  and  that  necessarily  involves 
contending  players. 

Scale  Hierarchy 

A  flat  view  of  the  agents  in  the  network  is  inadequate.  Every 
organization  has  a  management  hierarchy.  This  responds  to  the  emergence 
of  scales  from  partitioned  interaction. 

The  scale  hierarchy  is  a  structure  of  all  evo-devo  systems,  and  the 
congested  network  of  agents.  It  follows  from  the  Simon-Ando  principle  of 
partial  decomposability  [1961]:  Self-organized  networks  emerge  up-scale 
from  modular  interaction  that  overcomes  the  combinatorial  barriers  to 
unitary  self-assembly.  The  modules  themselves  have  some  mechanism  to 
instantiate  each  interaction.  We  have  here  a  conceptual  triad  that  resonates 
with  some  other  triadic  concepts  [Salthe/Nelson,  in  progress]: 


Spring  2006 


32 


Peircean  Categories 

Aristotelian  Causes 

Scale  Hierarchy  Triad 

Thirdness  (emergence) 

Final 

Emergent  Context 

(network) 

Secondness  (connection) 

Formal/Efficient 

Collective  of  peer  modules 

Firstness  (object) 

Material 

Mechanism 

There  are  rich  connections  and  implications  of  this  table.  From  the 
Peircean  perspective,  any  design  is  about  instantiating  a  concept  via  the 
connection  of  parts  whose  whole  corresponds  to  the  concept.  For  self¬ 
organization,  absent  a  designer,  we  just  eliminate  the  first  step  that  is 
recursive  with  the  emergent  result  anyway.  From  the  causes,  finality 
corresponds  to  the  concept  and/or  emergent.  The  material  cause  is  what  is 
assembled.  Formal  and  efficient  causes  must  be  seen  as  a  dialectic,  and 
can  be  interpreted  as  what  an  agent  decides  to  do  (efficiency)  being 
“formed”  by  information  from  the  emergent  environment. 

The  formal/efficient  dialectic  is  fundamental  to  either  biological 
development  (morphogenesis)  or  any  product  development.  An  engineer 
perceives  a  time  series  profile  in  design  and  construction  in  which  the 
early  phase  is  dominated  by  imposing  the  concept  on  unorganized  material 
(efficiency).  But  then  as  the  product  emerges,  it  is  actually  the  product  that 
highly  constrains  (cf.  “decides”)  the  further  organization  of  materials  and 
the  balance  shifts  to  formality.  This  time-series  profile  of  agency  (hence 
information)  is  essential  to  every  life  cycle  engineering  and  acquisition 
process.  The  locus  of  information  creation  that  shifts  from 
design/efficiency  to  product/formality  is  what  creates  and  then  resolves 
program  risk.  Risk  management  is  about  the  process  by  which  the 
information  is  created  over  time  with  respect  to  resources.  Fort  this 
reason,  a  normal  expenditure/progress  profile  in  earned  value  management 
(EVM)  is  S-shaped  like  the  profile  that  can  be  defined  for  the  entropy  rate 
of  any  development  [Salthe,  1993].  Organizations  show  this  profile:  when 
they  are  formative,  they  are  efficiently  creating  the  structure  to  adapt  to 
the  niche  they  are  filling.  When  mature,  the  organization  is  dominated  by 
its  own  structure,  that  is  good  at  what  it  was  originally  doing  but 
maladaptive  to  a  changing  environment. 

The  scale  hierarchy  has  manifold  interpretations.  For  the  congested 
network  of  agents,  what  we  visualize  as  the  tinker-toy  network  of  agents 
transacting  information  is  the  focal  scale  and  the  collective  of  peer  agents. 
They  partition  (decompose)  the  information  of  the  entire  network  into 
local  information  and  common  protocols  (rules,  laws,  norms).  The  fact  of 
each  agent’s  state  is  the  mechanism  by  which  the  pairwise  transactions  are 


Washington  Academy  of  Sciences 


33 


made  among  the  collective.  In  the  market  model  we  would  say  that  the 
context  scale  are  property  laws  and  price,  the  focal  collective  is  the  market 
itself  of  producers  and  consumers  exchanging  goods,  and  the  mechanism 
scale  is  the  production  process  and  processed  commodities  owned  by  each 
producer.  Note  that  the  goods  are  emergent  from  the  production  process, 
and  indeed  we  can  include  pure  money  or  information  transactions  here. 
There  is  a  similar  qualitative  shift  between  each  pair  of  scales. 

In  the  pure  information  network,  the  information  transacted  is 
essentially  the  predictor  of  the  source-agent’s  state,  and  that  state  is 
produced  by  the  source  agent’s  own,  hidden  mechanism.  That  is  the 
valued  information  in  the  risk-decision  game  of  commitments.  What  the 
receiving  agent  “owns”  as  a  result  of  the  transaction  is  the  power  to 
achieve  a  payoff  by  use  of  this  information.  The  source  agent  may  be  an 
opposing  player  or  a  surrogate  for  such  a  player  in  a  long  chain  of 
prediction.  This  model  makes  no  distinction  about  whether  the  ultimate 
opponent  is  another  sentient  agent  or  a  state  of  nature.  Having  a  prediction 
about  rain  serves  a  similar  payoff  to  a  prediction  that  someone  will  spray 
you  with  a  hose.  The  same  applies  to  preparedness  for  natural  disasters  or 
intentional  terror. 

Now,  self-similarity  of  scales  implies  that  if  information  of  value 
can  emerge  from  some  mechanism  to  a  focal  scale,  the  same  can  be 
replicated  over  several  scales.  This  applies  to  any  modular  or 
organizational  decomposition.  It  is  why  we  get  several  scales  of 
management  in  a  big  corporation  or  nation.  The  principle  of  prediction 
among  peer  agents,  no  matter  what  scale  they  represent,  is  the  same.  The 
scales  result  from  the  partitioning  of  local  information  and  there  will  be  a 
number  of  scales  according  to  the  total  information  in  the  total 
environment  (within  any  defined  system  and  its  external  interfaces).  The 
partitioning  results  from  the  finite  predictive  capability  of  any  module  or 
agent.  And  the  “finite”  limit  is  the  risk  tolerance  of  the  payoff.  No 
efficient  agent  will  decide  when  the  limits  are  so  large  that  decisions  are 
really  random.  No  system  will  “learn”  and  adapt  under  such 
circumstances.  But  this  limit  scales:  higher  managers  will  be  very 
uncertain  about  the  individual  state  of  agents  a  few  levels  down.  What 
they  have  to  be  reasonably  sure  about  is  the  performance  of  the 
organization  over  a  strategic  scale.  To  achieve  this,  formal  constraints  are 
imposed  (rules,  directives,  guidance,  physical  facilities)  on  lower  scales 
that  act  like  mechanisms  to  higher  scales. 


Spring  2006 


34 


There  are  many  fine  points  here  about  the  dynamics  of  the  total 
environment.  It  is  assumed  that  the  environment  is  non-equilibrium,  and 
statistically  non-stationary.  That  means  that  all  predictions  are  fallible,  in 
terms  of  exceeding  risk  limits.  This  time  series-effect  is  also  reflected  on 
the  ensemble  of  agents.  There  is  a  well-known  distribution  of  exploiters 
and  explorers  among  collectives  of  agents  (or  risk-avoiders  and  risk- 
takers).  This  amounts  to  the  variety  in  phenotypical  trials  that  creates  and 
occupies  niches.  This  is  how  any  relative  context  scale  learns  in  a  non- 
stationary  environment.  What  persists  (and  hence  what  we  see)  are  good 
mixtures  of  agent  types  organized  in  ways  that  lets  the  exploiters  get 
resources  for  the  organization  under  a  given  environment  and  lets  the 
explorers  change  the  organization  as  the  environment  changes. 

Layers 

The  scale  hierarchy  is  exhibited  as  layers  of  management  or  self¬ 
similar  layers  of  decomposition.  But  in  information  architectures,  layers 
have  another  meaning.  They  are  a  decomposition  of  an  end-to-end 
information  transaction,  or  channel.  For  the  purposes  here,  three  layers  are 
important: 

1 .  The  network  of  logical  agents,  who  value  information  and  make 
decisions  (create  information). 

2.  Decision  support  applications  that  assist  the  agents  by  processing 
the  valuable  information  (and  it  is  here  that  a  network  of  artificial 
agents  may  be  placed  relative  to  human  agents). 

3.  The  communication  of  information  in  channels  that  far  from 
creating  or  using  information  are  expected  to  slavishly  pass  it 
along  unaltered. 

The  third  layer  is  the  object  of  “information  technology”  and 
communications.  It  has  its  own  decompositions,  e.g.  into  a  seven  layer 
stack,  or  some  variation  on  those,  that  constitutes  the  Internet  (TCP/IP 
protocol  layers).  And  the  bottom  layer  is  always  a  physical  medium  that 
can  take  on  physical-bit  states. 

These  layers  must  be  kept  in  mind  for  proper  focus  in  the  scale 
hierard  There  may  be  any  number  of  scales  (layers)  of  logical  agents  or 
modules  in  a  system.  They  all  look  “flat”  if  viewed  as  their  physical 
instantiation  (people,  a  machine).  Anyone  in  a  suit  could  be  an  executive 
or  applicant  for  the  mail  room.  Similarly,  the  means  of  interaction  may  be 


Washington  Academy  of  Sciences 


35 


a  common  communications  infrastructure.  This  is  implicit.  How  that 
communications  works  (physical  mail  versus  Internet,  landline  telephone 
versus  cellphone,  horse  versus  airplane)  is  a  powerful  environmental 
context  for  the  agent  society.  And  so  a  society  of  designing  agents  may  be 
trying  to  design  a  system  for  interaction  (e.g.,  a  large-scale 
communications  and  computer  network  for  decision  support,  or  a  new 
transport  network)  that  will  alter  how  they  interact,  or  even  how  they 
design  such  systems.  This  evokes  a  more  radical  dialectic  of  formal  and 
efficient  cause:  It  puts  subject  and  object  at  the  same  focal  scale  of 
interaction. 


Where  is  Design  in  Complex  Systems? 

The  challenge  is  not  a  simple  widget  that  is  designed  and 
produced.  The  reason  for  the  interest  in  elaborate  system  engineering,  or 
life-cycle  development  and  acquisition  processes,  is  the  ambition  to  create 
complex  systems  with  the  ability  to  transform  the  very  organization  that 
designs  and  interacts  with  that  system.  This  is  the  sense  in  which 
“business  process  reengineering”  or  “enterprise  transformation”  is  used 
now  in  the  systems  community.  But  even  if  the  effect  of  the  system  is  not 
recursive,  the  organization  that  manages  a  complex  development  for 
“someone  else”  will  be  challenged  and  transformed  by  its  interaction  with 
its  object. 

The  past  decades  have  been  dominated  by  a  shift  from  the 
“waterfall”  model  of  systems  design  and  development  to  other  models  that 
encounter  the  confusion  between  evolution  and  development  at  the  same 
time  they  claim  to  both.  The  “waterfall”  represents  a  progressive 
instantiation  of  a  product  system  from  concept  to  operation.  But  in  fact 
this  must  always  occur  if  anything  real  is  to  be  produced  purposefully  (by 
design).  The  waterfall  may  be  associated  with  any  relation  between  an 
agent  and  the  mechanism  scale  in  the  scale  hierarchy,  because  every  agent 
locally  wants  to  see  a  progressive  sequence  toward  an  output  of  some 
kind.  The  confusion  arises  by  failing  to  put  that  into  a  full  scale  hierarchy 
where  most  focal  scales  have  an  iterative  interaction  with  peers,  and 
emergent  context  (that  is  iterative  with  the  focal  scale),  and  a  mechanism 
scale  that  is  not  the  final  product  instantiation.  More  often,  an  agent  is 
only  serving  as  the  coordinative  context  (management)  of  a  further  set  of 
peers.  The  ability  to  trace  what  appears  to  be  a  linear  thread  of  increasing 
specification  toward  instantiation  of  a  product  is  only  an  illusion  absent 
the  back  and  forth,  mistakes,  revisions  and  reconceptualizations  that  occur 
in  reality. 


Spring  2006 


36 


The  alternative  to  the  waterfall  is  evo-devo,  and  that  usually 
reduces  to  some  sort  of  “spiral”  concept.  Spirals  refer  to  an  image  of  a 
concept  going  through  successive,  tentative  instantiations  (e.g.,  rapid 
prototypes).  There  is  a  learning  feedback  that  adds  information  to  a  next 
phase  that  can  go  back  to  a  concept  before  progressing  to  a  more 
substantial  prototype  product.  This  is  a  good  approach,  but  all  it  is  doing  is 
formalizing  the  back-and-forth  that  always  occurred  in  the  engineering  of 
complex  systems  and  that  a  strict  waterfall  image  just  overlooks. 

Where  is  “design”  in  this?  The  best  we  can  say  is  that  it  represents 
a  transient  phase  local  to  each  agent  where  a  causal  relation  between  a 
concept  and  an  output  is  perceived.  But  that  applies  to  every  agent 
transaction.  Most  such  transactions  are  purely  informational:  a  decision  is 
in  fact  a  “design”  of  what  message  string  will  be  sent.  But  we  tend  to  think 
of  design  as  applying  to  a  physically  instantiated  product.  In  complex 
systems  that  design  is  as  decentralized  and  time-spread  as  the  agents 
themselves.  This  is  why  we  distinguish  self-organization  and  evolution 
from  design.  The  difficulty  arises  when  we  try  to  apply  a  causal  concept  of 
progressive  specification  to  a  collective  of  agents  in  any  complex  system. 

The  Design/Development/Acquisition  Life-Cycle  Process 

How  does  the  empirical  evidence  of  any  dew  ment  process 
compare  with  the  scale  hierarchy  theory?  All  formal  \  ocesses  define 
levels  of  interaction  that  progress  from  the  conceptual  to  the  formal 
instantiation.  These  levels  or  scales  are  roughly  as  follows: 

•  A  scale  deciding  what  to  do  (mission). 

•  A  systems  integration  scale  that  defines  modules  of  functionality 
and  is  concerned  with  their  interfaces  to  fulfill  the  mission. 

•  Teams  concerned  with  further  specifying  each  module,  meaning 
that  further  modular  levels  and  interfaces  are  created. 

•  Until  we  reach  the  scale  where  a  module  is  instantiated  and 
assembled  into  th-  ole  system  that  fulfills  the  mission. 

The  acquisition  process  generally  jumps  in  at  some  level  of 
maturity  of  modular  specification.  This  raises  an  issue  of  how  resources 
are  allocated  to  research  as  opposed  to  what  is  strictly  considered 
acquisition  of  a  product.  But,  at  some  point  in  a  supposed  progressive 
instantiation  of  a  design,  an  interest  is  excited  between  engineers  and 
executives  who  then  are  synchronized  into  series  of  benchmark  progress 


Washington  Academy  of  Sciences 


1 


37 


reviews  and  the  risk  management  of  the  formal/efficient  information 
profile.  We  simply  want  the  development  money  to  run  out  when  the 
product  is  “done.”  At  that  point  there  is  generally  another  “color”  of 
money  to  complete  the  operational  life  cycle.  Another  issue  here  is  how  to 
deal  with  the  significant  lag  in  the  budgeting  process  relative  to 
development.  About  three  years  are  necessary  to  get  a  budget  “wedge” 
established  for  a  project  big  and  complex  enough  to  excite  the  interest  of 
OMB  review.  That  planning  horizon  has  much  to  do  with  generating 
program  risk  because  of  the  uncertainty  about  the  development  profile: 
The  environment  may  change,  the  specification  may  change,  the  reality  of 
the  formal  cause  will  contend  with  the  efficient  concept.  But  these 
problems  are  inherent  in  an  interaction  between  engineering  and 
budgeting.  If  ideas  want  resources,  they  are  the  beggars. 

A  Prescription:  The  Scaled  CONOPS 

A  lot  of  evidence  can  be  given  on  the  problems  in  complex 
organizations  through  the  life  cycle.  To  cut  that  short,  a  jump  will  be  made 
to  a  simple  observation  and  prescription  for  the  process.  This  is  based  on 
the  scale  hierarchy  and  a  specific  peer  collaboration  process  that  will  be 
called  the  CONOPS. 

The  empirical  development  process  has  always  had  a  number  of 
self-similar  scales  as  alluded  to  above.  In  older  system  engineering 
processes,  levels  of  specifications  were  called  out.  Each  of  these 
progressively  represents  a  further  decomposition  of  modules  and  a  richer 
specification  of  them  and  their  interfaces.  But  fundamentally  the  same 
kind  of  scale-hierarchy  triad  was  occurring  around  each  scale,  now  also 
represented  as  a  time-series  of  increasing  specification  toward 
instantiation: 

•  A  contextual  “mission”  (general  specification)  is  received  by  all 
agents  at  a  scale  (a  collaborative).  The  agents  represent  a 
decomposition  of  functionality  also  passed  with  the  mission.  The 
decomposition  is  integrated  by  an  interface  specification  also 
maintained  at  the  upper  context  scale. 

•  The  agents  cluster  as  a  loosely  coupled  system.  There  may  be 
several  physical  agents  assigned  to  a  module.  The  inter-module 
interaction  is  less  intensive  but  can  reflect  back  to  the  context 
scale  to  change  protocols  and  specifications. 


Spring  2006 


38 


•  The  “design”  interaction  within  a  module  is  intensive.  It  results  in 
the  next  scale  of  specification.  When  this  is  passed  on,  it  appears 
as  context  to  a  next  scale  of  collaboration.  The  next  scale  is  the 
mechanism  for  progressing  toward  instantiation. 

While  this  looks  like  a  waterfall  description,  it  also  incorporates 
spirals,  especially  when  the  reverse  sequence  up  the  scales  is  stated,  and  as 
occurs  in  reality.  So,  a  sequence  down  the  scales  can  instantiate  one  phase 
of  prototype  as  well  as  “final  product.”  In  a  globally  evolutionary  system, 
the  finite  life  cycle  of  any  product  means  that  there  is  always  recursion 
back  to  the  “top”  scale,  although  often  that  is  with  different  physical 
agents,  in  another  thread  for  another  “project.” 

Dynamically  the  scale  hierarchy  implies  a  requirement  of  nested 
stability  between  scales.  This  means  that  a  context  scale  has  to  be  stable 
long  enough  to  complete  a  specification-to-mechanism  and  reverse  cycle. 
Stability  means  that  the  information  from  context  remains  sufficiently 
constant  over  the  cycle  not  to  alter  any  of  the  information  from  an  agent 
cluster  to  its  mechanism,  and,  in  addition,  to  allow  time  for  an  agent 
cluster  to  learn  from  feedback  from  its  relative  mechanism  scale  This  is 
the  condition  for  convergence  of  learning  from  the  development  cycle 
itself.  If  this  does  not  occur,  the  “vertical”  information  generated  in  the 
process  as  part  of  the  formal/efficient  dialectic  cannot  be  organized.  That 
is,  the  system  will  not  self-organize  toward  a  product.  This  criterion  is 
equivalent  to  “requirements  stability,”  something  that  is  often 
pathologically  violated.  However,  the  ability  to  adapt  the  requirements 
from  internal  information  is  allowed  and  this  is  often  what  is  meant  by  an 
evo-devo  process  in  the  product  life-cycle  context. 

Clearly,  the  internal  stability  criterion  can  be  overturned  by 
“external”  information.  A  new  market,  technical  or  threat  environment  can 
render  obsolete  any  development.  If  this  external  information  dominates, 
the  system  is  not  self-organizing.  However,  the  incorporation  of 
fluctuations  (in  Prigogine’s  sense)  is  part  of  the  adaptive  process.  This 
puts  constraints  on  the  absolute  cycle  times  of  the  scale  interactions. 

We  want  “fast”  development  because  we  want  a  return  on 
investment  of  the  product  relative  to  an  inevitably  changing  environment. 
We  can  enhance  this  quality  if  we  make  each  focal  scale  expeditious  in 
achieving  a  self-similar  cycle  of  receiving  context,  interacting,  and 
specifying.  A  principle  for  doing  this  is  to  convene  the  agents 
(stakeholders)  relevant  to  any  module  in  a  concurrent  process,  as  opposed 


Washington  Academy  of  Sciences 


39 


to  preventing  the  intensive  intra-module  interaction.  But  this  simple 
principle  scales;  in  order  for  any  relative  context  scale  to  adapt  the 
interfaces  (that  are  loosely  coupled  at  a  focal  scale  but  intensively  relative 
to  the  context  scale)  the  higher  manager  must  similarly  convene  and  make 
concurrent  the  lower-scaled  modules. 

Now,  it  happens  there  is  a  model  for  this  expedited  process  readily 
at  hand,  and  it  has  been  there  for  a  few  decades,  at  least.  It  is  the 
Operational  Concept  Description  (OCD),  often  called  a  concept  of 
operations  (CONOPS).  However,  a  CONOPS  can  also  be  construed 
narrowly  as  a  codification  of  what  an  operation  does,  and  the  OCD  proper 
is  strictly  a  document  as  product  early  in  a  development  thread  of  a 
system.  But  the  content  requirement  of  the  OCD  is  fundamental: 

1 .  Statement  of  the  functional  mission. 

2.  The  environmental  scenarios  that  stress  the  functions. 

3.  An  analysis  of  shortfalls  in  the  functionality. 

4.  Improvements  (and  alternatives)  to  make  the  functionality 
adequate. 

5.  An  impact  evaluation  of  the  prescribed  improvements. 

There  is  a  process  behind  producing  this  content,  and  a 
constituency.  Also,  strongly  implied  by  the  contents  and  process  is  the 
“architecture”  of  the  system,  a  notion  that  arrived  later  in  systems 
engineering  than  the  OCD,  but  that  is  really  the  system  description 
embedded  in  the  CONOPS.  The  architecture  picture  emerges  from  the 
interaction  of  the  constituents  of  the  CONOPS.  What  has  just  been  stated 
is  nothing  more  than  a  description  of: 

•  An  allocation  of  mission  from  context  scale. 

•  The  interaction  between  the  focal  scale  of  agents  within  the 
mission  context. 

•  The  focal  scale  making  a  further  specification  of  the  proposed 
system  to  a  next  scale  of  specification. 

•  The  emergence  of  an  architecture  as  the  integrative  concept,  back 
to  context  scale  (or  we  can  view  architecture  as  part  of  mission 
allocation  with  the  focal  scale  adding  a  level  of  specificity;  in  any 
case  it  performs  the  modular  integration  role). 


Spring  2006 


40 


The  agents  at  the  focal  scale  include  the  operational  domain 
experts,  who  properly  are  the  customers  for  the  system.  But  equally  the 
engineers  must  be  included.  And  we  can  specify  scenario  creators 
(simulators)  and  evaluators  as  additional  fields  of  expertise.  Proper 
facilitation  of  such  a  constituency,  with  modular  specification  being 
embodied  in  prior  and  offline  knowledge  of  the  participants,  achieves  the 
desired  interactive  and  concurrent  expedition  of  the  process,  at  any  scale. 
We  can  have  physical  agents  participate  at  multiple  scales  to  embody  the 
vertical  interfaces  between  scales. 

This  seems  simple  enough,  but  there  is  a  great  deal  of  resistance  to 
it.  Firstly,  the  standard  jargon  destroys  the  self-similarity  between  the 
scales  of  process  that  become  sequential  steps  of  process.  At  some  high 
level,  there  is  strategy,  policy,  possibly  mission  analysis.  In  fact  any 
discourse  by  agents  at  that  scale  will  include,  however  implicitly,  the  five 
related  contents  of  the  OCD.  The  mistake  would  be  to  generate  some 
unconsidered  high-level  specification  (policy)  and  leave  the  other  parts  to 
someone  else  sometime  later.  Then,  the  OCD/CONOPS  by  name  appears 
only  early  in  the  acquisition  process,  generally  long  after  the  content  at 
that  scale  has  been  generated,  as  part  of  formal  documentation.  It  is  pasted 
together  absent  the  proper  constituency,  perhaps  by  some  group  distant 
from  the  earlier  development  process,  absent  operational  stakeholders. 
This  is  where  the  problems  in  the  requirements  process  start,  and  it  is  quite 
contrary  to  the  intent  of  the  OCD  and  its  placement  in  the  development 
process.  Then  at  some  lower  scale  one  is  just  “designing”  some  modular 
system  component  and  the  process  again  becomes  implicit.  If  in  fact  the 
higher  scales  have  constrained  the  design  so  much  as  to  obviate  the  need 
for  a  systematic  look  at  the  modules  at  a  particular  scale,  the  process  has 
become  too  dominated  by  the  formal  constraints  too  soon,  unless  you  do 
indeed  believe  in  a  straight-through  waterfall  progression. 

In  addition  to  failing  to  recognize  the  necessary  self-similarity  of 
the  OCD  process  at  multiple  scales,  there  is  the  tendency  to  make  a  critical 
path  serialization  of  the  content  steps:  Someone  defines  the  operational 
functions  (that  may  or  may  not  be  part  of  a  separate  “architecture” 
process),  someone  comes  up  with  improvement  concepts  (researchers  and 
engineers),  someone  does  modeling  and  simulation  on  alternatives, 
someone  does  the  “investment  analysis”  of  impacts.  There  can  be,  and  is, 
a  great  deal  of  confusion  on  what  order  these  should  be  done  in  because 
there  is  a  great  deal  of  confusion  about  their  scaling.  The  tendency  is  to 
make  each  part  a  big  project  itself,  and  to  engulf  scales  that  should  have 


Washington  Academy  of  Sciences 


41 


been  specified  earlier,  or  should  be  specified  later.  And  so  there  is  often 
criticism  that  none  of  this  is  adaptive  to  either  external  or  internal 
information.  In  any  case  it  slows  and  bulks  the  process.  It  does  not  help 
that  the  acquisition  benchmark  reviews  are  established  as  yet  another 
chain  of  activities,  with  other  constituents. 

While  the  modem  life-cycle  development  processes  retain  the 
OCD  (more  or  less  as  it  has  been  for  at  least  20  years  now),  they  have  not 
fully  recognized  its  significance  and  how  it  fits  a  true  evo-devo  system. 
There  are  several  rather  vague  prescriptions  for  tailoring  the  prescribed 
steps  of  the  life  cycle  process  to  complex  systems.  But  a  scale  hierarchy  of 
self-similar  CONOPS  interactions  is  the  complex  system  of  agents 
appropriate  to  the  instantiation  of  not  just  one  product,  but  an  ongoing 
adaptive  set  of  such  products. 

The  challenge  is  to  see  how  the  ongoing,  scaled  organization 
(although  with  different  physical  agents  filling  ex  officio  logical  positions 
over  time)  relates  to  the  threads  of  finite  life-cycle  development.  So,  the 
tendency  is  to  customize  a  linear,  progressive  acquisition  thread  to  parallel 
the  maturity  of  the  product  thread.  That  by  itself  contradicts  another 
intuitive  approach  of  “portfolio  management.” 

The  portfolio  approach  is  just  a  modular  clustering  of  projects. 
This  may  or  may  not  be  used  properly  in  the  risk  management  sense  of 
having  an  ensemble  of  projects  to  accomplish  a  mission.  Often,  a  portfolio 
is  just  a  set  of  different  but  critically  interdependent  projects.  However, 
with  an  ensemble  of  projects  there  is  large-scale  management  of  the 
portfolio  and  smaller  scaled  management  of  its  constituents.  This  concept 
can  be  aligned  with  the  modular  decomposition  of  any  project.  It  is 
difficult  for  this  scaled  approach  to  supplant  the  dedicated  acquisition 
chain  if  the  projects  themselves  are  monolithically  discrete.  But  that  in 
itself  is  what  is  opposed  by  the  reaction  to  the  waterfall  approach  coupled 
with  “big  bang”  projects.  That  reaction  comes  directly  from  experience 
with  projects  such  as  the  A  AS,  although  it  must  be  noted  that  the 
predecessors  to  the  AAS  and  the  systems  approach  itself  are  famously 
successful  “big  bang”  projects  like  the  ICBM,  the  Polaris  submarine, 
SAGE  and  NAS  Stage  A.  An  alternative  hypothesis  is  that  the  AAS  was 
victim  to  formalized  process  that  interfered  with  the  natural  scale 
hierarchy,  tried  to  keep  constituents  at  arm’s  length,  and  made  the 
interactive  steps  excessively  sequential. 


42 


Today,  “open  systems”  are  the  norm.  This  concept  exactly  reflects 
partial  decomposability  and  the  scale  hierarchy.  But  given  this  concept  of 
an  evolutionary  system  with  life-cycle  modules,  there  is  no  reason  to  make 
projects  so  monolithic  that  they  require  a  separate  acquisition  chain  for 
each  thread  in  addition  to  the  ongoing  scale  hierarchy  of  modular 
interfacing  and  specification.  The  waterfall  model  can  come  back  as  the 
“rainfall”  model.  That  is,  because  of  the  dynamic  scaling  requirements, 
large-scale  specifications  are  general  but  enduring.  As  we  go  down  the 
scales  there  is  a  more  and  more  raindrop-like  “falling”  of  specifications 
that  hit  the  physical  ground  as  real  modular  instantiations.  Then,  the 
reality  of  the  whole  system  “evaporates  up”  again  as  aggregate 
information  to  emerge  as  new  general  concepts  that  precipitate. 

This  model  has  the  key  advantage  that  it  keeps  the  foci  of  agents 
properly  scaled.  For  instance,  executives  need  to  keep  a  strategic  view. 
When  we  entrain  them  in  acquisition  threads,  they  are  distracted. 
Conversely,  the  builders  are  often  left  with  inadequate  context  so  that 
failures  are  blamed  on  “poor  requirements”  traceable  all  the  way  up.  We 
can,  and  should,  modularize  real  instantiations  so  that  their  mechanism 
management  is  kept  small  scale.  This  approaches  biological  evo-devo: 
The  genome  launches  portfolios  of  phenotypical  trials  and  does  not  itself 
engage  in  the  real  trials  of  life  cycles.  But  for  purposeful  agent  systems, 
what  we  have  is  a  set  of  scales,  with  CONOPS  all  the  way  through  and  the 
reification  of  products  only  low  in  scale.  This  also  preserves  a  balanced 
profile  of  efficient  causality  pursuing  the  mission  (final  cause)  versus 
formal  cause  of  the  instantiations  made  material. 

Conclusion 

The  intent  here  is  to  use  basic  evo-devo  theory  to  address  the 
problems  of  complex  organizations  building  complex  systems.  In  complex 
systems,  development  is  not  opposed  to  evolution  but  is  its  mechanism. 
Life  cycle  instantiations  that  have  a  development  path  are  the  means  for 
the  evolutionary  system  to  adapt.  In  the  biological  application  of  this 
concept,  design  is  problematic.  In  organizations  of  purposeful  agents,  it  is 
inherent.  By  being  purposeful,  every  agent  designs,  if  only  to  decide  a 
message  or  an  action.  The  problem  is  that  we  engage  in  the  fallacy  of 
extrapolating  this  individual-agent  activity  to  the  collective  that  emerges 
with  scale-hierarchy  structure.  The  highest  ex  officio  position  in  such  a 
hierarchy  is  filled  by  a  designing  individual:  the  state,  or  the  corporation, 
or  the  organization  however  does  not  “design.”  This  fallacy  carries  over 
into  the  problem  of  how  the  process  for  life  cycle  development  and 


Washington  Academy  of  Sciences 


43 


acquisition  is  “designed.”  We  entrain  the  highest  scales  of  the 
organization  in  development  threads  as  if  there  were  corporate  design  of 
the  products.  But  there  is  not — there  is  only  the  scaled  functioning  of  the 
organization  that  must  continually  integrate  physical  instantiations  into  an 
indefinitely  large  system  that  was  not  by  any  means  designed.  Under  the 
“open  systems”  doctrine — that  is  synonymous  with  evo-devo — there  is  no 
need  to  scale-up  what  is  designed  until  it  approaches  the  scale  of  the 
emergent  organization. 


REFERENCES 


Allen,  P.M.  et.  al.,  The  Dynamics  of  Urban  Evolution,  Volume  1:  Inter-Urban  Evolution 
and  Volume  2:  Intra-Urban  Evolution ,  Final  Report,  October  1978.  USDOT-RSPA, 
Cambridge,  MA. 

_ ,  1981,  Urban  Evolution,  Self-Organisation  and  Decision  Making , 

Environment  and  Planning  A,  pp  167-183. 

Allen,  P.M.,  1997,  Cities  and  Regions  as  Self-Organizing  Systems:  Models  of 
Complexity,  Gordon  and  Breach,  Environmental  Problems  and  Social  Dynamics 
Series. 

Lendaris,  George  G.,  1964.  On  the  Definition  of  Self-Organizing  Systems,  Proc.  Of  the 
IEEE,  March  1964. 

Mankins,  John  C.,  1995.  TECHNOLOGY  READINESS  LEVELS,  A  White  Paper,  April 
6,  1995,  Advanced  Concepts  Office,  Office  of  Space  Access  and  Technology, 
NASA.  At  http://www.hq.nasa.gov/office/codeq/trl/trl.pdf 

Mesarovic,  M.D.,  1964.  Self-Organizing  Control  Systems,  pp.  265-269,  IEEE  Trans.  On 
Applications  and  Industry,  83. 

Nelson,  Gary  G.,  1990.  Hierarchies,  Noise  Like  the  Wind  and  Networks,  pp.  564-570, 
Vol.  II,  Proceedings  of  the  34th  Annual  Meeting,  International  Society  for  the 
Systems  Sciences,  Portland,  OR.  July  8-13,  1990. 

Nelson,  Gary  G.,  1992.  Adaptive,  Multi-Scaled  ATM:  Making  it  Work,  USDOT,  FAA, 
ARD-100. 

Nelson,  Gary  G.  and  Peter  M.  Allen,  Self-Organizing  Geography:  Scaled  Objects  and 
Regional  Planning  in  the  U.S.,  NECSI  Bi-Annual  Meeting,  2001,  Nashua,  N.H. 

Nelson,  Gary  G.,  2003.  “Next  TFM:  Collaboration  to  2015  and  Beyond”,  for  the  Federal 
Aviation  Administration  (FAA),  AUA  TAC. 

Nelson,  Gary  G.,  2005.  High-Level  Architecture  of  Homeland  Security,  Homeland 
Security  Institute. 

Nelson,  Gary  G.,  2006.  “Axioms  and  Architecture  for  Multi-Agent  Systems  in  Homeland 
Security”,  Submitted  to  the  3rd  International  ISCRAM  Conference,  May  2006. 

Norman,  Douglas  O.  and  Michael  L.  Kuras,  2004,  Engineering  Complex  Systems,  The 
MITRE  Corporation,  January  2004.  At 

http://www.mitre.org/work/tech_papers/tech_papers_04/norman_engineering/ 


Spring  2006 


44 


Salthe,  S.N.,  1985.  Evolving  Hierarchical  Systems:  Their  Structure  and  Representation, 
Columbia  University  Press. 

Salthe,  S.N.,  1993.  Development  and  Evolution:  Complexity  and  Change  in  Biology, 
MIT  Press. 

Salthe,  S.N.,  Summary  of  the  Principles  of  Hierarchy  Theory,  November  2001.  At 
http://www.nbi.dk/~natphil/salthe/hierarchy_th.html 
Simon,  H.A.  and  A.  Ando,  1961.  Aggregation  of  variables  in  dynamic  systems. 
Econometrica:  29:111-138,  1961. 


Washington  Academy  of  Sciences 


A  METHOD  FOR  DESIGNING  IMPROVEMENTS  IN 
ORGANIZATIONS,  PRODUCTS,  AND  SERVICES 


45 


Dragon  Tevdovski 
University  of  Sts.  Cyril  and  Methodius, 
Skopje,  Macedonia 

Irina  Naoumova 

Kazan  State  University, 

Kazan,  Russia 

Stuart  Umpleby 
The  George  Washington  University 
Washington,  D.C. 


Abstract 

A  Quality  Improvement  Priority  Matrix  (QIPM)  may  be  used  for 
identifying  priorities  for  improving  an  organization,  a  product,  or  a 
service.  This  paper  reports  on  the  use  of  the  QIPM  method  by  members 
of  the  Department  of  Management  Science  at  The  George  Washington 
University  and  members  of  the  Department  of  Management  at  Kazan 
State  University  in  Kazan,  Russia,  in  2002.  Features  of  a  Department, 
such  as  salaries,  teaching  assistants,  computer  hardware,  etc.  (a  total  of 
5 1  features),  were  evaluated  on  the  scales  of  importance  and 
performance.  Recent  research  has  significantly  improved  the  method 
as  a  way  of  determining  priorities,  monitoring  progress,  identifying 
consensus  or  disagreement,  and  comparing  two  organizations.  This 
paper  discusses  additional  statistical  improvements  and  ways  of 
presenting  the  results  of  statistical  analysis.  The  QIPM  method  is  a 
way  of  achieving  agreement  among  a  group  of  people  on  the  most 
important  actions  to  be  taken. 


Introduction 

The  features  of  an  organization,  as  evaluated  by  employees,  might 
include  salaries,  health  benefits,  office  space,  secretarial  help,  and 
computer  equipment.  The  features  of  a  product,  as  evaluated  by 
customers,  might  include  price,  styling,  reliability,  and  resale  value. 
Assuming  an  organization  wants  to  improve  its  performance,  where 


Spring  2006 


46 


should  it  focus  its  attention?  How  can  an  organization  use  limited 
resources  so  as  to  achieve  the  greatest  return  in  customer  and  employee 
satisfaction? 

A  Quality  Improvement  Priority  Matrix  (QIPM)  can  be  used  to 
determine  priorities  among  features  and  to  monitor  performance 
improvement.  Customers  or  employees  evaluate  various  features  of  an 
organization  or  product  or  service  on  two  scales:  importance  and 
performance.  The  intent  is  to  identify  features  that  are  rated  high  in 
importance  and  low  in  performance.  A  Quality  Improvement  Priority 
Matrix  was  first  described  by  the  managers  at  GTE  Directories 
Corporation  in  1995.  They  conducted  a  customer  satisfaction 
measurement  program  for  determining  what  was  important  to  their 
customers,  how  well  the  company  was  performing,  and  how  the  comp° 
could  do  better.  (Chapman,  1995  and  Carlson,  1995) 

A  similar  method  called  a  “strategic  improvement  matrix”  was  used  uy 
the  people  from  Armstrong  Building  Products  Operation.  (Wellendorf, 
1996)  A  QIPM  was  found  to  be  useful  for  evaluating  the  Junior  Faculty 
Development  Program  by  Naoumova  and  Umpleby  (2002).  Melnychcnko 
and  Umpleby  (2001)  and  Karapetyan  and  Umpleby  (2002)  used  a  QIPM 
to  identify  priorities  in  a  University  department.  Prytula,  et  al.  (2004) 
devised  the  Importance/  Performance  Ratio.  Dubina  and  Umpleby  (2006) 
applied  cluster  analysis  and  suggested  that  standard  deviation  be  used  as  a 
measure  of  lack  of  agreement. 

The  aim  of  this  paper  is  to  compare  the  assessments  by  faculty 
members  of  the  Department  of  Management  Science  at  the  George 
Washington  University  (GWU),  USA,  and  the  Department  of 
Management  at  Kazan  State  University  (KSU),  Russia,  and  to  further 
develop  the  QIPM  method  as  a  guide  for  improvement  efforts.  We  define 
a  high  priority  feature  as  having  high  importance  and  low  performance. 
Naoumova  and  Umpleby  (2004)  earlier  compared  priorities  of  thusa 
Departments,  but  in  their  analysis  they  used  simple  quantitative  methods 
In  this  paper  we  shall  improve  the  comparison  of  the  features  and  thd* 
priorities  by  using  more  advanced  statistical  techniques.  We  present  the 
data  in  Part  II.  Evaluation  and  standardization  of  the  measures  is  made  in 
Part  III.  Part  IV  presents  and  discusses  the  data  in  matrix  form  for  the  two 
departments.  Parts  V  and  VI  compare  the  priorities  of  the  two 
departments  and  show  the  results  of  a  cluster  analysis. 


Washington  Academy  of  Sciences 


1 


47 


The  Data 

Data  were  collected  by  means  of  a  questionnaire.  The 
questionnaire  covered  51  features.  The  features  included  in  the 
questionnaire  were  issues  that  had  been  discussed  by  the  GWU  faculty  in 
recent  years.  The  same  features  were  used  in  the  KSU  questionnaire  in 
order  to  make  comparisons.  The  questionnaire  was  given  to  management 
faculty  members  at  both  GWU  and  KSU  in  2002.  Twenty  responses  were 
received  from  GW  faculty,  and  eighteen  from  KSU  faculty.  Faculty 
members  evaluated  the  importance  and  performance  of  each  feature  of  the 
department.  A  scale  from  0  to  10  was  used.  On  the  importance  scale  0 
means  that  the  feature  has  no  importance  at  all  and  10  means  that  the 
feature  has  a  very  high  importance  for  the  department.  On  the 
performance  scale  0  means  that  the  department’s  performance  is  very 
poor,  whereas  10  means  the  department’s  performance  is  excellent. 

Evaluation  of  Importance  and  Performance 

The  scores  for  each  feature  were  averaged.  Descriptive  statistics 
for  GWU  and  KSU  are  shown  in  Table  1 . 

Table  1:  Descriptive  statistics  for  GWU  and  KSU 
Importance  -  Performance 


Std. 

N 

Range 

Minimum 

Maximum 

Mean 

Deviation 

Importance 

(GWU) 

51 

4.80 

4.20 

9.00 

7.5408 

1.25207 

Performance 

(GWU) 

51 

4.90 

3.25 

8.15 

5.4890 

1.18905 

Importance 

(KSU) 

51 

6.00 

4.00 

10.00 

7.3371 

1.84934 

Performance 

(KSU) 

51 

8.39 

.50 

8.89 

4.3529 

2.49989 

Spring  2006 


48 


For  all  the  features  the  mean  value  on  importance  at  GWU  was 
7.54.  At  KSU  the  mean  value  for  importance  was  7.34.  These  results 
imply  that  the  features  are  considered  to  be  quite  important  by  faculty 
members  at  both  universities.  The  mean  scores  on  performance  at  GWU 
and  KSU  are  5.49  and  4.35,  respectively.  These  scores  suggest  that 
corrective  actions  should  be  taken  in  order  to  improve  the  functioning  of 
both  university  departments. 

Dispersion  is  a  measure  of  consensus  among  the  faculty  members. 
A  standard  deviation  of  0  implies  that  faculty  members  evaluate  a  feature 
the  same  way.  The  higher  the  standard  deviation  is,  the  higher  are  the 
evaluation  differences  among  the  faculty  members.  GWU  standard 
deviations  are  1.25  and  1 .19  on  importance  and  performance,  respectively. 
KSU  has  much  higher  standard  deviations,  1.85  and  2.50  on  importance 
and  performance,  respectively.  In  order  to  compare  the  evaluation 
differences  we  measured  the  coefficient  of  variation.  It  ranges  between  0% 
and  100%.  If  the  coefficient  of  variation  is  0%,  this  means  that  there  is 
consensus  among  faculty  members.  If  its  value  is  100%,  this  means  that 
all  faculty  members  differ  in  their  evaluations.  The  coefficients  of 
variation  of  GWU  and  KSU  are  presented  in  Table  2. 


Table  2:  Coefficients  of  Variation  of  GWU  and  KSU 


Coefficient  of  Variation 

Importance  (GWU) 

16.60% 

Performance  (GWU) 

21.66% 

Importance  (KSU) 

25.21% 

Performance  (KSU) 

57.43% 

GWU  has  higher  agreement  among  faculty  members.  Or,  in 
other  words,  differences  among  KSU  faculty  members  are  higher.  Table  2 
also  suggests  that  the  differences  on  performance  are  higher  than  those  on 
importance.  Especially  note  the  high  values  of  dispersion  on  the  KSU 
performance  measures. 

In  order  to  equalize  the  level  of  consensus  among  faculty  members 
at  the  two  universities  we  standardized  the  importance  and  performance 
measures.  Every  feature  was  divided  by  the  respective  standard  deviation. 


Washington  Academy  of  Sciences 


49 


Standardized  importance  and  performance  measures  are  presented  in 
Table  3. 


Table  3:  Standardized  Importance  and  Performance 


N 

Range 

Minimum 

Maximum 

Mean 

Std. 

Deviation 

Importance 

Standardized 

(GWU) 

51 

3.84 

3.35 

7.19 

6.0225 

1.00 

Performance 

Standardized 

(GWU) 

51 

4.12 

2.73 

6.85 

4.6157 

1.00 

Importance 

Standardized 

(KSU) 

51 

3.25 

2.16 

5.41 

3.9661 

1.00 

Performance 

Standardized 

(KSU) 

51 

3.36 

0.20 

3.56 

1.7408 

1.00 

Note  that  standard  deviations  are  equal  to  one.  This  means 
that  the  evaluations  of  importance  and  performance  have  the  same  level  of 
consensus  among  members  of  the  GWU  and  KSU  departments.  GWU 
faculty  members  found  the  features  to  be  more  important  than  the 
members  of  the  KSU  department.  GWU  has  a  standardized  importance 
mean  1.51  times  higher  than  KSU.  The  comparison  of  performance  is 
even  more  significant.  GWU  has  a  standardized  performance  mean  2.65 
times  higher  than  KSU.  This  indicates  that  KSU  faculty  members  rate  the 
performance  of  their  department  lower  than  do  GWU  faculty  members.  A 
visual  comparison  of  GWU  and  KSU  standardized  importance  and 
performance  means  is  shown  in  Figure  1 . 

We  used  the  QIPM  as  a  tool  for  determining  the  priority  of  the 
features.  A  QIPM  consists  of  four  quadrants.  The  northeast  quadrant 
contains  features  with  high  importance  and  high  performance.  The 
features  in  this  quadrant  do  not  need  corrective  action.  The  features  in  the 
northwest  quadrant  have  low  importance  and  high  performance.  Resources 
of  the  department  should  be  transferred  from  the  features  of  this  quadrant 
to  features  with  high  importance  and  low  performance.  The  third  quadrant 


Spring  2006 


50 


is  southwest.  The  features  in  it  are  characterized  by  low  importance  and 
low  performance.  Using  department  resources  on  the  features  in  this 

quadrant  depends  on  their  importance.  The  last  quadrant  is  the  southeast 

Figure  1:  GWU  and  KSU  Standardized  Importance  and  Performance 

Means 


Standardized  Importance  Standardized  Performance 


Quality  improvement  Priority  Matrix 

quadrant.  These  features  have  high  importance  and  low  performance. 
These  features  have  the  highest  priority  for  the  department.  For  these 
features  corrective  action  is  necessary.  Hence,  we  focus  our  attention  on 
the  features  in  this  quadrant. 

Figure  2  shows  the  QIPMs  for  the  GWU  and  KSU  departments. 
The  data  are  non-standardized.  Fifteen  features  are  found  in  the  GWU 
southeast  quadrant  and  nineteen  in  the  KSU  southeast  quadrant.  These 
numbers  suggest  that  there  are  many  features  in  both  departments  that 
need  corrective  action. 

In  order  to  focus  our  attention  on  urgent  features  we  changed  the 
borders  of  the  quadrants.  See  Figure  3.  The  new  borders  are  average 
values  of  the  total  GWU  and  KSU  features.  The  joint  GWU  and  KSU 
importance  average  is  7.44,  and  joint  GWU  and  KSU  performance 
average  is  4.92.  Six  features  are  found  in  the  GWU  southeast  quadrant: 
office  security,  building  physical  environment,  conference  room  and  other 
space,  secretarial  support,  department  strategic  plan  and  computer 
laboratories.  Nine  features  are  found  in  the  KSU  southeast  quadrant:  travel 
support,  projection  equipment,  salaries,  classroom  facilities,  copiers, 
building  physical  environment,  accounts  payable,  computer  hardware  and 
teaching  assistants. 


Washington  Academy  of  Sciences 


Figure  2:  GWU  and  KSU  QIPM 


Importance  Standardized  (GWU) 


Importance  Standardized  (KSU) 


Figure  3:  GWU  and  KSU  QIPM  Based  on  Joint  Averages 


Importance  Standardized  (GWU) 


Importance  Standardized  (KSU) 


Figure  4  shows  the  matrices  for  the  GWU  and  KSU  departments 
based  on  the  standardized  values  of  importance  and  performance. 
Standardization  is  used  to  achieve  the  same  level  of  consensus  among  the 
members  of  both  departments  on  the  evaluation  of  importance  and 


Spring  2006 


52 


performance.  However,  this  approach  can  be  misleading.  If  the  importance 
and  performance  scales  were  reversed,  so  that  0  was  high  importance  or 
performance  and  10  was  low,  dividing  by  standard  deviation  would  raise 
rather  than  lower  importance  and  performance  scores.  To  see  the  impact 
of  standardization  compare  the  coordinates  of  Figures  3  and  4.  Note  that 
the  coordinates  of  the  features  are  shifted  from  their  original  positions. 
The  shift  in  coordinates  is  proportional  to  the  standard  deviations  of  the 
respective  importance  and  performance  scales.  A  higher  standard 
deviation  leads  to  a  larger  downward  shift  of  the  corresponding 
importance  or  performance  features.  In  Figure  4  not  a  single  feature  at 
KSU  has  a  performance  score  above  the  combined  performance  mean. 
Standardization  by  dividing  by  standard  deviation  may  be  more  useful 
when  comparing  a  large  number  of  universities.  In  this  case,  comparing 
just  two  very  different  universities,  the  unstandardized  means  may  be 
more  informative. 


Figure  4:  GWU  and  KSU  QIPM  Based  on  Standardization 


* 

•o 

S4£ 


c 
« 

£  2  .00  ■ 


- 

0 

\  <&>o 

1  0 

■o  0 

1  > 

b° 0 

O  0! 

Co  0  0  ] 

1  o\ 

O  O  0 

9o 

©  © 

H - 1 - 1 

i  %  ° 

i — - 1 - 

Importance  Standardized  (KSU) 


In  the  QIPM  of  the  GWU  department  most  of  the  features  have 
high  evaluations  on  the  importance  scale.  Only  2  from  all  5 1  features  have 
low  importance.  The  rest  of  the  features  are  in  the  quadrants  with  high 
importance.  Among  them  12  are  features  in  the  southeast  quadrant.  They 
have  high  importance  and  low  performance.  A  list  of  these  priority 
features  is  given  in  Table  4. 


Washington  Academy  of  Sciences 


53 


The  main  characteristic  of  the  KSU  department  is  the  low 
performance  ratings  of  the  features.  There  are  no  features  with  a  high 
standardized  performance  evaluation.  There  were  26  priority  features  in 
the  southeast  quadrant.  These  are  listed  in  Table  5. 

In  general,  the  priorities  differ  between  the  GWU  and  KSU 
departments.  Only  4  features  are  found  to  be  in  the  southeast  quadrants  for 
both  departments:  building  physical  environment,  accounts  payable, 
department  strategic  plan  and  department  organization  to  implement  its 
strategic  plan.  Note  that  the  number  of  priority  features  of  the  KSU 
department  is  more  than  double  the  number  for  the  GWU  department.  This 
result  can  be  explained  by  the  lower  average  performance  evaluations  by 
the  KSU  faculty  members  relative  to  the  GWU  faculty  members.  It  seems 
that  more  work  needs  to  be  done  to  improve  performance  at  KSU  than  at 
GWU. 


Table  4:  GWU  Features  in  the  Southeast  Quadrant 


GWU  Priority  Features 

Standardized 

Importance 

Standardized 

Performance 

Office  security 

7.15 

3.62 

Building/  physical  environment 

5.99 

3.36 

Dept,  organization  to  implement  its  strategic  plan 

5.67 

3.23 

Dept,  strategic  plan 

5.97 

3.45 

Help  with  writing  research  proposals 

4.71 

2.73 

Use  of  continuous  improvement  methods  in  the 
Department 

5.13 

3.01 

Conference  room  and  other  space 

5.91 

3.57 

Secretarial  support 

5.91 

3.70 

Accounts  payable 

5.5 

3.55 

Working  papers  series 

4.22 

2.92 

Course  evaluations 

4.47 

3.74 

Social  activities 

4.12 

3.69 

Spring  2006 


- 


Table  5:  KSU  Features  in  the  Southeast  Quadrant 


KSU  Priority  Features 

Standardized 

Standardized 

Importance 

Performance 

Funds  to  support  research 

4.84 

0.20 

Travel  support 

4.83 

0.20 

Office  space  for  faculty 

5.41 

0.44 

Projection  equipment 

4.69 

0.50 

Salaries 

5.33 

0.80 

Classroom  facilities 

5.14 

0.89 

Copiers 

5.03 

1.12 

Building/  physical  environment 

5.03 

1.24 

Accounts  payable 

4.60 

1.24 

Computer  hardware 

4.89 

1.50 

Consulting  opportunities  in  area 

4.98 

2.05 

Teaching  assistants 

4.08 

1.88 

Dept,  organization  to  implement  its  strategic  plan 

4.40 

2.12 

Computer  labs 

4.88 

2.37 

Computer  software 

4.88 

2.40 

General  ability  of  students 

4.92 

2.45 

Dept,  strategic  plan 

4.82 

2.61 

Transparency  promotion  process 

4.84 

2.88 

Opportunities  to  work  with  faculty  in  other  departments 

4.54 

2.80 

Library  journal  collection 

5.08 

3.27 

Library  book  collection 

4.81 

3.29 

Opportunities  to  meet  local  businessmen  and  managers 

4.54 

3.17 

Coordination  with  other  depts. 

4.70 

3.32 

Dept,  head  protects  faculty  from  administrative 
interference 

4.46 

3.20 

A  supportive  climate  in  the  dept. 

4.89 

3.56 

Opportunities  for  academic  work  with  Dept,  faculty 

4.87 

3.55 

Washington  Academy  of  Sciences 


55 


Ranking  the  Priorities 

These  results  define  the  priorities  in  the  two  departments,  as 
judged  by  their  respective  faculty  members.  But  department  resources  are 
limited.  In  order  to  highlight  the  features  where  corrective  action  is  most 
needed  we  ranked  the  priorities.  For  this  purpose  we  used  two  methods: 
an  index  method  and  cluster  analysis.  In  this  part  we  describe  the  index 
method. 

A  standardized  importance-performance  ratio  (SIP)  is  defined  as: 

SIP  =  — 

p, 

where  Is  is  standardized  importance  and  Ps  is  standardized  performance. 

The  higher  the  value  of  the  index  the  higher  the  priority  that  should  be 
given  to  that  feature. 

It  is  important  to  note  that  the  SIP  ratio  has  one  weakness.  It  gives 
the  same  value  to  features  on  the  same  linear  distance.  For  example,  a 
feature  with  standardized  importance  8  and  standardized  performance  4 
has  the  same  priority  as  a  feature  with  standardized  importance  4  and 
standardized  performance  2  (the  SIP  is  2  in  both  cases).  This  is  a 
significant  weakness,  because  one  might  easily  decide  that  only  the  first 
feature  has  priority.  In  order  to  avoid  this  problem  we  only  ranked  features 
in  the  southeast  quadrant. 

In  Table  6,  we  present  the  five  features  with  the  highest  priority  for 
the  GWU  faculty  members,  according  to  SIP.  The  KSU  department’s  top 
five  priority  features  according  to  SIP  are  presented  in  Table  7.  (The 
features  in  Tables  4  and  5  are  also  ranked  in  order  by  SIP.) 


Table  6:  Highest  Ranking  GWU  Priorities  According  to  SIP  Ratio 


Rank 

GWU  Priority  Features 

SIP 

1 

Office  security 

1.977 

2 

Building/  physical  environment 

1.781 

3 

Dept,  organization  to  implement  strategic  plan 

1.756 

4 

Dept,  strategic  plan 

1.729 

5 

Help  with  writing  research  proposals 

1.724 

Spring  2006 


56 


Table  7:  Highest  Ranking  KSU  Priorities  According  to  SIP  Ratio 


Rank 

KSU  Priority  Features 

SIP 

1 

Funds  to  support  research 

24.19 

7 

2 

Travel  support 

24.17 

0 

3 

Office  space  for  faculty 

23 

9 

4 

Projection  equipment 

9.387 

5 

Salaries 

6.631 

The  Tabic  6  and  7  rankings  show  two  main  differences  between 
the  GWU  and  KSU  departments.  First,  the  KSU  top  priorities  are  directly 
related  to  the  improvement  of  the  conditions  of  the  individual  faculty 
members.  The  GWU  top  priorities,  on  the  other  hand,  are  mainly 
concerned  with  improving  the  functioning  of  the  department.  Second,  SIP 
ratios  are  much  higher  in  KSU  than  in  GWU.  This  is  a  consequence  oi  the 
low  KSU  performance  scores. 


Clustering  the  Priorities 

We  used  cluster  analysis  in  order  to  sort  different  priorities  into 
clusters  so  that  the  dissimilarity  between  two  priorities  is  minimized  if 
they  belong  to  the  same  cluster  and  maximized  otherwise.  The  measure  of 
dissimilarity  is  Euclidean  distance.  This  is  the  geometric  distance  in  the 
two-dimensional  space,  in  this  case  the  distance  between  the  features  in 
the  space  importance  -  performance.  It  is  computed  as  follows: 

distance^,  p)  =  ^{i2  )2  +(p2  -  ptf 

Where  /,  and  /?,  are  importance  and  performance  of  the  first 
feature,  and  i2  and  p2  are  importance  and  performance  of  the  second 
feature.  In  this  method,  the  distance  between  two  clusters  is  calculated  as 


Washington  Academy  of  Sciences 


57 


the  weighted  average  distance  between  all  pairs  of  scores  in  the  two 
clusters. 

We  divided  the  GWU  department  features  into  five  clusters.  The 
cluster  analysis  is  presented  in  a  working  paper  by  the  same  authors 
available  at  www.gwu.edu/~umpleby/qipm.html.  The  mean  values  of 
each  cluster  are  shown  in  Table  8.  We  ranked  the  clusters  according  to 
their  SIP  ratios.  Cluster  1  should  have  the  top  priority  for  the  GWU 
management.  This  cluster  contains  only  one  feature:  office  security.  The 
next  cluster  in  priority  is  cluster  2.  It  contains  two  features:  help  with 
writing  research  proposals  and  use  of  continuous  improvement  methods  in 
the  department.  Cluster  3  contains  six  features:  building  physical 
environment,  department  organization  to  implement  its  strategic  plan, 
department  strategic  plan,  conference  room  and  other  space,  and 
secretarial  support.  Figure  5  visually  presents  the  GW  priority  clusters. 


Table  8:  GWU  Cluster  Centers 


Cluster 

1 

2 

3 

4 

5 

Importance 

Standardized 

7 

4 

5 

4 

4 

(GWU) 

.15 

.92 

.83 

.22 

.3 

Performance 

Standardized 

3 

2 

3 

2 

3 

(GWU) 

.62 

.87 

.48 

.92 

.72 

1 

1 

1 

1 

1 

SIP 

.97 

.71 

.67 

.44 

.15 

KSU  priorities  are  divided  into  seven  clusters.  The  cluster  analysis 
is  presented  in  the  working  paper  available  at 
www.gwu.edu/~umpleby/qipm.html.  The  mean  values  of  each  cluster  are 
shown  in  Table  9.  The  highest  priority  for  KSU  management  should  be 
cluster  1.  It  contains  three  features:  funds  to  support  research,  travel 
support  and  projection  equipment.  Cluster  2  contains  three  features: 
office  space  for  faculty,  salaries  and  classroom  facilities.  The  third  cluster 
by  priority  contains  four  features:  copiers,  building  physical  environment. 


Spring  2006 


58 


accounts  payable  and  computer  hardware.  The  KSU  clusters  are  presented 
in  the  Figure  6. 


Figure  5:  GWU  SE  Quadrant 


Table  9:  KSU  Cluster  Centers 


Cluster 

1 

2 

3 

4 

5 

6 

7 

Importance 

Standardized 

(KSU) 

4.79 

5.29 

4.89 

4.24 

4.90 

4.60 

4.87 

Performance 

Standardized 

(KSU) 

0.30 

0.71 

1.27 

2.00 

2.38 

3.01 

3.40 

SIP 

15.97 

7.45 

3.85 

2.12 

2.06 

1.53 

1.43 

Washington  Academy  of  Sciences 


Figure  6:  GWU  SE  Quadrant 


A  supportive  c  limrte  in  the  de 


Importance  Standardized  (KSU) 


Conclusion 

We  used  the  method  of  a  Quality  Improvement  Priority 
Matrix  combined  with  statistical  methods  in  order  to  determine  the 
priorities  of  the  Department  of  Management  Science  at  The  George 
Washington  University  and  the  Department  of  Management  at  Kazan 
State  University  and  to  learn  how  a  QIPM  can  be  used  to  compare  two 
organizations.  We  found  that  priorities  differ  between  the  GWU  and  KSU 
departments.  In  addition,  after  standardization  of  the  measures,  the 
number  of  priorities  (features  in  the  SE  quadrant)  of  the  KSU  department 
is  more  than  double  the  number  for  the  GWU  department.  This  is  a 


Spring  2006 


60 


consequence  of  the  lower  performance  ratings  given  by  the  KSU  faculty 
relative  to  the  GWU  faculty.  However,  the  features  used  in  the  study  were 
based  on  discussions  in  the  GWU  department.  If  the  list  of  features  had 
come  from  both  the  GWU  and  KSU  departments,  the  results  would  have 
been  somewhat  different. 

The  paper  also  experimented  with  standardization  by 
dividing  mean  importance  and  mean  performance  by  the  standard 
deviation  to  achieve  the  same  level  of  agreement  for  the  two  groups.  This 
procedure  seemed  to  bias  the  results.  So,  it  should  be  used  carefully. 
Furthermore,  we  experimented  with  clustering  the  features  in  the  southeast 
quadrant.  This  is  an  alternative  means  of  prioritization  to  using  the 
importance/  performance  ratio.  The  ratio  may  be  a  simpler  guide  to 
action.  As  this  and  previous  papers  describing  experiments  with  the 
QIPM  method  show,  the  QIPM  is  a  conceptually  simple  but  surprisingly 
informative  means  of  prioritizing  actions  and  tracking  results. 


REFERENCES 

Carlson,  M.  (1995),  “GTE  Directories:  Customer  Focus  and  Satisfaction,”  The  Quest  for 
Excellence  VII ,  Official  Conference  of  the  Malcolm  Baldrige  National  Quality 
Award,  February  6-8,  1995,  Washington,  DC. 

Chapman,  C.R.  (1995),  “Conference  Report:  Quest  for  Excellence  VII.”  Journal  of 
Research  of  the  National  Institute  of  Standards  and  Technology ,  Volume  100, 
Number  3,  pp.  287-299. 

Dubina,  I.,  S.  Umpleby  (2006),  “Agenda  Setting  and  Improvement  Monitoring  in  a 
University  Department,”  Twelfth  Annual  Deming  Research  Seminar,  New  York 
City. 

Karapetyan,  A.,  S.  Umpleby  (2002),  “How  a  Quality  Improvement  Priority  Matrix 
Reveals  Change  in  a  University  Department,”  Russell  J.  Meyer  and  David 
Keplinger  (eds.),  Perspectives  in  Higher  Education  Reform ,  Volume  12, 
Alliance  of  Universities  for  Democracy,  Texas  Review  Press,  pp.  315-322. 

Melnychenko,  O.,  S.  Umpleby  (2001),  “Using  a  Quality  Improvement  Priority  Matrix  in 
a  University  Department,”  Customer  Satisfaction  Management  Frontier  -  VI, 
Johnson  A.  Edosomwon  (eds.),  Fairfax,  VA:  Quality  University  Press,  pp.  6.1- 
6.12. 

Naoumova,  I.,  S.  Umpleby  (2002),  “Two  Methods  Useful  for  Starting  a  Quality 
Improvement  Program,”  in  Russell  J.  Meyer  and  David  Keplinger  (eds.), 


Washington  Academy  of  Sciences 


61 


Perspectives  in  Higher  Education  Reform ,  Volume  1 1,  Alliance  of  Universities 
for  Democracy,  Texas  Review  Press,  pp.  185-193. 

Naoumova,  I.,  S.  Umpleby  (2004),  “A  Comparison  of  Priorities  in  an  American 
Academic  Department  and  a  Russian  Academic  Department”,  Proceeding  of  the 
Annual  Meeting  of  the  Alliance  of  Universities  for  Democracy,  Vilnus, 
Lithuania. 

Wellendorf,  J.A.  ( 1 996),  “Armstrong  Building  Products  Operations: 
Information  and  Analysis,”  The  Quest  for  Excellence  VIII ,  Official 
Conference  of  the  Malcolm  Baldrige  National  Quality  Award, 
February  4-7,  1996. 


Spring  2006 


AFFI1ATED  INSTITUTES 

The  National  Institute  for  Standards  and  Technology 
Meadowlark  Botanical  Gardens 
The  John  W.  Kluge  Center  of  the  Library  of  Congress 
Potomac  Overlook  Regional  Park 


Washington  Academy  of  Sciences 


63 


THE  PHILOSOPHICAL  SOCIETY  OF  WASHINGTON 

SELECTED  MINUTES 
Ronald  O.  Hietala,  Recording  Secretary 

The  Phibdhical  Soiet©  Washingts  is  the  area’s  oldest 
scientific  society,  founded  in  1871 .  The  society  meets  twice  a  month  from 
October  through  May  in  the  Powell  Auditorium  of  the  Cosmos  Club.  The 
meetings,  on  Friday  evenings  at  8:30,  are  free  and  the  public  is  cordially 
invited  to  attend,  and  to  remain  after  the  meetings  for  refreshments  and 
conversation. 

Each  meeting  of  the  Philosophical  Society  features  a  lecture  by  a 
distinguished  scientist.  By  tradition,  these  lectures  are  summarized  by  the 
Recording  Secretary  in  the  Minutes  of  the  meeting  and  are  read  aloud  at 
the  next  meeting.  Selected  Minutes,  chosen  for  broad  interest  in  the 
subject  matter  of  the  lecture,  are  printed  below  with  the  kind  permission  of 
the  Directors  of  the  Society,  the  speakers,  and  the  Recording  Secretary, 

Mr.  Ronald  Hietala. 

For  further  information,  go  to  www.philsoc.org. 

Minutes  of  the  2189th  Meeting 
March  77,  2005 

Lecture:  Genetics  Testing  and  Personalized  Medicine: 

Ms  Joann  Boughman 

President  Robert  Hershey  called  the  2189th  meeting  to  order  at  8:18  p.m. 
on  March  11,  2005.  The  minutes  of  the  2188th  meeting  were  read  and 
approved. 

Mr.  Hershey  introduced  the  main  speaker  of  the  evening,  Ms.  Joann 
Boughman  of  the  American  Society  for  Human  Genetics.  Ms.  Boughman 
spoke  on  “Genetic  Testing  and  Personalized  Medicine,  The  Genome  and  a 
Health  Care  Revolution.” 

Ms.  Boughman  reviewed  the  recent  history  of  human  genetics.  In  1990, 
the  Human  Genome  Project  was  launched,  NIH  started  the  Ethical,  Legal, 
and  Social  Implications  Program,  and  the  first  gene  for  breast  cancer  was 
mapped.  Since  then,  developments  have  been  rapid  and  many.  They 
include  sequencing  of  the  first  bacterial  gene  in  1995,  sequencing  of  a 


Spring  2006 


64 


mouse  gene  and  mapping  of  the  human  genome  in  1996.  Sequencing  of 
the  human  genome  was  begun  in  1999  and  draft  sequences  appeared  in 
2000  and  were  published  in  2001.  ahead  of  schedule  and  under  budget. 
The  finished  version  appeared  in  2003. 

Despite  all  that  progress,  we  are  only  entering  the  genome  era.  Ms. 
Boughman  believes. 

We  are  now  beginning  to  understand  that  a  gene  does  not  cause  a 
disease.  It  is  only  one  of  many  factors,  some  of  them  environmental. 
Making  use  of  genetic  disease  information  involves  identifying  the 
controlling  gene,  understanding  the  basic  defect,  and  developing 
diagnostics,  preventive  measures,  drug  therapies,  and  genetic  therapies. 

The  most  important  genetic  test,  she  said,  is  family  history.  She  urges 
her  audiences  to  collect  this  history’  soon;  she  has  seen  many  cases  where 
passage  of  time  has  made  it  very  difficult  to  collect  this  information.  She 
recommended  two  web  sites  that  can  be  helpful  with  this:  www.ashg.org 
and  www.hhs.gov  familvhistorv.  and  she  recommended  the  Surgeon 
Generafs  family  history  tool.  Beyond  that,  current  genetic  tests  include 
diagnostic  tests,  newborn  screening,  carrier  testing,  prenatal  testing,  and 
predictive  testing. 

Diagnostic  testing  is  often  used  to  confirm  or  rule  out  a  diagnosis.  An 
example  was  a  41 -year-old  male  in  an  emergency  room  with  chest  pain. 
When  the  doctor  learns  the  man’s  father  and  paternal  uncle  had 
myocardial  infarctions  at  ages  40  and  44.  the  diagnosis  and  treatment 
proceed  quickly.  This  shows  how  powerful  genetic  information  can  be. 

Infant  screening  is  mandated  for  a  number  of  diseases  and  this  leads  to 
much  more  effective  follow-up  and  treatment,  as  does  carrier  screening. 
Prenatal  screening  is  often  used  to  assess  the  health  of  a  fetus,  especially 
when  there  is  a  known  genetic  risk.  Predictive  testing  indicates 
presymptomatic  and  predispositional  conditions;  presymptomatic  meaning 
the  disease  will  develop  if  the  relevant  mutation  exists  and 
predispositional  meaning  the  development  of  symptoms  is  likely  but  not 
certain.  On  the  matter  of  predictions,  she  quoted  Yogi  Berra  - 
“Predictions  are  tricky. ...  especially  ones  about  the  future.” 

Information  about  the  future  is  one  distinct  feature  of  genetic  tests. 
They  are  also  exceptional  in  how  they  affect  family  members  and  in  their 
complex  and  probabilistic  nature. 

She  sketched  two  hypothetical  scenarios  of  a  woman  and  her  use  of 
genetic  information.  In  one,  the  woman  used  the  Surgeon  General’s  family 
history  tool  early  and  as  a  result  had  a  complete  gene  sequence 
determined.  Following  a  preventive  diet  and  exercise  regime  and  taking 


Washington  Academy  of  Sciences 


65 


needed  treatments  promptly,  she  lived  a  long  and  relatively  healthy  life.  In 
the  other  scenario,  she  never  heard  of  the  Surgeon  General’s  family 
history  tool.  She  declined  gene  sequencing  because  her  brother  had  lost 
his  health  insurance  because  of  genetic  information.  She  ate  an  unhealthy 
diet,  gained  weight  and  developed  high  blood  pressure.  She  began  a  drug 
treatment  for  the  hypertension  but  developed  a  hypersensitivity  reaction 
and  stopped  taking  it.  At  50,  she  developed  pain  in  her  left  arm.  Her  M.D., 
unaware  of  her  risk,  diagnosed  it  as  musculoskeletal  and  prescribed  rest. 
The  next  day  she  was  back  in  the  ER  in  cardiogenic  shock.  Lack  of 
genetic  information  prevented  quick  choice  of  optimal  treatment.  She  died 
in  the  emergency  room. 

Anticipating  some  of  the  possibilities  suggested  by  these  two  scenarios, 
37  states  have  made  laws  prohibiting  discrimination  in  health  insurance  or 
employment  based  on  genetic  information.  The  U.S. .  Senate  passed  a 
similar  bill  by  98  -  0.  A  bill  has  been  introduced  in  the  House. 

Ms.  Boughman  made  some  predictions  for  the  future  of  genomics  in 
medicine.  She  believes  that  primary  care  providers  will  practice  genetic 
medicine,  that  cancer  therapies  will  be  targeted  to  the  specific  types  of 
tumors,  that  individualized  pharmacogenomic  treatments  will  become 
common,  and  that  there  will  be  interventions  available  that  will  use 
targeted  genetic  switches.  She  pointed  out,  however,  that  the  effectiveness 
of  genetic  treatment  will  depend  on  accurate  transmission  of  complex  and 
predictive  accurate  information  and  on  understanding  how  to  act  on  that 
information. 

She  briefly  discussed  stem  cells.  There  are  three  types  -  adult,  cord 
blood,  and  embryonic  stem  cells.  Embryonic  cells,  which  appear  to  have 
the  most  promise,  are  grown  by  taking  the  inner  cell  mass  out  of  a 
blastocyst  and  putting  those  cells  in  a  culture  to  reproduce.  There  are  more 
than  400,000  blastocysts  in  freezers  that  are  not  going  to  be  used 
otherwise.  However,  federal  funds  cannot  be  used  on  stem  cells  outside 
the  permitted  lines.  Another  possibility  is  to  remove  the  nucleus  from  an 
egg  and  replace  the  nucleus  with  a  somatic  cell  to  stimulate  division  to 
produce  stem  cells  that  will  reproduce  in  a  culture. 

Ms.  Boughman  offered  to  answer  questions. 

A  hardy  Finn  from  the  Gulf  of  Bothnia,  having  survived  knife  fights 
and  wolf  attacks,  remains  concerned  about  prostate  cancer.  Is  there  a  test 
for  it,  he  asked?  Not  a  specific  one,  Ms.  Boughman  said. 

What  are  the  non-genetic  diseases?  Too  numerous  to  count.  Viral  and 
bacterial  infections,  certainly,  and  AIDS.  She  mentioned,  however,  that 
some  people  seem  to  be  naturally  resistant  to  AIDS,  which  is  a  fascinating 


Spring  2006 


66 


clue  to  how  it  might  be  treated. 

One  person  asked  if  it  is  true  that  there  are  about  100,000  genes  and 
that  thousands  of  them  do  nothing.  Although  it  once  was  thought  there 
were  80,000  to  100,000  genes,  it  now  appears  tb~re  are  about  25,000, 
which  is  interesting,  because  some  other  species  have  far  more.  The  so- 
called  unused  genes  are  important  as  place-holders  and  it  appears  they  are 
somehow  otherwise  very  important.  They  are  like  junk,  not  like  garbage. 
Garbage  is  thrown  out,  junk  is  put  in  the  attic.  The  chimpanzee  and  human 
are  98  percent  alike  in  their  genes.  Perhaps  the  spacing  will  explain  some 
of  the  differences. 

Mr.  Hershey  announced  the  next  meeting  and  made  the  parking 
announcement.  He  invited  visitors  to  join  the  Society,  and  then  adjourned 
the  2189th  meeting  at  9:35  to  the  social  hour. 

Attendance:  33 

Weather:  Misty,  occasional  sprinkles 

Temperature:  5  C. 

Respectfully  submitted, 

Ronald  O.  Hietala,  Recording  Secretary 
©  2005  Ronald  O.  Hietala 

Minutes  of  the  21 91st  Meeting 
April  8,  2005 

Lecture:  Nuclear  Magnetic  Resonance-based 
Quantum  Computing:  Ms  Karen  Sauer 

President  Robert  Hershey  called  the  2191st  meeting  of  the  Philosophical 
Society  of  Washington  to  order  in  the  Powell  Auditorium  of  the  Cosmos 
Club  at  8:20  p.m.  on  April  8,  2005.  The  minutes  of  the  2190th  meeting 
were  read  by  William  Saalbach,  acting  recording  secretary,  and  approved. 

This  was  the  occasion  of  the  David  Franklin  Bleil  Memoria*  ‘  .*  cture 
in  Physics,  sponsored  by  David  Frederick  Bleil. 

Mr.  Hershey  introduced  the  speaker,  Ms.  Karen  Sauer  of  George 
Mason  University.  Ms.  Sauer  spoke  on  “Nuclear  Magnetic  Resonance- 
based  Quantum  Computing.” 

Why  study  quantum  computing?  Ms.  Sauer  reminded  us  of  Moore’s 
law,  which  states  that  the  number  of  transistors  on  a  chip  doubles  every  1 8 
months.  The  inside  story  says  that  chip  development  is  not  the  controlling 
factor;  it  is  that  18  months  must  pass  before  introducing  a  new  product 


Washington  Academy  of  Sciences 


67 


which  will  infuriate  the  existing  customers.  No  matter  which  is  closer  to 
the  truth,  current  computer  architecture  does  limit  the  effectiveness  of 
computing. 

There  would  seem  to  be  two  major  advantages  of  a  quantum 
computer,  large  searches  and  factoring  large  numbers.  Using  quantum 
computing,  the  effort  of  finding  a  needle  in  a  haystack  of  N  elements 
increases  with  the  square  root  of  N  instead  of  N.  Finding  a  certain  word 
for  a  crossword  puzzle  that  would  take  a  conventional  computer  500,000 
steps  would  take  a  quantum  computer  1 000  steps. 

Factoring  would  also  be  much  faster.  Using  Shor’s  factoring 
algorithm,  factoring  would  be  exponentially  faster  than  the  best  known 
classical  algorithm,  which  would  have  major  implications  for 
cryptosystems. 

The  building  block  of  quantum  computing  is  called  the  qubit  instead 
of  a  bit.  A  bit  can  represent  any  intersection  of  two  lines;  a  qubit  can 
represent  any  point  of  a  solid.  It  can  be  a  combination  of  a  1 -state  and  a  0- 
state,  or  a  number  of  them.  N  qubits  can  stand  for  2N  power  at  once.  A 
mere  50  qubits  can  represent  every  binary  number  from  zero  to  more  than 
a  trillion,  simultaneously.  The  readout,  however,  would  be  very  long. 

Qubits  are  also  unique  by  their  entangled  state.  In  classical  computing 
architecture,  two  bits  can  be  00,  01,  10,  or  11.  The  value  of  one  bit  does 
not  affect  another.  In  quantum  computing,  it  does,  because  of  the  way  the 
states  of  quantum  factors  affect  others. 

Ms.  Sauer  reviewed  the  experimentation  to  date.  Nuclear  magnetic 
resonance  computation  has  been  accomplished  with  seven  cubits.  An  ion- 
trap  system  has  been  developed  to  control  and  read  three  qubits.  Using 
superconducting  electronics,  computation  has  been  accomplished  with  two 
qubits.  A  quantum  dot  system,  using  the  spin  of  the  electrons,  has  been 
used  in  computing  with  one  cubit. 

Almost  any  interaction  with  the  environment  collapses  the  quantum 
state  into  a  very  definite  state,  or  decoherence,  and  calculation  stops. 
Therefore,  the  computer  must  be  isolated  from  the  environment,  which 
makes  it  difficult  to  control  and  read  the  quantum  states. 

A  magnetic  nucleus  in  a  magnetic  field  behaves  like  a  gyroscope  and 
precesses  about  the  field.  The  magnetic  moment  precesses  in  proportion  to 
the  field,  and  this  has  led  to  great  concern  and  competition  in  the  size  of 
the  superconducting  magnets  used.  They  are  now  working  with  magnets 
of  1 1 .7  tesla. 

Fundamentally,  the  way  an  NMR  (Nuclear  Magnetic  Resonance) 
computer  works  is  this:  a  magnetic  field  is  used  to  align  the  magnetic 


Spring  2006 


68 


moments.  Radio-frequency  pulses  at  the  Larmor  frequency  tip  the 
magnetic  moments  out  of  balance.  The  transverse  then  rotates  at  the 
Larmor  frequency  and  is  detected  by  a  coil  using  Faraday’s  law.  The  coil 
is  part  of  a  “tuned”  circuit  that  is  sensitive  to  a  limited  band  of 
frequencies.  The  ability  of  the  pulses  to  reverse  spins  has  given  NMR 
computing  the  boost  to  seven  qubits. 

Isolating  the  computers  is  done  by  dissolving  the  molecules  in  a 
liquid.  The  effect  of  the  magnetic  field  of  one  nucleus  on  another  is 
largely  averaged  away  in  a  liquid. 

The  major  problems  with  NMR  Quant  computing  are  that: 

•  It  is  best  done  on  single  molecule,  but  NMR  needs  108  molecules 
to  see  the  signal.  The  solution  is  to  use  a  large  number  of 
molecules  and  read  out  a  collective  answer. 

•  NMR  is  too  weak  to  determine  the  outcome  and  cause  the  state’s 
collapse  into  specific  states  for  each  molecule.  However,  it  is  often 
good  enough  to  see  an  NMR  signal  that  represents  the  average 
over  all  the  molecules. 

•  The  equilibrium  states  of  the  molecules’  nuclear  spins  are  nearly 
random,  with  only  a  relative  few  pointing  in  the  right  direction. 
The  solution  is  to  use  temporal,  spatial,  or  logical  labeling  methods 
to  single  out  the  small  fraction  that  do  represent  the  desired  initial 
state. 

These  methods  produce  “pseudo-pure”  states,  where  the  readout 
shows  what  is  designed  even  though  most  of  the  mix  does  not. 

Ms.  Sauer  predicts  that  the  field  will  turn  to  a  solid  instead  of  a  liquid 
form  for  the  chemicals.  Liquid  state  NMR  computers  will  likely  be  limited 
to  about  10  to  20  qubits;  50  to  300  are  needed.  She  believes  quantum 
computing  has  enormous  potential,  particularly  for  large-scale  searches 
and  factorization  of  large  numbers.  She  recognizes  that  actual  quantum 
computing  is  far  behind  current  theory,  but  she  says  NMR  has  provided  a 
good  test  bed  for  quantum  computing,  with  its  precise  control  of  magnetic 
moments.  Other  quantum  computers  with  better  scalability  should  benefit 
from  the  ideas,  concepts  and  solutions  that  NMR  experiments  have 
yielded. 

Ms.  Sauer  offered  to  answer  questions. 

One  questioner  observed  that  the  systems  she  described  do  not  seem 
fast.  They  won’t  do  traditional  calculations  well,  she  said,  but  that’s  not  a 
fair  comparison.  They  will  do  different  calculations. 

How  do  you  input  a  number?  someone  asked.  We  do  not  know  how  to 
put  a  number  in  a  molecule.  In  solids,  this  may  not  be  the  problem  it 


Washington  Academy  of  Sciences 


69 


seems  to  be  now.  Why  can’t  the  same  problems  be  put  on  a  classical 
computer?  another  asked.  Because  a  quantum  mechanical  state  can  mean  a 
numericity  of  states. 

Ms.  Sauer  admitted  there  is  no  advantage  to  quantum  computing 
currently.  There  is  no  problem  the  current  machinery  can’t  do  better.  It  is 
the  potential  that  is  great. 

Mr.  Hershey  encouraged  visitors  to  join  the  Society.  He  announced  the 
next  meeting  and  made  the  parking  announcement.  He  invited  everyone  to 
stay  for  the  social  hour.  Finally,  he  adjourned  the  2191st  meeting  at  9:50 
p.m. 

Attendance:  50 

Weather:  Moist  and  mild 

Temperature:  17  C. 

Respectfully  submitted, 

Ronald  O.  Hietala,  Recording  Secretary 
©  2005  Ronald  O.  Hietala 

Minutes  of  the  2192nd  Meeting 
April  22,  1005 

Lecture:  Smallpox  and  Ebola  Viruses  as  Agents 
Of  Bioterrorism:  Mr.  Peter  Jahrling 

President  Robert  Hershey  called  the  2192nd  meeting  of  the  Philosophical 
Society  of  Washington  to  order  in  the  Powell  Auditorium  of  the  Cosmos 
Club  at  8:16  p.m.  April  22,  2005.  The  minutes  of  the  2191st  meeting  were 
read  and  approved. 

Mr.  Hershey  introduced  the  speaker  of  the  evening,  Mr.  Peter 
Jahrling.  Mr.  Jahrling  is  chief  scientist  of  the  National  Institute  of  Allergy 
and  Infectious  Diseases,  part  of  the  National  Institutes  of  Health.  Mr. 
Jahrling  spoke  on  “Smallpox  and  Ebola  Viruses  as  Agents  of 
Bioterrorism.” 

Emerging  infections  arise,  Mr.  Jahrling  said,  by  unnatural  means. 
New  organisms  have  always  arisen  or  evolved  as  a  result  of  adaptation  or 
environmental  pressures.  New  pressures  come  from  the  extension  of  civil 
engineering  into  new  geography,  international  travel,  political  instability, 
natural  disasters  (often  involving  sanitation),  war  and  famine  and 
displaced  persons  that  result  from  them,  and  intentional  release.  Only  the 
last  is  the  action  of  terrorists.  The  reality  of  bioterrorism  did  not  sink  in 


Spring  2006 


70 


until  anthrax  attacks  of  2001. 

In  the  past  10-15  years,  new  viruses  that  have  broken  out  include 
the 

-  Hantavirus  Outbreaks  in  the  U.S. 

-  Ebola-related  Reston  Filovirus 

-  Andes  Virus  -  Argentina 

-  Ebola  -  Ivory  Coast,  Zaire,  C^bon,  Uganda,  Sudan 

-  SARS  -  China,  Canada,  (wo  wide?) 

-  Monkeypox,  almost  indistinguishable  from  smallpox 

-  Marburg  Virus  (Angola)  -  presently  ongoing 

-  Influenza  H5N 1  (poised  to  emerge) 

Mr.  Jahrling  said  he  had  never  intended  to  work  with  ebola.  There 
is  no  treatment  for  it,  it  is  fatal,  and  results  in  a  gruesome  death. 

He  gave  several  examples  of  how  tricky  and  dangerous  it  is  to 
work  with  these  viruses.  One  of  them  was  the  dying  golden  tamarins  at  the 
National  Zoo.  It  turned  out  to  be  a  virus  which  was  being  transmitted  to 
the  tamarins  by  the  baby  mice  the  caretaker  was  feeding  them.  The 
tamarins  were  being  bred  and  some  of  the  infected  ones  had  been 
scheduled  to  be  shipped  to  Brazil.  It  was  a  narrow  escape  that  they  were 
not. 

Monkeypox  came  to  the  United  States  in  a  shipment  of  Giant 
Gambian  rats.  Here  they  were  cohoused  with  prairie  dogs,  and  the  prairie 
dogs  were  sold  as  pets  in  the  Midwest.  Humans  were  infected  by  prairie 
dog  bites.  A  disturbing  fact  was  that  it  was  extant  in  the  country  for  about 
30  days,  even  though  the  symptoms  look  very  much  like  smallpox,  before 
it  was  reported  to  appropriate  authorities.  It  infected  30  people  in  15 
states.  It  was  just  fortunate  that  it  happened  to  be  a  weak  strain  of  the 
monkeypox  virus. 

Smallpox  was  declared  dead  in  1980.  The  campaign  to  defeat  it 
was  successful  because  it  has  no  natural  reservoir;  it  has  only  one  natural 
host,  humans.  Its  last  vestiges  were  wiped  out  in  Africa  by  identifying 
cases  and  then  sending  in  teams  to  vaccinate  everyone  the  infected  person 
came  in  contact  with. 

At  that  time,  however,  the  Soviet  Union  was  manufacturing  it.  They 
were  making  20  tons  of  smallpox  every  six  months.  They  were  planning  to 
put  it  in  the  nose  cones  of  missiles  to  wipe  out  any  survivors  of  a  nuclear 
war.  That  was  Russian  military  doctrine.  They  say  they  don’t  do  it  any 
more,  and  Mr.  Jahrling  believes  that,  but  we  can  only  wonder  if  all  that 
material  has  been  effectively  destroyed,  although  the  evidence  he  has  seen 
has  not  been  very  alarming.  They  also  produced  anthrax,  and  an  accidental 


Washington  Academy  of  Sciences 


71 


release  of  that  killed  80  people  downstream. 

There  was  an  outbreak  of  smallpox  in  Kazakhstan  in  1971.  It 
originated  in  the  Russian  smallpox  factory  on  an  island  in  the  Aral  Sea. 
The  woman  who  contracted  it  had  been  vaccinated,  but  for  some  reason 
the  vaccination  was  ineffective. 

He  described  a  planning  exercise  called  “Dark  Winter.”  It  assumes 
3000  people  arc  exposed  to  smallpox  in  simultaneous  attacks  on  three 
shopping  centers  in  different  states.  Since  42%  of  the  population  have 
never  been  vaccinated,  and  assuming  that  each  case  exposes  ten  others  and 
that  only  15  million  doses  of  vaccine  are  available,  by  the  fourth 
generation  of  the  disease,  in  nine  weeks,  three  million  people  would  have 
caught  smallpox  and  one  million  would  have  died.  The  planning  exercise 
demonstrated  two  important  points:  the  lack  of  vaccines  limits 
management  options  in  dealing  with  diseases,  and  the  United  States  lacks 
the  resources  to  deal  with  a  mass  outbreak  of  smallpox,  or  indeed  any 
contagious  bioterrorism  agent.  As  an  interim  policy,  the  best  they  could  do 
is  targeted  vaccination  of  contacts.  Asymptomatic  contacts  would  be 
monitored  but  not  isolated.  Any  delay  in  vaccination  would  increase 
mortality  greatly. 

Smallpox,  he  concludes,  is  a  very  credible  agent  of  terrorism.  Ring 
vaccination  would  be  only  a  partial  solution  to  a  large  or  multi-centered 
attack.  Mass  vaccination,  it  appears,  works  better  than  targeted 
vaccination,  largely  because  of  the  time  required  to  find  the  people  at  risk. 

There  is  surprisingly  little  material  around  to  show  what  smallpox 
looks  like.  They  do  have  a  preserved  human  arm  and,  of  course,  there  are 
old  pictures. 

They  are  working  to  develop  an  animal  model  of  the  disease. 
People  in  moon  suits  have  infected  monkeys  with  variola  in  labs  of  the 
Centers  for  Disease  Control.  Another  model  under  development  is 
infection  of  monkeys  with  monkeypox  virus.  It  looks  much  like  smallpox 
and  its  spread  appears  to  be  very  similar.  It  appears  that  infected  monkeys 
are  not  contagious  until  they  are  obviously  sick. 

They  did  a  study  of  the  effectiveness  of  cidofovir  against  variola. 
A  massive  variola  exposure  resulted  in  death  for  all  the  controls.  Three  of 
the  six  treated  animals  survived,  but  they  were  the  ones  treated  at  24  hours 
after  infection.  The  three  treated  at  12  hours  all  died. 

An  aerosol  exposure  of  the  monkeys  was  not  very  effective.  It 
takes  too  much  variola  material  to  infect  monkeys  through  the  air. 

He  showed  some  pictures  of  monkeys  with  variola  and  of  humans 
with  smallpox.  Both  are  pretty  scary.  His  description  was  “Really  ugly.” 


Spring  2006 


72 


The  infection  produces  nasty,  boil-like  sores  all  over,  inside  and  out. 

Mr.  Jahrling  concluded  that  effective  countermeasures  against 
smallpox  terrorism  can  be  developed  through  research  by  an  international 
community  of  scientists.  They  will  presumably  include  better  vaccines, 
antiviral  drugs,  and  methods  to  distribute  and  use  them. 

Mr.  Hershey  made  the  usual  announcements.  Finally,  at  9:46  p.m., 
he  adjourned  the  2192nd  meeting  to  the  social  hour. 

Attendance:  20 

Weather:  Misty  to  sprinkly 

Temperature:  12  C. 

Respectfully  submitted, 

Ronald  O.  Hietala,  Recording  secretary 
©  2005  Ronald  O.  Hietala 

Minutes  of  the  2194th  Meeting 
September  9,  2005 

Lecture:  Life  in  the  Age  of  Risk  Management 
Ms.  Kimberly  M.  Thompson 

President  Robert  Hershey  called  the  2194th  meeting  to  order  at  8: 17  p.m. 
on  September  9,  2005.  The  minutes  of  the  2192nd  meeting  were  read  and 
approved. 

Mr.  Hershey  introduced  the  speaker  of  the  evening,  Ms.  Kimberly 
M.  Thompson  of  the  Harvard  School  of  Public  Health,  where  she  is 
Associate  Professor  of  Risk  Analysis  and  Decision  Science  and  Director 
of  the  Kids  Risk  Project.  Ms.  Thompson  observed  that  it  was  a  pleasure  to 
be  speaking  in  the  John  Wesley  Powell  room;  she  attended  John  Wesley 
Powell  High  School  in  Mesa,  Arizona,  where  her  interest  in  science  was 
kindled. 

Ms.  Thompson  spoke  on  “Life  in  the  Age  of  Risk  Management” 
and  delighted  the  audience  with  cartoons  from  her  book  called  Risk  In 
Perspective :  Insight  and  Humor  in  the  Age  of  Risk  Management.  She 
posed  a  question  of  whether  we  are  living  in  a  state  of  fear,  and  quoted 
several  observations  of  Michael  Crichton,  who  said  that  “...  all  reality  is 
media  reality,”  and  that  the  politico-legal-media  complex  is  dedicated  to 
promoting  fear. 

Ms.  Thompson  said  we  are  living  in  a  time  when  life  is  full  of 
risks,  choices  often  involve  tough  trade-offs,  and  good  data  and  risk 


Washington  Academy  of  Sciences 


73 


analysis  play  a  critical  role  in  decisions,  both  individual  and  collective. 
These  are  the  characteristics  of  the  age  of  risk  management. 

At  this  time,  we  have  the  benefits  of  enormous  advancements  in 
science  and  technology.  We  have  high-quality  information  about  risks,  a 
large  spectrum  of  choices,  computational  tools,  and  better  understanding 
of  problems  and  solutions.  Success  still  depends  on  our  ability  to 
understand  and  communicate  how  things  work  and  to  effectively  manage 
variability  and  uncertainty. 

There  is  great  good  news  about  risks.  Since  1 900,  life  expectancy 
has  gone  from  less  than  50  years  to  over  75. 

The  big  questions  about  risks  are,  which  risks  are  big  and  which 
are  small,  what  can  and  should  we  do  about  them,  are  we  investing  the 
right  resources  and  spending  wisely,  what  data  do  we  need,  how  do  we 
deal  with  challenges  of  scale,  and  how  do  we  effectively  communicate? 
Communication  regarding  risks  especially  needs  our  attention. 

Risk  analysis  uses  mathematical  models  to  characterize 
information  -  what  can  happen,  how  likely  is  it,  and  if  it  happens,  what  are 
the  consequences?  The  models  are  used  to  evaluate  options  and  weigh 
trade-offs  -  what  can  we  do,  what  happens  if  we  do  it,  what  is  the  best 
option?  Finally,  we  need  to  communicate  risk  information. 

She  distinguished  variability  from  uncertainty.  Variability  is  the 
degree  of  heterogeneity  or  variability  in  a  population.  Uncertainty  is 
ignorance  about  a  poorly  characterized  phenomenon. 

She  discussed  three  examples  of  risk  problems  -  the  effectiveness 
of  airbags  in  cars,  the  mortality  risk  to  people  on  the  ground  from  crashing 
airplanes,  and  the  dynamics  of  managing  the  risks  of  polio. 

The  air  bag  question  had  its  roots  in  the  concern  about  the  40,000 
deaths  a  year  in  car  crashes.  There  was  a  time,  a  short  time,  when  a  car 
would  not  start  if  the  seat  belt  was  not  fastened.  People  did  not  like  that, 
and  it  did  not  last  long.  After  that,  rates  of  seat  belt  use  were  poor,  and 
interest  in  a  passive  method  developed.  This  interest  led  to  several  errors 
related  to  estimating  the  benefits  of  airbags.  The  fact  that  airbags  would 
kill  people  was  overlooked.  A  compliance  test  was  needed  to  show 
whether  airbags  worked,  and  the  standard  involved  in  the  test  led  to 
ignoring  variability  in  human  physique  and  behavior.  Not  everyone  is  a 
50th  percentile  male  and  not  everyone  sits  still  in  the  seat  facing  forward 
all  the  time,  like  a  dummy.  Also,  the  early  experience  with  seat  belts  led  to 
pessimism  regarding  seat  belt  use,  and  estimates  of  the  effectiveness  of 
airbags  inadvertently  included  people  whose  outcomes  should  have  been 
credited  to  seat  belts. 


Spring  2006 


74 


From  this  several  insights  were  drawn.  Technologies  perform 
differently  in  controlled  versus  uncontrolled  settings.  Efforts  should  be 
made  to  quantify  the  risks  of  safety  technologies.  Distributional  issues 
should  be  examined.  We  need  to  guard  against  overconfidence  and  we 
need  to  be  aware  of  political  realities. 

It  is  an  irony  that  airbags  are  an  engineering  solution  to  a 
behavioral  problem  and  that  the  result  of  children  being  killed  by  airbags 
necessitates  a  new  behavioral  solution  of  putting  kids  in  the  back. 
However,  we  have  learned  that  both  the  engineering  and  the  behavior 
matter. 

She  turned  to  the  problem  of  airplanes  crashing  on  people  on  the 
ground.  A  paper  in  1992  indicated  the  risk  of  this  was  four  in  a  million 
over  a  lifetime,  which  is  above  the  one  in  a  million  threshold  used  to 
identify  actionable  risks.  A  closer  look  at  the  data  now  reveals  some 
interesting  facts.  The  data  show  that  the  rates  of  accidents  have  gone  down 
substantially,  and  more  importantly  that  the  risks  differ  greatly  for  people 
near  and  far  from  airports.  The  risk  is  hundreds  of  times  higher  close  to 
major  airports.  This  leads  to  very  different  ideas  about  what  should  be 
done  about  the  risk.  It  also  shows  that,  for  most  of  us,  the  risk  is 
negligible. 

The  story  about  polio  is  one  of  a  brilliant  success  of  vaccines. 
Polio  paralysis  cases  in  the  U.S.  peaked  in  1952  at  over  21,000  cases.  Ms. 
Thompson  mentioned  that  we  recently  celebrated  50  years  of  polio 
vaccine  and  reminded  the  audience  about  the  major  headlines  on  April  12, 
1955  announcing  the  effectiveness  and  safety  of  the  vaccine.  The  decrease 
in  the  disease  after  that  was  dramatic,  and  for  some  time  now  we  have 
been  within  striking  distance  of  eradication. 

Oral  vaccine,  a  weakened -virus  vaccine,  causes  polio  at  a  very  low 
rate.  In  1979,  wild  cases  of  polio  virtually  disappeared.  In  the  late  1990s, 
the  U.S.  decided  the  risk  of  polio  from  oral  vaccine  was  unacceptable  and 
we  went  back  to  using  the  killed-virus,  injectable  vaccine. 

Since  the  oral  vaccine  is  a  live  virus  that  spreads  through  populations 
and  is  easier  to  administer,  it  is  the  preferred  method  if  we  want  to 
eradicate  polio.  However,  what  to  do  after  successful  eradication  remains 
a  difficult  choice. 

Ms.  Thompson  discussed  how  the  risks  and  concerns  changed  over 
time.  Sometimes  choices  get  tougher  as  the  original  problem  is  reduced. 
The  devil  is  in  the  details.  The  best  options  for  some  people  will  not  be  the 
best  for  others. 

In  concluding,  Ms.  Thompson  emphasized  the  need  for  good 


Washington  Academy  of  Sciences 


75 


science  to  realize  the  potential  of  risk  analysis  and  get  the  most  out  of  life 
in  the  age  of  risk  management. 

Mr.  Hershey  announced  the  next  meeting,  invited  people  to  apply 
for  membership  in  the  Society,  and  invited  everyone  to  enjoy  the  social 
hour.  He  adjourned  the  2194th  meeting  at  9:54  p.m.  to  the  social  hour. 

Attendance:  38 

Weather:  Clear,  mild,  beautiful 

Temperature:  15C. 

Respectfully  submitted, 

Ronald  O.  Hietala,  Recording  Secretary 

Minutes  of  the  2198th  Meeting 
December  10,  2005 
Lecture:  Einstein ’s  Warped  Universe 
Mr.  Ted  Jacobson 

President  Robert  Hershey  called  the  2198  meeting  to  order  at  8:20  p.m., 
on  December  10,  2005.  The  minutes  of  the  2197th  meeting  were  read  and 
approved. 

Mr.  Hershey  then  introduced  the  speaker  of  the  evening,  Mr.  Ted 
Jacobson  of  the  University  of  Maryland.  Mr.  Jacobson  spoke  on 
“Einstein’s  Warped  Universe.” 

“I’m  glad  to  have  the  opportunity  to  tell  you  some  things  about 
Einstein’s  warped  universe,”  Mr.  Jacobson  began.  He  noted  that  we  are  in 
the  anniversary  of  Einstein’s  “miraculous  year,”  1905,  in  which  he 
published  four  seminal  papers  that  spanned  all  of  the  revolutions  of 
modem  physics  of  that  time. 

One  introduced  the  idea  of  photons  and  the  particle  nature  of  light. 
Another  gave  calculations  of  Brownian  motion,  which  helped  to  establish 
the  atomic  nature  of  matter.  Third,  he  introduced  the  idea  of  mass  as  not 
independent  of  energy,  but  as  an  aspect  of  energy.  And  fourth,  he  showed 
that  simultaneity  and  space  and  time  measurements  depend  on  the  motion 
of  the  observer.  Einstein  was  26  at  the  time  and  a  clerk  in  a  patent  office. 
He  could  not  get  an  academic  job,  Mr.  Jacobson  joked. 

Einstein  did  say  that  “a  practical  profession  is  a  salvation  for  a  man  of  my 
type.  An  academic  career  compels  a  young  man  to  scientific  production, 
and  only  strong  characters  can  resist  the  temptation  of  superficial 
analysis.” 


Spring  2006 


76 


The  1905  work  on  time  and  space  is  called  special  relativity.  In  it, 
time,  space,  and  mass  are  taken  as  not  absolute.  This  work  omits  the 
whole  matter  of  gravity.  It  wasn’t  until  1915  that  he  incorporated  gravity 
and  inertia.  He  conceived  gravity  as  a  warping  of  time  and  space. 

What  does  it  mean  to  say  that  space  and  time  are  not  absolute?  In  1905, 
Einstein  had  inherited  Maxwell’s  theory  of  electrodynamics  and  Newton’s 
theory  of  mechanics.  Neither  of  them  implied  a  preferred  state  of  rest. 
Maxwell’s  theory  did  predict  that  light  and  electromagnetic  waves 
propagate  at  a  definite  speed.  Light  from  a  flashlight  travels  at  the  same 
speed  regardless  of  whether  the  flashlight  moves.  Einstein  saw  that 
therefore  we  cannot  attach  any  absolute  signification  to  the  concept  of 
simultaneity. 

Consider  a  flash  of  light  in  a  box.  If  the  flash  originates  in  the 
middle  of  the  box,  it  reaches  both  ends  at  the  same  time,  as  reckoned  by  an 
observer  at  rest  with  respect  to  the  box.  If  another  observer  is  running 
toward  the  box,  he  sees  the  box  approaching  him,  so  he  sees  the  light 
reach  the  back  end  before  it  reaches  the  front  end.  Thus,  a  time  ordering 
between  two  events  can  depend  on  who  is  observing.  This  is  the  relativity 
of  simultaneity. 

One  implication  of  this  relativity  is  that  time  elapsed  between  two 
events  depends  on  the  path  in  spacetime  that  connects  them.  One  who 
travels  a  straight  line  from  one  point  to  another  might  age,  say,  fifty  years. 
One  who  visits  a  distant  intermediate  point  on  the  interim  ages  less,  and 
the  difference  is  relative  to  the  additional  distance  traveled.  The  terms 
time  and  space  are  not  used.  Instead,  they  use  timelike  and  spacelike, 
because  what  it  means  to  remain  at  “the  same  point  of  space”  depends  on 
the  observer. 

Newton  conceived  gravity  as  a  universal  force,  an  attraction  of 
masses  to  each  other.  It  explained  both  the  falling  of  an  apple  and  the  orbit 
of  the  planets  around  the  sun. 

Einstein  said  he  was  sitting  in  the  patent  office  in  Bern  when  all  of 
a  sudden  a  thought  occurred  to  him:  If  a  person  falls  freely,  he  won’t  feel 
his  own  weight.  He  was  startled. 

This  simple  thought  made  a  deep  impression  on  him.  Then  he  had 
the  happiest  thought  of  his  life,  that  a  gravitational  field  has  only  a  relative 
existence. 

To  illustrate  Einstein’s  happy  thought,  Mr.  Jacobson  showed  a 
picture  of  a  dancer  in  an  airplane  following  a  parabolic  free-fall  arc, 
dancing  in  the  air  in  a  space  that  was  static  relative  to  the  confines  of  the 


Washington  Academy  of  Sciences 


77 


airplane  and  feeling  none  of  the  weight  of  Newton’s  force.  She  seemed  to 
be  having  a  good  time. 

Gravity  in  Einstein’s  conception  is  the  curvature  of  spacetime.  A 
freely  falling  object  follows  a  straight  line  in  spacetime.  When  an  apple 
falls,  what  really  happens  is  that  the  apple  and  the  earth  approach  each 
other.  Their  initially  parallel  paths  in  spacetime  do  not  remain  parallel, 
because  the  spacetime  is  curved.  This  is  analogous  to  two  lines  of 
longitude  on  the  earth  that  start  out  parallel  at  the  equator,  yet  converge  as 
they  proceed  north  towards  the  pole.  This  convergence  is  due  to  the 
curvature  of  the  surface  of  the  earth,  while  the  approach  of  a  falling  apple 
and  the  earth  is  due  to  curvature  of  spacetime.  He  described  a  little  of  how 
the  global  positioning  system  works.  The  earth  devices  have  clocks  on 
them.  Locations  are  determined  by  comparing  the  times  of  the  origin  of 
the  different  signals.  These  times  are  measurably  affected  by  relativity, 
due  to  both  the  motion  of  the  clocks  and  the  gravitational  time  dilation 
effect:  a  higher  clock  runs  more  quickly  than  a  lower  one. 

Gravity  bends  light.  Mr.  Jacobson  showed  some  pictures  of  a 
galaxy  viewed  through  the  gravitational  lens  of  another  galaxy.  Rings  of 
light  from  the  distant  object  that  appear  around  the  nearer  object  are  called 
Einstein  rings.  Actually,  you  usually  see  only  arcs,  not  complete  rings.  In 
some  cases,  the  time  difference  between  two  paths  of  light  from  the  same 
source  is  over  a  year. 

To  bring  the  matter  closer,  he  showed  a  picture  of  the  Smithsonian 
castle  and  another  picture  of  it  as  it  might  appear  through  a  gravitational 
lens,  as  if  a  black  hole  were  between  the  camera  and  the  castle.  Parts  of 
the  building  seemed  to  bend  around  and  enclose  parts  of  the  clouds  behind 
it. 

He  showed  a  picture  of  an  antenna  in  Puerto  Rico.  This  instrument 
first  detected  signals  from  a  binary  pulsar  in  1974.  The  binary  has  an  orbit 
period  of  eight  hours  and  a  pulse  period  of  59  ms.  Such  systems  emit 
gravitational  waves.  Gravitational  waves  carry  energy,  the  orbits  arc 
reduced,  and  the  orbit  period  has  been  observed  since  1974  to  decrease  at 
the  predicted  rate. 

As  another  approach,  there  is  an  attempt  underway  to  measure 
gravitational  waves  directly  using  very-long-baseline  laser  interferometry. 
Instruments  have  been  placed  far  apart  on  the  earth,  one  in  the  state  of 
Washington,  one  in  Louisiana.  Other  interferometers  elsewhere  on  earth 
are  also  being  used,  and  eventually  interferometers  in  space  are  planned. 
As  yet,  no  detection  has  been  achieved. 


Spring  2006 


78 


Then  he  took  up  some  other  questions  about  the  universe.  Is  it 
closed  or  open?  Is  it  curved  or  flat?  How  did  it  begin?  Does  it  expand 
forever?  What  about  the  beginning  of  time?  What  is  time  like  inside  a 
black  hole?  He  offered  some  speculations  on  these  questions. 

Mr.  Jacobson  offered  to  answer  questions  from  the  audience.  In 
response  to  questions,  he  noted  that  Einstein  made  gravity  equal  to  the 
curvature  of  spacetime,  which  means  that  gravity  is  the  warping  of  inertia. 

Someone  complained  that  he  would  never  understand  the 
spacetime  thing.  Where  does  it  get  started?  How  do  you  measure  time 
without  an  instrument?  You  can’t,  Jacobson  said.  Time  is  what  a  clock 
measures. 

Is  the  cosmos  as  a  whole  warped?  Yes,  it  is. 

Does  the  3-dimensional  universe  have  an  analog  in  a  2 
dimensional  universe? 

Yes,  this  is  an  outgrowth  of  string  theory.  It  is  a  theory  in  two 
dimensions  that  has  no  gravity. 

Electromagnetic  waves  are  harnessed  for  useful  purposes,  is 
anything  like  that  possible  for  gravitational  waves?  Yes,  for  astronomy. 
Could  we  generate  them?  Anything  we  can  generate  would  be  extremely 
weak.  He  guessed  detecting  waves  we  have  generated  won’t  happen  in  the 
next  50  years. 

The  annual  business  meeting  was  held  (the  business  records  are 
kept  by  the  corresponding  secretary). 

Mr.  Hershey  announced  the  next  lecture  and  invited  guests  to  join 
the  Society.  Finally,  he  made  the  parking  announcement,  invited  everyone 
to  enjoy  the  social  hour,  and  at  9:50  p.m.,  adjourned  the  2197th  meeting. 

Attendance:  71 

Weather:  Unremarkable 

Temperature:  6  C. 

Respectfully  submitted, 

Ronald  O.  Hietala,  Recording  Secretary 
©  2005  Ronald  O.  Hietala 


Washington  Academy  of  Sciences 


79 


MARINE  TECHNOLOGY  SOCIETY  NEWS 


The  Marine  Techndy  Soiety  has  published  MTS 
Journal:  Promoting  Lifelong  Ocean  Education  (Winter  2005/2006).  This 
issue  of  the  quarterly,  peer-reviewed  journal  describes  solutions  to  the 
problem  of  ocean  literacy,  and  includes  information  on  successful 
programs  that  are  currently  promoting  learning.  According  to  the  issue's 
editor,  Blanche  W.  Meeson,  of  Oceans  US  and  the  National  Oceanic  and 
Atmospheric  Administration  (NOAA),  "To  my  knowledge  this  is  a  first: 
an  issue  of  a  major  science  and  technology  society's  signature  journal 
dedicated  to  education." 

MTS  has  developed  an  Experts  Directory  and  a  Speakers  Bureau, 
both  of  which  are  searchable  databases  on  the  MTS  Web  site  at 
www.mtsociety.org.  The  directories  are  available  to  anyone  wishing  to 
find  marine  technologists  and  engineers  for  speaking  engagements, 
collaborations,  general  questions,  etc. 

Justin  Manley  is  the  new  editor  of  the  MTS  Journal.  As  editor, 
Manley  has  three  goals:  to  continue  to  improve  the  Journal's  quality;  to 
capitalize  on  the  diverse  interests  of  MTS  as  a  strength  so  as  to  provide  a 
unique  perspective  on  the  intersection  of  science/technology,  business  and 
policy;  and  to  frame  MTS  and  public  discussion  of  key  issues,  such  as 
ocean  energy  and  the  role  of  advanced  technology  in  the  oceans.  Manley 
is  a  senior  engineer  at  Battelle.  He  chairs  both  the  National  Atmospheric 
and  Space  Administration's  Autonomous  Underwater  Vehicle  Working 
Group  and  the  MTS  AUV  Professional  Committee.  Manley  replaces 
outgoing  editor  Dan  Walker,  senior  program  officer  of  the  National 
Research  Council. 

In  April,  MTS  sponsored  the  online  Pre-Engineering  Times,  a 
publication  of  JETS,  which  works  to  increase  interest  and  awareness  of 
engineering  and  technology-based  careers.  As  sponsor,  MTS  provided  two 
articles  to  the  publication,  one  aimed  at  young  people  who  might  be 
interested  in  focusing  their  engineering  interest  in  the  marine  sciences  and 
another  on  the  society-sponsored  student  outreach  programs  and 
scholarships. 

The  MTS-sponsored  Oceans  2006  Conference  is  scheduled  for 
September  18-21  in  Boston.  The  conference  plans  to  highlight  several  "hot 


Spring  2006 


80 


topic"  areas,  including  homeland  security  applications,  tsunami  early- 
warning  systems,  autonomous  underwater  vehicle/unmanned  undersea 
vehicle/glider  technology,  distributed  sensors  and  networks,  tracking  and 
data  fusion,  non-acoustic  sensing  and  imaging,  integrated  ocean 
observatories,  marine  mammal  classification,  Artie  Ocean  science,  optical 
properties  of  water,  aquaculture  engineering  and  marine  archaeology.  For 
more  information,  visit  www.oceans06mtsieeeboston.org. 

The  MTS-sponsored  Dynamic  Positioning  Conference  2006  is 
scheduled  for  October  17-18  in  Houston,  Texas.  The  annual  DP 
conference  attracts  leading  DP  professionals  from  around  the  world. 
Check  the  Web  site  to  find  out  when  registration  will  begin.  Exhibitors  are 
encouraged  to  reserve  space  now,  since  space  is  limited.  For  more 
information,  visit  www.dynamic-positioning.com. 

The  MTS-sponsored  Underwater  Intervention  2007  is  scheduled 
for  January  30-February  1,  in  New  Orleans,  La.  Prospective  speakers  are 
invited  to  submit  proposals  to  the  conference  in  one  of  the  following 
areas:  commercial  diving,  ROVs,  AUVs,  sonar,  acoustics,  underwater 
inspections,  underwater  construction/repairs,  training/education,  legal  and 
regulatory,  safety  issues,  certification,  underwater  cutting/welding, 
equipment  maintenance,  bid  specifications  and  military  issues.  Deadline 
for  abstract  submission  is  July  15,  2006,  and  final  papers  are  due 
December  15,  2006.  The  conference  draws  over  2,000  attendees — 
purchasing  agents,  project  managers,  engineers,  operations  managers, 
owners,  directors  and  many  other  key  decision  makers  from  the  United 
States  and  30  other  countries.  For  more  information,  visit 
www.underwaterintervention.com. 

MTS  members  participated  in  the  Consortium  for  Oceanographic 
Research  and  Education  (CORE)  Public  Policy  Forum  in  Washington, 
D.C.,  in  March.  Among  those  speaking  was  Vice  Adm.  Conrad 
Lautenbacher.  Andrew  Clark  participated  in  a  panel  discussion  on  Ocean 
Observing  Systems.  Richard  Spinrad  and  Shirley  Pomponi  were  part  of  a 
panel  discussion  on  Ocean  Research  Priorities  Plan  and  Implementation 
Strategy. 

MTS  is  one  of  the  hosts  of  the  Conference  on  Ocean  Literacy 
(CoOL),  June  7-8,  at  the  Ronald  Regan  Building  and  International  Trade 
Center  in  Washington,  D.C.  The  two-day  forum  will  bring  together 
members  of  government,  education,  textbook  publishing,  industry,  science 
centers,  non-profits  and  other  interested  entities  to  discuss  ocean  literacy 


Washington  Academy  of  Sciences 


81 


and  the  challenges  and  opportunities  for  educating  the  public  to  make 
informed,  responsible  decisions  about  the  ocean  and  its  resources.  The 
conference  chair  is  MTS  member  Sharon  Walker. 

MTS  member  Shirley  Pomponi  has  been  appointed  to  the  Florida 
Oceans  and  Coastal  Council.  The  council  will  develop  priorities  for  ocean 
and  coastal  research,  and  establish  a  statewide  ocean  research  plan.  The 
group  will  also  coordinate  public  and  private  ocean  research  for  more 
effective  coastal  management. 

MTS  member  JDR  Cable  Systems  has  appointed  John  R.  Havey  as 
technical  sales  manager  of  the  JDR  Oil  and  Gas  Division,  North  America. 
The  announcement  was  made  by  MTS  member  Paul  Gahm,  executive  vice 
president  of  sales  and  marketing,  JDR  Oil  and  Gas  Division.  Formerly 
sales  manager  at  Dril-Quip,  Havey  is  a  30-year  veteran  in  the  offshore  oil 
and  gas  industry. 

The  Supervisory  Board  of  MTS  member  INTEC  Engineering  and 
Heerema  Holding  have  named  Bruce  Crager  to  be  chief  executive  officer 
of  INTEC  Engineering  based  in  Houston.  Crager  brings  more  than  30 
years  of  experience  in  the  oil  and  gas  industry. 

MTS  member  Eric  Steimlc  of  the  University  of  South  Florida-St. 
Petersburg,  developed  a  radio-controlled  guided  surface  vehicle  (GSV) 
that  carried  a  D1DSON  imagining  sonar  and  a  hydrophone  listening  device 
to  eavesdrop  on  the  spawning  sounds  of  black  drum  fish.  The  fish  raise  a 
loud  chorus  when  they  spawn.  The  instrumentation  helped  to  determine 
whether  the  sound  production  was  matched  by  real  results— tight  clusters 
of  newly  fertilized  fish  eggs. 

MTS  member  Teledyne  RD  Instruments  announced  the 
appointment  of  William  Kikendall  as  general  manager.  Kikendall  has 
served  as  general  manager  of  Teledyne  Geophysical  Instruments, 
Houston,  Texas,  for  the  past  six  years  and  will  now  oversee  operations  at 
both  Teledyne  Technologies'  facilities. 

Jill  Zande  is  the  new  chair  of  the  Monterey  Section  of  the  Marine 
Technology  Society.  Zande  is  outreach  director  and  ROV  Competition 
coordinator  of  the  MATE  Center  in  Monterey,  Calif. 

MTS  member  John  Moore  moderated  a  panel  discussion  on 
American  Security  Interests  and  the  Law  of  the  Sea  at  a  Senate 
appropriations  hearing  in  Washington,  D.C.,  in  April.  Moore  is  the 
director  of  the  Center  for  Oceans  Law  and  Policy  at  the  University  of 


Spring  2006 


82 


Virginia  School  of  Law.  Among  the  panelists  contributing  to  the 
discussion  was  MTS  member  Douglas  Burnett,  an  international  law 
adviser  and  a  partner  with  Holland  and  Knight,  LLP. 

MTS  member  William  Kuperman  of  the  University  of  California, 
Scripps  Institution  of  Oceanography  is  on  the  National  Research  Council- 
approved  Committee  to  Review  the  Joint  Subcommittee  on  Ocean  Science 
and  Technology  (JSOST)  Ocean  Research  Priorities  Plan. 

MTS  member  Dr.  Reginald  Beach  has  served  as  the  Consortium 
for  Oceanographic  Research  and  Education  (CORE)  director  of  research 
for  the  last  four  years  and  is  leaving  to  become  the  chief  scientist  for  the 
Ocean  Exploration  program  at  NOAA. 

Offshore  survey  company  and  MTS  member  Fugro  Chance 
recently  promoted  Greg  Pilgrim  to  operations  manger.  Marine 
Construction  Survey  in  Houston,  Texas.  The  Marine  Construction  Survey 
Group  develops  innovative  survey  methods  and  customized  procedures  for 
deepwater  construction  projects. 

MTS  was  a  signatory  to  letters  sent  to  U.S.  Senate  and  House 
appropriations  subcommittees  on  Science,  State,  Justice  and  Related 
Agencies  encouraging  them  to  provide  NOAA  with  an  appropriation  of 
$4.5  billion  in  fiscal  year  2007.  "NOAA  is  critical  to  protecting  ou*  ocean 
resources,  coastal  communities  and  economy.  In  fact,  weather  and  nate 
sensitive  industries  account  for  about  one-third  of  the  Nation’s  GDP.  An 
investment  of  $4.5  billion  averages  out  to  $15  per  person  annually,"  the 
letters  stated,  adding  that  this  "small  amount"  provides  an  enormous 
amount  of  information,  research  and  local  community  assistance.  "A  better 
understanding  of  the  oceans  not  only  benefits  coastal  communities.  In  fact, 
economists  have  estimated  that  altering  planting  decisions  based  on 
improved  El  Nino  and  La  Nina  forecasts  would  save  U.S.  farmers  $265- 
$300  million." 

Susan  M.  Branting 

Communications  Manager,  MTS 


Washington  Academy  of  Sciences 


DELEGATES  TO  THE  WASHINGTON  ACADEMY  OF  SCIENCES 
REPRESENTING  AFFILIATED  SCIENTIFIC  SOCIETIES 


Acoustical  Society  of  America 

Paul  Arveson 

American/Intemational  Association  of  Dental  Research 

J.  Terrell  Hoffeld 

American  Association  of  Physics  Teachers 

Frank  R.  Haig,  S .J. 

American  Ceramics  Society 

VACANT 

American  Fisheries  Society 

Ramona  Schreiber 

American  Institute  of  Aeronautics  and  Astronautics 

David  W.  Brandt 

American  Institute  of  Mining,  Metallurgy  &  Exploration 

Michael  Greeley 

American  Meteorological  Society 

Kenneth  Carey 

American  Nuclear  Society 

Steven  Arndt 

American  Phytopathological  Society 

Kenneth  L.  Deahl 

American  Society  for  Cybernetics 

Stuart  Umpleby 

American  Society  for  Microbiology 

VACANT 

American  Society  of  Civil  Engineers 

Kimberly  Hughes 

American  Society  of  Mechanical  Engineers 

Daniel  J.  Vavrick 

American  Society  of  Plant  Physiology 

Mark  Holland 

Anthropological  Society  of  Washington 

Marilyn  London 

ASM  International 

Toni  Marechaux 

Association  for  Women  in  Science  (AWIS) 

Emanuela  Appetiti 

Association  for  Computing  Machinery 

Lee  Ohringer 

Association  for  Science,  Technology,  and  Innovation 

F.  Douglas  Witherspoon 

Association  of  Information  Technology  Professionals 

Barbara  Saffanek 

Biological  Society  of  Washington 

VACANT 

Botanical  Society  of  Washington 

Alain  Touwaide 

Chemical  Society  of  Washington 

James  J.  Zwolenik 

District  of  Columbia  Institute  of  Chemists 

James  J.  Zwolenik 

District  of  Columbia  Psychology  Association 

David  Williams 

Eastern  Sociological  Society 

Ronald  W.  Mandersheid 

Electrochemical  Society 

Robert  L.  Ruedisueli 

Entomological  Society  of  Washington 

F.  Christian  Thompson 

Geological  Society  of  Washington 

Bob  Schneider 

Historical  Society  of  Washington,  DC 

VACANT 

History  of  Medicine  Society 

Alain  Touwaide 

Human  Factors  and  Ergonomics  Society 

Douglas  Griffith 

Institute  of  Electrical  and  Electronic  Engineers 

Sajjad  Durrani 

Institute  of  Electrical  and  Electronic  Engineers 

Murty  Polavarapu 

Institute  of  Food  Technologies 

Isabel  Walls 

Institute  of  Industrial  Engineers 

Neal  F.Schmeidler 

Instrument  Society  of  America 

Hank  Hegner 

Marine  Technology  Society 

Judith  T.  Krauthamer 

Mathematical  Association  of  America 

Sharon  K.  Hauge 

Medical  Society  of  the  District  of  Columbia 

Duane  Taylor 

National  Capital  Astronomers 

Jay  H.  Miller 

National  Geographic  Society 

VACANT 

Optical  Society  of  America 

Jim  Cole 

Pest  Science  Society  of  America 

VACANT 

Philosophical  Society  of  Washington 

Vary  T.  Coates 

Society  of  American  Foresters 

G.  Foster 

Society  of  American  Military  Engineers 

VACANT 

Society  of  Experimental  Biology  and  Medicine 

Darren  Roesch 

Society  of  Manufacturing  Engineers 

VACANT 

Soil  and  Water  Conservation  Societyw 

Bill  Boyer 

Technology  Transfer  Society 

Clifford  Lanham 

Washington  Evolutionary  Systems  Society 

Jerry  L.R.  Chandler 

Washington  History  of  Science  Club 

Albert  G.  Gluckman 

Washington  Chapter  of  the  Institute  for  Operations 

Research  and  Management  Science 

Russell  R.  Vane  III 

Washington  Paint  Technology  Group 

VACANT 

Washington  Society  of  Engineers 

Alvin  Reiner 

Washington  Statistical  Society 

Michael  P.  Cohen 

World  Future  Society 

Diane  Pickar 

Washington  Academy  of  Sciences 
Room  637 

1200  New  York  Ave.  NW 
Washington,  DC  20005 
Return  Postage  Guaranteed 


ERNST  MAYR  LIBRARY 
MUSEUM  COMP  ZOOLOGY 
HARVARD  UNIVERSITY 
26  OXFORD  ST 

CAMBRIDGE.  MA  02138-2902 


MEDIA  MAIL 
U.S.  POSTAGE 
PAID 

PERMIT  1502 
MERR1FIELD.  VA 


MCZ 

LIBRARY 


OCT  1  1  2006 

HARVARD 

UNIVERSITY 


Volume  92 
Number  2 
Summer  2006 


Journal  of  the 


WASHINGTON 
ACADEMY  OF  SCIENCES 


Contents 


Instructions  for  Authors .  i 

Incoming  President’s  Message .  ii 

Affiliated  Institutions  . iv 

William  T.  Franz,  Bottle  Rockets,  Teacups  and  the  Real  World . 1 

Carl  E.  Mungan,  Relative  Speeds  of  Interacting  Astronomical  Bodies . 7 

Colin  F.  Mackenzie  and  Yan  Xiao,  Videos  of  Emergency  Care  Show  Challenges  for  Patient 
Safety . 15 

M.  Sue  Bogner,  It’s  Not  Who  in  98,000  Medical  Error  Deaths,  It’s  What  . 29 

Gerald  P.  Krueger,  Fatigue,  Drowsy  Decision-Making  and  Medical  Error:  Issues  of  Quality  Health 
Care . 41 

Frank  R.  Haig,  S.J.  and  Peg  Kay,  The  Role  of  Academies  of  Science  in  the  Critical  Examination 
of  New  Ideas:  Looking  at  Gaia  . 61 

Affiliated  Societies . Inside  back  cover 


ISSN  0043-0439 


Issued  Quarterly  at  Washington  DC 


^asJjtngton  Ucabemp  of  ^cicntcs! 

Founded  in  1898 


Board  of  Managers 
Elected  Officers 

President 

William  Boyer 
President  Elect 

Alain  Towaide 
Treasurer 

Harvey  Freeman 
Secretary 

James  Cole 

Vice  President,  Administration 
Rex  Klopfenstein 
Vice  President,  Membership 

Thomas  Meylan 

Vice  President,  Junior  Academy 

Paul  L.  Hazan 

Vice  President,  Affiliated  Societies 

Mark  Holland 
Members  at  Large 

Sethanne  Howard 
Donna  Dean 
Frank  Haig,  S  .J. 

Jodi  Wesemann 
Vary  Coates 
Peg  Kay 

Past  President:  F.  Douglas  Witherspoon 

AFFILIATED  SOCIETY  DELEGATES: 
Shown  on  back  cover 


Editor  of  the  Journal 

Vary  T.  Coates 

Associate  Editors: 

Alain  Touwaide 
Sethanne  Howard 

Academy  Office 

Washington  Academy  of  Sciences 
Room  63 1 


The  Journal  of  the  Washington  Academy  of 
Sciences 

The  Journal  is  the  official  organ  of  the  Academy. 
It  publishes  articles  on  science  policy,  the  history  of 
science,  critical  reviews,  original  science  research, 
proceedings  of  scholarly  meetings  of  its  Affiliated 
Societies,  and  other  items  of  interest  to  its  members. 
It  is  published  quarterly.  The  last  issue  of  the  year 
contains  a  directory  of  the  current  membership  of 
the  Academy. 

Subscription  Rates 

Members,  fellows,  and  life  members  in  good 
standing  receive  the  Journal  free  of  charge. 
Subscriptions  are  available  on  a  calendar  year  basis, 
payable  in  advance.  Payment  must  be  made  in  U.S. 


currency  at  the  following  rates. 

US  and  Canada .  $25.00 

Other  Countries  .  30.00 

Single  Copies  (when  available)  .  10.00 


Claims  for  Missing  Issues 

Claims  must  be  received  within  65  days  of  mailing. 
Claims  will  not  be  allowed  if  non-delivery  was  the 
result  of  failure  to  notify  the  Academy  of  a  change 
of  address. 

Notification  of  Change  of  Address 

Address  changes  should  be  sent  promptly  to  the 
Academy  Office.  Notification  should  contain  both 
old  and  new  addresses  and  zip  codes. 

POSTMASTER: 

Send  address  changes  to  WAS,  Rm.631, 

1200  New  York  Ave.  NW 
Washington,  DC.  20005 

Journal  of  the  Washington  Academy  of  Sciences 
(ISSN  0043-0439) 

Published  by  the  Washington  Academy  of  Sciences 
202/326-8975 


1200  New  York  Ave.  NW 

Washington,  DC  20005  website:  www.washacadsci.org 

Phone:  202/326-8975 
email:  was@washacadsci.org 


MCZ 

LIBRARY 


INSTRUCTIONS  FOR  AUTHORS  OCT  1  1  2006 

HARVARD 

UNIVERSITY 

THE  JOURNAL  of  the  Washington  Academy  of  Sciences  is  a  peer- 
reviewed  journal.  Exceptions  are  made  for  papers  requested  by  the  editors 
or  positively  approved  for  presentation  or  publication  by  one  of  our 
affiliated  scientific  societies. 

We  welcome  disciplinary  and  interdisciplinary  scientific  research  reports 
and  papers  on  technology  development  and  innovation,  science  policy, 
technology  assessment,  and  history  of  science  and  technology.  Book 
reviews  are  also  welcome. 

Contributors  of  papers  are  requested  to  follow  these  guidelines  carefully. 

Papers  should  be  submitted  as  e-mail  attaclunents  to  the  chief  editor,  vcoafes  r/  mac.com. 
along  with  full  contact  information  for  the  primary  or  corresponding  author. 

Papers  should  be  presented  in  Word;  do  not  send  PDF  files. 

Papers  should  be  6000  words  or  fewer.  If  more  than  6  graphics  are  included  the  number 
of  words  allowed  will  be  reduced  accordingly. 

Graphics  must  be  in  black  and  white  only.  They  must  be  easily  resized  and  relocated.  It  is 
best  to  put  graphics,  including  tables,  at  the  end  of  the  paper  or  in  a  separate  document, 
with  their  preferred  location  in  the  text  clearly  indicated. 

References  should  be  in  the  form  of  endnotes,  and  may  be  in  any  sty  le  considered 
standard  in  the  discipline(s)  represented  by  the  paper. 


The  editor  for  this  edition  of  the  Journal  is  Sethanne  Howard.  Those 
papers  by  Mackenzie,  Bogner,  and  Krueger  are  from  the  Potomac  Chapter 
of  the  Human  Factors  and  Ergonomics  Society  session  of  Capital  Science 
2006,  The  papers  by  Mungan  and  Franz  are  from  the  American 
Association  of  Physics  Teachers,  Chesapeake  Section  session. 


Summer  2006 


II 


At  the  Academy ' s  Animal  Awards  Banquet ,  on  May  9  at  the  United  States 
Botanic  Garden ,  Bill  Boyer  was  installed  as  president  for  2006-2007 


A  MESSAGE  FROM  THE  PRESIDENT: 

At  my  first  meeting  of  the  Board  of  Managers  of  the  Washington 
Academy  of  Sciences  I  was,  and  still  am  struck  by  their  positive  team 
work  and  dedication.  At  a  time  when  many  non-profit  professional 
organizations  are  struggling  just  to  keep  going,  the  Academy  stands  out. 
Even  among  other  academies  of  sciences  in  other  states  and  cities,  the 
Washington  Academy  stands  out. 

Several  months  ago  Peg  Kay  asked  me  what  plans  I  have  to  lead 
this  organization.  Sometimes  this  organization  reminds  me  of  a  sailboat 
gliding  across  the  Chesapeake  Bay.  The  sails  are  out,  the  breeze  is  steady, 
the  crew  is  working  like  an  organized  team,  and  I’m  holding  on  for  dear 
life. 


After  our  recent  Capital  Science  event  I  asked  members  of  the 
Board  what  THEY  wanted  for  the  future  direction  of  the  Academy.  Here 
is  what  they  said  are  the  strong  points  of  the  Academy: 

•  Younger  people  are  stepping  up  to  leadership  roles; 

•  Traditions  remain  strong  and  are  cherished; 

•  Unlike  Academia  or  much  of  the  government,  the  Board  of 
Managers  can  disagree  without  getting  personal; 

•  Our  Junior  Academy  and  STARS  program  (Science  and 
Technology  Aptitude  Recognition  for  Schools)  encourages  a 
future  supply  of  young  scientists; 

•  Sponsorship  of  conferences,  like  Capital  Science  and 
specialized  symposia  serve  the  local  and  regional  scientific 
community; 

•  The  Journal  provides  a  venue  for  interdisciplinary  and  offbeat 
research  and  analysis ; 

•  The  Awards  Banquet  recognizes  outstanding  contributors  to 
science,  scientific  education,  and  public  service; 


Washington  Academy  of  Sciences 


Ill 


•  The  dissemination  of  news  about  the  meetings  and  activities  of 
affiliated  scientific  societies  encourages  networking  across 
disciplines,  professions,  and  institutions. 

Areas  where  the  Academy  needs  some  attention  and  improvement 
include: 

•  We  need  to  do  a  better  job  of  involving  and  serving  our 
Affiliates; 

•  The  Junior  Academy  should  move  beyond  Science  Fair 
judging,  perhaps  to  offer  scholarships,  for  example; 

•  We  need  to  secure  a  stronger  financial  foundation; 

•  We  should  improve  our  benefits  and  activities  for  members  and 
fellows  who  are  not  on  the  Board; 

•  We  should  have  more  academic  members,  considering  the 
many  colleges  in  the  area. 


Other  specific  ideas  were  offered  that  will  be  discussed  in  the 
coming  year. 


I  asked  the  Board  members  to  envision  the  Academy  5-10  years 
from  now  and  here’s  what  they  saw  for  the  future: 

•  The  Washington  Academy  as  a  full-fledged  member  of  the 
area’s  scientific  establishment,  called  on  for  Congressional 
testimony,  interviewed  by  the  Post  on  matters  of  science  and 
science  policy; 

•  A  vibrant  Junior  Academy  built  along  the  lines  of  the  Academy 
and  graduating  members  able  to  step  into  careers  in  science  and 
roles  in  the  Academy; 

•  Involvement  of  more  people  with  increased  diversity; 

•  An  operations  manual  with  guidance  for  new  board  members; 

•  Regular  activities  for  all  members,  in  addition  to  Board  of 
Managers  meetings; 

•  Membership  should  double  or  triple  what  it  is  today. 


Summer  2006 


IV 


As  you  can  tell,  there  is  belief  in  the  Board  that  the  Academy 
should  be  used  to  spread  the  word  that  science  is  an  important  part  of  the 
lives  of  the  people  living  in  this  region.  Even  when  science  seems  to  be  a 
pawn  of  politics,  there  are  scientists  and  scientific  organizations  working 
hard  to  make  sure  science  is  recognized  as  a  tool  to  make  our  lives  better. 

In  1898,  the  purpose  of  the  new  Academy  was  to  encourage  the 
advancement  of  science  and  “to  conduct,  endow,  or  assist  investigation  in 
any  department  of  science.”  That  purpose  guided  the  Academy  throughout 
its  first  100  years  and  will  continue  to  be  our  guide  through  the  coming 
century. 

Bill  Boyer ,  President 


Other  newly  elected  officers  of  the  Washington  Academy  of  Sciences 
(2006-2007): 

President  Elect:  Alain  Touwaide 
Treasurer:  Harvey  Freeman 
Secretary:  James  Cole 

Vice  President  for  Administrative  Affairs:  Rex  Klopfenstein 
Vice  President  for  Membership  Affairs:  Tom  Meylan 
Vice  President  for  Affiliate  Affairs:  Mark  Holland 
Vice  President  for  Junior  Academy:  Paul  Hazan 
At  Large  Members  of  the  Board:  Peg  Kay  and  Vary  Coates, 

Journal  Editor 

Immediate  Past  President:  Douglas  Witherspoon 

AFFILIATED  INSTITUTIONS 

The  National  Institute  for  Standards  and  Technology 
Meadowlark  Botanical  Gardens 
The  John  W.  Kluge  Center  of  the  Library  of  Congress 
Potomac  Overlook  Regional  Park 


Washington  Academy  of  Sciences 


Bottle  Rockets,  Teacups  and  the  Real  World: 

A  senior  seminar  to  bridge  the  gap  between  physics 
students  and  life  after  college** 

William  T.  Franz 

Department  of  Physics,  Randolph-Macon  College,  Ashland,  VA  23005 


Abstract 

One  of  the  peculiar  aspects  to  being  a  professional  physicist  is  the 
authority  we  all  seem  to  have  to  comment  on  'real  life'  phenomena.  I 
have  been  asked  about  event  hi  ng  from  divining  rods  to  space  junk 
during  my  career.  The  senior  seminar  at  Randolph-Macon  College  is 
designed  to  be  a  culminating  experience  that  asks  students  to  synthesize 
their  course  and  research  experience  and  improve  their  presentation 
skills.  The  most  recent  iteration  focused  on  problems  that  varied  from 
urban  legends  to  wacky  theories  with  an  emphasis  on  laboratory 
measurement,  practical  calculation,  and  presentation  of  results. 

Methods  for  heating  water  to  make  tea  and  the  practicality  of  launching 
people  with  bottle  rockets  will  be  discussed. 

Introduction 

The  new  curriculum  at  Randolph-Macon  College  asks  each 
department  to  create  a  culminating  or  “capstone”  experience  for  its  major 
students.  The  goals  of  the  college  curriculum  include  the  statement:  “Each 
student  should  participate  in  some  activity  which  draws  together  principles 
from  various  courses  of  study,  examines  a  topic  of  special  interest  using 
skills  and  abilities  drawn  from  several  courses,  or  invites  comparisons  and 
contrasts  about  components  of  the  major  courses  of  instruction.”1 

The  physics  department  has  instituted  a  number  of  courses  in 
guided  research  and  a  senior  seminar.  Through  the  guided  research 
courses,  students  engage  in  an  independent  research  project  of  their  own 
design  under  the  tutelage  of  a  departmental  faculty  member.  Several 
projects  have  resulted  in  publication  and/or  presentation.  The  senior 
seminar  course  includes  as  its  goals  -  development  of  communication 
skills,  both  written  and  oral;  integration  of  course  material  with  real-world 
applications;  and  building  a  bridge  between  the  academic  world  and  the 


Frank  Haig  Prize  winner,  presented  at  the  Chesapeake  Section  of  the  American 
Association  of  Physics  Teachers 


Washington  Academy  of  Sciences 


2 


so-called  real  world  after  college.  The  emphasis  is  on  skills  so  that  the 
choice  of  content,  while  important,  is  largely  irrelevant. 

The  bridge  to  life  after  college  is  taken  quite  seriously  in  the  design 
of  the  senior  seminar.  Many  of  our  graduates  will  find  themselves  in  front 
of  a  classroom  at  some  point  in  their  post-graduate  lives.  For  some,  they 
are  intending  a  career  in  education,  for  others,  there  will  be  opportunities 
as  a  teaching  assistant  in  graduate  school.  Therefore  the  seminar  includes  a 
component  designed  to  teach  some  basic  educational  theory,  and  it 
provides  an  opportunity  to  practice  teach  in  both  informal  settings  and  in 
our  introductory  class. 

Life  after  college  is  also  defined  by  the  need  to  search  for  solutions 
where  the  answer  is  not  in  the  back  of  a  book  and  the  methodology  is  not 
taught  in  one  particular  course.  The  seminar  therefore  provides  an 
opportunity  for  problem  solving  in  “non-traditional”  applications.  This 
work  is  devoted  to  two  examples  of  problems  chosen  for  the  seminar,  the 
student  results,  and  their  presentations. 


The  Research  Problems 


PROBLEM  #J:  The  Great  Tea  Cup  Controversy 

An  e-mail  from  a  colleague2  posed  an  interesting  question.  A 
discussion  in  a  local  law  office  had  reached  a  spirited  level  regarding  the 
proper  method  of  heating  water  for  preparing  tea.  One  person  alleged  that 
water  heated  in  a  teapot  retained  its  heat  longer  than  water  heated  in  a 
microwave  oven  and  was  therefore  a  better  method  for  brewing  tea.  In 
response  to  the  colleague,  it  was  suggested  that  water  had  a  notoriously 
poor  memory  and  that  the  method  used  to  heat  the  water  would  have 
nothing  to  do  with  the  rate  at  which  it  cooled.  Nonetheless,  there  was  the 
possibility  that  the  container  into  which  the  water  was  poured  could  be 
influencing  the  observations  as  heated  water  poured  into  a  cold  vessel 
would  be  found  cooler  than  water  remaining  in  a  teapot  that  had  been 
heated  along  with  the  water. 

Despite  the  sound  physics  and  the  reasonable  attempt  at  forcing  a 
compromise,  the  colleagues  remained  unconvinced  by  the  solution.  The 
problem  was  posed  to  the  seminar  students  and  they  were  asked  to  conduct 
a  series  of  definitive  experiments  to  prove  the  point. 


Summer  2006 


3 


Experiments  were  conducted  with  water  being  heated  by  three 
different  methods  -  electric  teapot,  tea  kettle  on  a  hot  plate,  and  in  a 
microwave  oven.  Water  was  brought  nominally  to  boil  and  then  poured 
into  cups  initially  at  room  temperature.  Experiments  were  conducted  using 
Styrofoam  cups  and  aluminum  beakers.  Using  temperature  probes 
interfaced  through  a  Pasco  interface  system3,  the  students  measured  the 
temperature  of  the  water  as  it  cooled  for  several  minutes  and  then  fit  their 
data  to  a  traditional  Newton’s  Law  of  Cooling  curve  and  determined  the 
fitting  parameters.  Figures  1  and  2  show  typical  data  obtained. 


Water  cooling  rates  in  Styrofoam  Cups 


100  150 

Time  (seconds) 


[—♦—Conventional  «  Electrical  Microwave  - Expon.  (Conventional)  —  Expon.  (Electrical) - Expon.  (Microwave)  | 


Figure  1 

Results  of  the  student  experiments  revealed  that  the  variation  in 
cooling  rate  from  one  method  of  heating  to  another  was  smaller  than  the 
experimental  uncertainty  where  the  uncertainty  was  determined  by 
examining  the  variation  among  values  obtained  in  several  repetitions  of 
the  cooling  experiment  for  a  single  set  of  experimental  variables  (same 
method  of  heating,  same  vessel). 

Students  presented  their  results  in  a  formal  talk  before  an  audience 
of  other  students  and  faculty  in  a  seminar  setting.  They  prepared  a  set  of 
power  point  slides  that  described  the  nature  of  the  problem,  the 
experimental  protocol  they  had  developed,  the  results,  and  a  discussion  of 
the  validity  of  their  results. 


Washington  Academy  of  Sciences 


4 


Water  cooling  rates  in  Aluminum 


— 


— 


150 

Time  (seconds) 


Conventional  -»•-  Electrical  Microwave - Expon  (Conventional) - Expon  (Electrical) - Expon  (Microwave)  | 


Figure  2 


PROBLEM  #2:  Launching  a  person  with  water  rockets 

A  cable  television  network4  shows  a  series  called  “MXC”  based  on 
the  Japanese  game  show  “Takeshi’s  Castle.”  The  program,  a  rather 
slapstick  reality  show,  has  something  of  a  cult  following.  A  film  clip  from 
this  program  showing  a  young  Japanese  man  being  launched  using  a  back 
pack  consisting  of  water  rockets  has  been  circulating  on  the  Internet.  One 
version  of  the  clip  has  been  found  on  a  web  page  devoted  to  water 
rocketry.' 

The  “power  pack”  strapped  to  the  back  of  the  adventurer  consists 
of  a  set  of  about  20  bottles  appearing  to  be  2  liters  in  capacity.  They  are 
partially  filled  with  water  and  then  pressurized  using  a  bicycle  pump  and 
sealed.  A  mechanism  allows  for  the  seals  on  the  bottles  to  be 
simultaneously  broken  such  that  the  water  is  ejected  and  the  adventurer  is 
launched  a  considerable  distance  in  the  direction  opposite  to  the  ejection 
direction  of  the  water. 

The  seminar  students  were  asked  to  determine  the  plausibility  of 
this  film  clip  being  real.  Using  the  laws  of  conservation  of  energy  and 
momentum,  reasonable  assumptions  regarding  the  pressure  capacity  of  a 


Summer  2006 


5 


plastic  soda  bottle,  and  basic  kinematic  equations,  they  determined  the 
maximum  distance  a  typical  person  could  be  launched.  The  adventurer  in 
the  film  clip  appears  to  fly  a  distance  of  perhaps  100  meters  during  a  flight 
that  lasts  over  6  seconds. 

The  students  performed  various  test  launches  of  water  rockets  and 
determined  that  the  maximum  thrust  delivered  to  the  rocket  occurred  when 
the  bottle  was  about  14  full  of  water.  Using  20  bottles  at  0.5  liters  of  water 
per  bottle  (and  therefore  1.5  liters  of  air)  and  a  maximum  of  10 
atmospheres  of  pressure  per  bottle,6  the  students  estimated  the  energy 
content  of  the  power  pack  at  300  liter-atmospheres  or  30,000  Joules. 
Assuming  the  mass  of  water  to  be  10  kg  and  the  mass  of  the  adventurer  to 
be  60  kg,  the  momentum  available  to  the  ejected  water  (backward)  and  the 
adventurer  (forward)  is  720  kg  m/s.  The  maximum  forward  velocity  of  the 
adventurer  at  launch,  presuming  100%  energy  conversion  efficiency  and 
the  most  optimistic  assumptions,  is  12  m/s.  This  corresponds  to  a  range  of 
about  15  meters  and  a  flight  time  of  about  1.7  seconds,  far  below  the 
apparent  flight  distance  and  time  in  the  film  clip. 

While  toy  water  rockets  fly  impressively,  it  is  their  relatively  small 
mass  as  compared  with  the  mass  of  the  water  ejected  that  leads  to  this 
phenomenon.  The  bulk  of  the  energy  in  such  a  “reverse  collision  problem” 
is  carried  away  by  the  lighter  mass.  In  the  case  where  the  payload  mass 
dominates  the  fuel  mass,  it  is  the  fuel,  not  the  payload  that  gets  most  of  the 
energy. 

Results  were  presented  to  an  audience  of  introductory  students 
who  were  studying  the  laws  of  conservation  of  momentum  and  energy  at 
the  time.  A  week  later,  several  students  commented  that  they  had  seen  an 
experimental  analysis  of  the  same  phenomenon  on  another  TV  show  on 
another  network.7  This  experimental  analysis  confirmed  the  calculations 
as  the  payload  launched  by  the  water  rocket  fizzled  immediately. 
Furthermore,  more  advanced  concepts  such  as  the  stability  of  the 
adventurer  against  torques  leading  to  wild  rotations  could  not  be  avoided. 


Conclusion 

Students  in  a  senior  seminar  were  exposed  to  non-traditional 
problems  that  crossed  the  boundaries  between  traditional  sub-disciplines 
of  physics.  Using  simple  experimentation,  ideas  from  mechanics, 
electromagnetism  and  thermodynamics,  they  evaluated  assumptions  and 


Washington  Academy  of  Sciences 


6 


rendered  opinions  on  “real  world”  possibilities.  They  presented  their 
results  formally  and  defended  their  theories  and  experimental  results. 

The  students  reported  great  satisfaction  with  the  course.  Of  even 
greater  importance,  their  abilities  to  synthesize  and  process  information 
improved  and  their  confidence  and  presentation  skills  prepared  them  for 
life  after  college. 

A  cbiow  I  edge  merits 

I  am  indebted  to  Elizabeth  Griffin  and  Paulo  Garcia  whose  data  are 
presented  here.  I  am  also  indebted  to  R.  Ferrell  Newman  whose  initial 
inquiry  sparked  the  “Tea  Cup  Controversy”  and  began  a  chain  of  events 
that  led  to  this  course  structure. 


References 

1 .  Randolph-Macon  College  curriculum  goals. 

2.  R.  Ferrell  Newman,  private  correspondence. 

3.  Model  CI-6525  Temperature  Sensor  and  Science  Workshop  750  Interface,  both 
available  from  Pasco  Scientific. 

4.  Spike  Television  Network  has  broadcast  programming  initially  called  "Most 
Extreme  Challenge"  which  lias  eventually  been  abbreviated  as  MXC. 

5 .  http://www.ast.  leeds.ac.uk/~knapp/rockets/ 

6.  ibid 

7.  "Mythbusters"  is  shown  by  The  Discovery'  Channel. 


Summer  2006 


7 


RELATIVE  SPEEDS  OF  INTERACTING 
ASTRONOMICAL  BODIES 

Carl  E.  Mungan 

U.S.  Naval  Academy.  Annapolis.  MD 


Abstract 

Simultaneous  conservation  of  linear  momentum  and  of  mechanical 
energy  can  be  used  to  calculate  the  relative  speed  of  an  isolated  pair  of 
astronomical  bodies  as  a  function  of  the  distance  separating  them.  An 
exact  treatment  is  straightforward  and  has  application  to  such 
contemporary  topics  as  the  launch  velocities  of  rockets,  and  collisions 
between  an  asteroid  and  the  Earth.  In  contrast,  when  these  topics  are 
discussed  in  introductory  physics  courses,  an  infinite-Earth-mass 
approximation  is  typically  invoked.  In  addition  to  being  unphvsical. 
this  denies  students  an  opportunity  for  a  richer  exploration  of  the 
conservation  laws  of  mechanics. 


Introduction 

Consider  two  spherically  symmetric  bodies  1  and  2  moving  through 
space  and  interacting  with  each  other  gravitationally  but  not  subject  to  any 
other  forces  (such  as  gravitational  forces  from  other  bodies  or  thrusts  from 
propulsion  systems).  This  configuration  is  depicted  in  Fig.  1.  Object  1  has 
mass  ni\  and  velocity  x>\,  while  the  second  body  has  mass  m2  and  velocity 
x>2  The  distance  between  the  centers  of  the  two  objects  is  r.  Then 
conservation  of  linear  momentum  implies  that 

+m2V2i  =  m\°\{  +  "M>2f  ’  (D 

while  conservation  of  energy  states 

1  O  1  O  Grthm^  1  o  1  o  Gm,m^ 

2  m\  wii  +  2  m2 v~2, - -  =  2  m\ "Tf  +  2  W/2  y2r  -  >  (2) 

where  the  subscripts  /  and / denote  initial  and  final  instants  in  time,  and  G 
is  the  universal  gravitational  constant. 


Summer  2006 


8 


Fig.  1  Geometry  of  two  objects  moving  under  the  influence  of  their 
mutual  gravitational  attraction.  Object  2  is  represented  as  being  larger 
than  object  1  because  we  will  think  of  2  as  being  the  Earth  and  1  as  a 
meteoroid  or  rocket.  Since  object  1  is  the  body  whose  motion  is  of 
primary  interest,  we  define  the  relative  velocity  to  specify  its  velocity 
relative  to  that  of  object  2. 


Define  the  relative  speed  v  of  the  two  objects  as  the  magnitude  of  the 
relative  velocity  vector  Then  Eqs.  (1)  and  (2)  can  be 

combined  (see  the  Appendix)  to  find 


(\  o 


Vr  =  .  V~  +  2 GjM  I - 


U 


(3) 


where  M  =  m]+  is  the  total  mass  of  the  system.  It  is  worth 

emphasizing  that  this  result  is  independent  of  the  directions  of  the  initial 
and  final  relative  velocity  vectors1,  they  need  not  be  directed  one- 
dimensionally  along  the  line  joining  the  two  bodies.  This  angle 
independence  is  akin  to  the  fact  that  we  can  use  energy  conservation  to 
predict  the  landing  speed  of  a  projectile  tossed  off  a  building  of  known 
height  with  a  known  launch  speed  regardless  of  the  launch  angle. 

Also  note  that  Eq.  (3)  can  be  generalized  to  the  motion  of  particles 
under  the  action  of  other  mutual  inverse-square  forces.  For  example,  it  can 
be  applied  to  the  electrostatic  interaction  of  two  charges  qx  and  q2  if  we 
replace  G  by  -k(qx  /  ml  )(q0  /  w0)  where  k  is  the  Coulomb  constant. 


Washington  Academy  of  Sciences 


9 


Three  applications  of  Eq.  (3) 


An  immediate  application  of  this  result  is  to  compute  the  escape  speed. 
This  is  the  minimum  initial  speed  that  enables  the  two  objects  to  climb  out 
of  each  other’s  gravitational  potential  wells,  or  in  other  words  that  causes 
their  relative  speed  to  fall  to  zero  as  they  approach  infinite  separation. 
Putting  =  0  at  rf  =  °°  implies  that  the  launch  speed  V-  =  Vesc  is 


(4) 


where  R  =  r  is  the  distance  between  the  centers  of  the  two  objects  at 
launch.  (In  the  case  of  a  terrestrial  rocket,  R  is  the  distance  of  the 
spacecraft  from  the  center  of  the  Earth  after  the  engines  have  been  shut  off 
and  the  booster  stages  ejected.  Unless  one  is  launching  off  a  high-orbit 
platform,  R  is  essentially  equal  to  Earth’s  radius  in  this  case.)  Note  that 
Eq.  (4)  differs  from  the  usual  approximate  textbook  expression2  in  that  M 
is  the  sum  of  both  masses,  rather  than  just  m2  alone.  This  difference  is  of 
negligible  consequence  when  launching  a  rocket  off  Earth,  but  can  be 
significant  in  the  case  of  two  astronomical  bodies  of  more  comparable 
mass  trying  to  escape  from  each  other  (< e.g .,  the  Moon’s  original 
breakaway  from  the  Earth,  or  the  response  of  a  pair  of  orbiting  bodies 
after  a  third  body  sweeps  past  or  collides  into  one  of  them). 

Another  important  application  of  Eq.  (3)  is  to  calculate  the  impact 
speed  of  a  meteoroid  (object  1)  striking  Earth  (object  2).  In  that  case,  the 
final  distance  is  Earth’s  radius,  =  RE  =  6380  km  .  Suppose  the 

meteoroid  is  initially  detected  when  it  is  far  from  the  Earth,  rf  In  ~  0,  and 
that  it  is  then  traveling  at  about  the  same  speed  as  the  Earth  because  of  the 
Sun’s  gravitational  pull,  vE  =  v2[  =  vE  where  Earth’s  orbital  speed  about 

the  Sun  is  vE  =(Gws  //?ES)  “  =29.8  km/s.  (Here  is  the  solar  mass 
and  RE$  is  one  astronomical  unit  or  150  million  kilometers.  This 
expression  is  derived  by  setting  the  Sun-Earth  gravitational  force 
GmsmE/RES  equal  to  the  product  of  Earth’s  mass  mE  and  centripetal 

acceleration  vE/RES.)  If  we  take  the  dot  product  of  the  expression 

u  =t)Ii  -\)oj  with  itself,  we  get  v\  =  2*^(1 -cos#)  where  #is  the  angle 
between  the  initial  directions  of  travel  of  the  meteoroid  and  the  Earth  (so 


Summer  2006 


10 


that  Ujj  v>2i  =V\\V2\  cos^  =  cos#).  Equation  (3)  now  becomes 

v( = ^I2vK]-cos6))+vL.e  (5) 

where  vcscE  =  (2GmE  / RE)  =  11.2km/s  from  Eq.  (4),  assuming  the 

meteoroid  is  small.  This  impact  speed  v$  is  plotted  in  Fig.  2  as  a  function 

of  the  angle  0.  The  results  are  in  good  agreement  with  astronomical  data 
collected  for  actual  meteoroid  arrival  speeds  at  Earth’s  upper  atmosphere.3 


angle  (degrees) 


Fig.  2  Speed  relative  to  Earth  with  which  a  meteoroid  strikes  our 
atmosphere  (assuming  the  meteoroid  is  much  smaller  in  size  and  mass 
than  the  Earth).  The  abscissa  is  the  angle  between  Earth's  orbital 
velocity  (assumed  fixed  in  direction)  and  the  meteoroid's  initial 
velocity.  Large  angles  imply  a  head-on  collision  (so  that  the  relative 
impact  speed  is  approximately  2V0.  while  small  angles  imply  that 
either  the  asteroid  strikes  Earth  from  behind  or  vice-versa  (so  that  the 
intercept  along  the  ordinate  is  i>esc  E).  as  the  inset  diagrams  suggest. 

A  third  important  application  of  Eq.  (3)  is  Solar  System  escape:  How 
should  a  rocket  be  launched  from  Earth’s  surface  so  that  it  escapes  both 
the  Earth  and  Sun?  A  solution  can  be  obtained  by  separately  considering 
the  escape  from  each  of  these  bodies.  This  is  called  the  “independent 
escape”  approximation  and  its  validity  has  been  confirmed  by  numerical 
solution  of  the  exact  three-body  problem.4  Substitute  into  Eq.  (3)  the 


Washington  Academy  of  Sciences 


values  M  ~  mE  ,  /•  =  ,  rf  =  «> ,  launch  speed  vi  =  VQSC  ss  relative  to  the 

Earth  in  order  to  escape  from  the  Solar  System,  and  final  velocity  Desc  s 
relative  to  the  Sun  [in  order  to  escape  from  it  with  speed 
v  s  =  (2G/7?s  /  ^ES)1/2  =  42. 1  km/s  ]  which  implies  a  final  speed  relative 
to  Earth  of  =  t>esc  s  “  ve  >  assuming  the  rocket  is  launched  in  the 

direction  of  Earth's  orbital  velocity  t)E  (Earth’s  axial  velocity  can  also  be 
included  if  the  rocket  is  launched  eastward  from  the  equator,  as  is  often 
done  for  deep-space  satellites.)  Rearranging,  one  thereby  obtains 

yesc,SS  =  \l(l'csc.S~Vl  J  +  VL.E  =  1 6  7  km/s  •  (6) 

which  is  only  a  little  larger  than  the  escape  speed  from  Earth  alone!  In 
particular,  this  speed  is  much  smaller  than  the  42.1  km/s  escape  speed 
from  the  Sun  starting  at  rest  relative  to  the  Sun  at  Earth’s  distance.  Taking 
advantage  of  Earth’s  motion  by  launching  in  the  direction  of  its  orbital 
velocity  confers  a  huge  assist.  (Additional  boosts  are  possible  using  the 
gravitational  slingshot  effect  as  the  spacecraft  passes  other  planets  on  its 
way  out  of  the  Solar  System.) 

It  is  important  to  note  that  Eq.  (6)  cannot  be  obtained  by  assuming  that 
the  sum  of  the  kinetic  energy  of  the  rocket  (in  Sun’s  frame  of  reference) 
and  the  potential  energy  of  the  rocket  relative  to  the  Sun  and  Earth  is 
conserved,  i.e.,  by  letting  ni  be  the  rocket’s  mass  and  writing 


\m(Vesc.SS  +  Vi:j 


Gm^m 


Gm^m  ? 

^ES 


?  7  I 

=  0  ^  Vesc,SS  =yVlsc,S  +  *4sc,E  ”  VE 


which  does  not  agree  with  Eq.  (6).  The  error  is  that  the  change  in  Earth’s 
kinetic  energy  (in  Sun’s  frame  of  reference)  has  been  neglected.  In  the 
solar  frame  the  Earth  is  moving,  and  the  rocket  is  exerting  a  gravitational 
force  on  it  in  its  direction  of  motion.  Therefore  work  is  done  on  the  Earth, 
so  that  Earth’s  kinetic  energy  must  increase.  To  put  it  another  way,  work 
(and  hence  the  change  in  kinetic  energy)  are  dependent  on  the  reference 
frame  of  the  observer.  (In  the  terrestrial  frame,  no  work  is  done  on  the 
Earth.)  It  is  only  the  sum  of  the  work  that  the  Earth  and  rocket  do  on  each 
other  that  is  frame  independent  (namely  it  equals  the  decrease  in 
gravitational  potential  energy  of  the  Earth-rocket  system),  as  can  be  seen 
from  Eq.  (13)  in  the  Appendix. 


Summer  2006 


12 


Conclusions 

In  summary,  calculation  of  the  relative  speed  between  two 
astronomical  bodies  resulting  from  their  mutual  gravitational  interaction 
(or  between  two  point  charges  interacting  electrostatically)  is  an  elegant 
and  useful  application  of  the  conservation  laws  of  energy  and  momentum. 
The  math  is  considerably  simplified  by  measuring  the  positions  and 
velocities  of  the  bodies  in  the  center-of-mass  reference  frame,  so  that  an 
exact  derivation  is  within  the  scope  of  an  introductory  physics  course.  In 
contrast,  standard  treatments  such  as  Eq.  (7)  only  consider  the  mechanical 
energy  of  a  single  body.  The  latter  approach  not  only  violates  conservation 
of  linear  momentum,  it  is  not  even  properly  defined  because  potential 
energy  is  actually  a  property  of  the  system  of  interacting  bodies  and  not  of 
one  body  alone.  That  standard  approach  only  gives  the  correct  answer, 
such  as  Eq.  (4),  when  one  body  is  much  more  massive  than  the  other  and 
the  velocities  are  measured  in  the  rest  frame  of  the  heavy  body,  as 
required  by  the  work-kinetic-energy  theorem. 

One  application  of  the  exact  result  given  by  Eq.  (3)  is  to  compute  the 
escape  speed  of  one  body  relative  to  another.  It  is  given  by  Eq.  (4) 
regardless  of  the  sizes  of  the  two  objects  (unlike  the  usual  textbook 
expression).  That  explains  why  the  formula  is  symmetric  in  the  radii  and 
masses  of  the  two  bodies.  The  escape  speed  for  object  1  to  escape  from  2 
must  be  the  same  as  for  body  2  to  escape  from  1 . 

A  second  application  is  the  calculation  of  the  speeds  of  meteoroids 
impacting  the  Earth.  Most  of  the  variation  in  speed  here  is  due  to  the  large 
range  of  angles  between  the  meteoroid’s  and  Earth’s  velocities,  as  can  be 
seen  from  Eq.  (5).  A  head-on  collision  approximately  doubles  the  impact 
speed  (ignoring  the  small  boost  due  to  Earth’s  gravity  described  by  the 
vesc£  term),  while  a  rearward  collision  almost  cancels  it,  assuming  the 
Earth  and  meteoroid  have  similar  initial  speeds  relative  to  the  Sun. 

Finally  Eq.  (6),  describing  escape  from  the  Solar  System,  depends  on 
three  separate  speeds:  the  escape  speed  from  Earth’s  surface,  the  escape 
speed  from  the  Sun  at  Earth’s  distance,  and  the  orbital  speed  of  Earth 
about  the  Sun.  The  two  escape  terms  are  added  in  quadrature  because 
kinetic  energy  depends  on  speed  squared.  Meanwhile,  the  orbital  speed  is 
subtracted  from  the  solar  escape  speed  because  Earth’s  motion  about  the 
Sun  boosts  the  rocket  toward  escape,  provided  one  launches  in  the 
direction  that  takes  advantage  of  this  assist.  In  fact.  Earth’s  orbital  speed  is 
71%  (2~1/2)  of  the  required  escape  speed  from  the  Sun,  which  explains 


Washington  Academy  of  Sciences 


13 


why  Solar  System  escape  is  actually  dominated  by  escape  from  the  Earth 

Appendix — Derivation  of  Eq.  (3) 

The  simultaneous  solution  of  Eqs.  (1)  and  (2)  is  simplified  by  the  wise 
choice  of  coordinate  system.  Since  the  two  bodies  1  and  2  are  isolated 
from  external  forces,  the  total  linear  momentum  of  the  system  is 
conserved,  and  hence  the  center  of  mass  has  constant  velocity.  We  can 
thus  choose  the  origin  to  be  fixed  at  the  center  of  mass  and  to  move  with 
it,  which  properly  defines  an  inertial  reference  frame.  In  that  case,  the  total 
linear  momentum  of  the  system  is  always  zero,  and  Eq.  (1)  implies  that 

mxX)Xx  =  -m2X) 2i  =  -m2  (6^  -X*x )  (8) 

since  X)x  is  the  initial  velocity  of  object  1  relative  to  2.  This  equation  can 
be  rearranged  to  obtain 

=  <9) 

where  Mis  the  total  mass  of  the  system.  Similar  reasoning  for  the  second 
body  gives 


l) 


2i 


Hh 

M 


(10) 


(The  minus  sign  here  reflects  the  symmetry  in  the  definition  of  the  relative 
velocity.)  Equations  (9)  and  (10)  imply  that  the  initial  kinetic  energy  of 
the  system  is 


1  2  1  2  1  2 

-mxvXl  +  -m2v2l  =  -MV{ 


(11) 


where  ju  =  mxm~,  I M  is  called  the  reduced  mass  of  the  system.  (The 

reason  for  this  name  is  that  it  is  a  quantity  with  units  of  mass  and  is 
smaller  than  both  mx  and  m2.  One  can  think  of  the  total  mass  as  the 

“series”  sum  of  the  individual  masses,  M  -  mx  +  m0,  while  the  reduced 
mass  is  the  “parallel”  sum,  1  /  n  - 1  /  mx  + 1  /  m^  .)  In  like  fashion,  the  final 
kinetic  energy  is 


Summer  2006 


14 


1  2  1  2  1  2 
wij  %  +  m2  v2i-  nv{  • 


(12) 


Equation  (2)  in  the  main  text  can  therefore  be  compactly  expressed  in 
terms  of  the  relative  speeds  and  distances  as 


2  Gmxm2  _  1  Grthm 


2^ 


=~m  - 


1"'2 


(13) 


Finally,  noting  that  mxm^  =  Mju ,  Eq.  (13)  can  be  immediately  rearranged 
to  give  Eq.  (3). 


References 

1.  In  contrast,  if  we  wished  to  determine  the  final  velocities.  u if  and  0)2  f.  and  not  merely 

the  relative  speed,  then  we  would  have  6  unknowns  (i.e..  three  components  of  each 
velocity).  Equations  (1)  and  (2)  only  provide  4  independent  relationships  (3 
components  of  linear  momentum  plus  1  scalar  energy  expression).  One  would  then 
need  to  invoke  conservation  of  angular  momentum  to  get  2  more  relations.  The 
resulting  analysis  is  no  longer  introductory  level  but  instead  invokes  non  collinear 
scattering  theory,  treated  in  texts  such  as  S.T.  Thornton  and  J.B.  Marion.  Classical 
Dynamics  of  Particles  and  Systems.  5  th  ed.  (Thomson.  Belmont  CA.  2004). 

2.  See.  for  example.  R.A.  Serway  and  J.W.  Jewett.  Jr.,  Principles  of  Physics .  4th  ed. 

(Thomson.  Belmont  CA.  2006).  p.  348. 

3.  A.  Diaz-Jimenez  and  A.P.  French.  "A  note  on  'Solar  escape  revisited  .”  Am.  J.  Phvs ., 

56.  85-86(1988). 

4.  N.J.  Hannon.  C.  Leidel.  and  J.F.  Lindner,  "Optimal  exit:  Solar  escape  as  a  restricted 

three-body  problem.”  Am.  J.  Phvs.  71.  871-877  (2003).  Also  see  A.Z.  Hendel  and 
M.J.  Longo.  ‘'Comparing  solutions  for  the  solar  escape  problem,”  Am.  J.  Phvs..  56. 
82-85  (1988). 


Washington  Academy  of  Sciences 


15 


VIDEOS  OF  EMERGENCY  CARE  SHOW  CHALLENGES 
FOR  PATIENT  SAFETY 

Colin.  F.  Mackenzie 
Yan  Xiao 

National  Study  Center  for  Trauma  &  EMS 
Program  in  Trauma  and  Department  of  Anesthesiology 
University  of  Mary  land 

Abstract 

Video  data  collection  and  analysis  is  simple  and  fruitful  and  is  currently 
underused  in  healthcare  to  understand  unsafe  acts,  pre-cursor  events 
and  system  failures  leading  to  patient  safety  issues.  Video  recorded 
examples  of  two  brief,  risky,  but  beneficial  emergency  tasks  performed 
routinely  in  emergency  care,  were  used  in  this  paper  to  illustrate,  yy  ith 
human  factors  and  ergonomic  methods,  lioyv  video  can  identity  and 
potentially  provide  solutions  for  correction  of  safety  deficiencies  in 
emergency  and  routine  clinical  care.  In  comparison  with  safety' 
recommendations  of  expert  clinicians  and  best  clinical  practice  models, 
video  recorded  performance  of  routine  and  emergency  tasks  showed 
that  rarely  were  either  expert  recommendations  or  best  practices  used 
consistently.  The  safety7  issues,  what  really  occurred,  and  potential 
solutions  to  prevent  recurrences  yvere  revealed  by  the  video  record. 

Because  of  the  fine-grained  analyses  possible,  the  video  record 
captured  pre-cursor  and  fleeting  events,  subtle  cues,  brief  utterances, 
and  unsafe  acts  leading  to  the  safety  deficiencies. 

The  Problem 

Traditional  data  collection  methodologies  have  difficulty  capturing 
fleeting  events,  subtle  cues,  brief  utterances,  or  team  interactions  and 
communications  (Rogers,  1992).  There  is  a  paucity  of  data  about  what 
occurs  in  uncertain  emergency  medicine  workplaces,  where  risky  but 
beneficial  procedures  are  carried  out,  often  in  non-optimal  circumstances. 
Such  data  may  be  critical  to  identification  of  what  Reason  (1990)  has 
termed  unsafe  acts,  pre-cursor  events,  accident  opportunities,  latent  and 
systems  failures.  This  paper  discusses  how  patient  safety  shortcomings  in 
the  emergency  medical  domain  can  be  identified  and  potentially  rectified 
through  a  video-based  data  collection,  analysis,  and  educational  feedback 
approach.  Successful  preventive  strategies  were  identified  for  patient  and 
clinician  safety  performance  problems  that  were  revealed  using  this 


Summer  2006 


- 


16 


robust,  inexpensive  video  technology  through  which  fine-grained  data 
analyses  are  possible. 

Significance 

Video  data  collection  and  analysis  are  simple,  fruitful  and  currently 
underused  to  examine  the  real-life  medical  workplace  and  understand 
what  is  really  happening  and  how  improvements  can  be  made  (Mackenzie, 
Xiao,  &  Horst  2004).  Emergency  medical  departments  have  many 
different  physical  characteristics,  personnel  work  routines,  and  team 
organization  structures;  yet  video  data  collection  is  a  methodology  that  is 
applicable  in  all  domains.  Video  analysis  provides  both  the  systems-based 
solutions  that  can  be  generalized  across  many  emergency  medical  domains 
and  unique  solutions  to  a  specific  location.  Video  data  captures  real-life 
events  that  can  be  used  to  develop  simulations  and  training  material  to 
prevent  a  recurrence  (Weinger  el  a /  2004).  This  approach  for  improving 
patient  care  outcomes  in  healthcare  used  in  a  systematic  manner  can 
identify  many  of  the  deficiencies  in  knowledge  about  pre-cursor  events, 
error  opportunities,  and  provide  solutions  for  correction  of  deficiencies. 

Methods 

Video  clips  from  the  University  of  Maryland,  Baltimore,  video 
library  and  15  year  experience  of  video  data  collection  and  analysis 
methodologies  captured  during  emergency  care  of  trauma  patients  were 
used  as  source  material.  The  challenges  in  identification  of  safety, 
organizational,  and  systems  based  problems  in  technical  work  in 
emergency  care,  were  characterized  using  human  factors  and  ergonomic 
methods.  A  multidisciplinary  approach  for  analysis  and  data  extraction 
included  experienced  trauma  clinicians,  experts  in  industrial  engineering, 
psychology,  and  applied  technology. 

As  an  example  of  the  patient  safety  data  collection  from  video¬ 
recording  in  the  trauma  workplace,  video  records  were  made  of  airway 
management  (placement  of  a  plastic  tube  into  the  trachea  -  called  tracheal 
intubation).  Misplacement  of  the  airway  is  a  major  source  of  adverse 
outcome  in  anesthesia  and  during  trauma  patient  resuscitation.  In  an 
analysis  of  2,046  closed  claims  from  medical  insurance  company  files, 
762,  or  37%,  of  such  events  were  caused  by  misplacement  of  the  tracheal 
tube  into  the  esophagus,  resulting  in  no  oxygen  delivery  to  the  lungs  and 
an  adverse  patient  outcome  (Caplan  et  al  1990).  In  a  300  patient  sub¬ 
group  of  these  patients  who  had  traumatic  injury,  the  incidence  of  brain 
damage  and  death  was  47%  (Cheney  el  al  1991).  The  misplacement  of  the 
tracheal  tube  into  the  esophagus  also  occurred  with  the  pre-hospital  use  of 


Washington  Academy  of  Sciences 


17 


this  airway  management  technique,  resulting  in  50%  mortality  (Katz  & 
Falk,  2001).  The  task  of  tracheal  intubation  is  therefore  a  risky,  but  at  the 
same  time  a  potentially  life-saving  procedure,  likely  to  be  a  fruitful  source 
of  patient  safety  issues  captured  on  video  records. 

A  second,  brief,  risky,  but  beneficial  medical  procedure  performed 
frequently  in  injured  patients  is  chest  tube  insertion.  Video  recording  was 
used  in  a  similar  manner  to  that  employed  for  tracheal  intubation  to 
evaluate  performance  of  insertion  of  the  tube  through  the  chest  wall.  This 
procedure  is  used  to  relieve  pressure  from  air  or  fluids  (such  as  blood)  that 
accumulate  after  trauma,  between  the  chest  wall  and  lung  tissue, 
collapsing  one  or  both  lungs.  The  risks  of  chest  tube  insertion  include 
damage  to  the  lungs,  heart,  diaphragm,  liver,  stomach,  and  spleen  (if  the 
tube  is  misdirected  or  inserted  too  low  or  too  far  into  the  chest).  In  trauma 
patients  there  is  a  reported  morbidity  of  6%  to  36%  of  all  chest  tube 
insertions  (Etoch  el  al.  1995).  At  our  own  institution  there  was  a  16% 
incidence  of  infection  within  the  chest  following  chest  tube  insertion 
(Caplan  et  al.  1984),  about  four  times  the  incidence  of  many  other  similar 
institutions  (Ernst  el  al.  2003).  The  suspected  mechanism  for  infection  was 
contamination  during  the  procedure  of  chest  tube  insertion.  Management 
of  such  infections  within  the  chest  requires  prolonged  hospital  stay, 
lengthy  drainage  from  an  indwelling  tube,  and  often  extensive  surgery  to 
peel  the  infection  from  the  lung.  Video  recording  of  chest  tube  insertion 
seemed  likely  to  be  able  to  identify  causes  of  possible  contamination 
during  insertion 

Results 

Tracheal  Intubation 

Among  the  first  50  video  recordings  of  tracheal  intubation,  there 
was  a  single  video  record  of  a  prolonged  undetected  esophageal 
intubation.  This  video  was  reviewed  a)  by  the  anesthesia  care  providers 
whose  care  was  video  recorded;  b)  by  subject  matter  experts  (SMEs), 
experienced  anesthesiology  clinicians;  c)  in  comparison  to  performance  of 
tracheal  intubation  and  a  consensus  airway  management  algorithm  agreed 
upon  by  20  experienced  trauma  anesthesiologists;  and,  d)  in  comparison  to 
the  other  49  video  records  of  tracheal  intubation,  performed  in  both 
elective  and  emergency  circumstances  during  patient  resuscitation  and 
routine  anesthesia  induction  in  the  Operating  Room.  Each  of  these  four 
analyses  yielded  different  aspects  of  the  pre-cursor  events,  unsafe  acts,  and 
system  failures  that  lead  to  the  error  and  identified  factors  that  allowed  the 
window  of  error  opportunity  to  occur. 


Summer  2006 


18 


Participant  Anesthesia  Care  Providers  Review  The  care  providers  were 
part  of  an  ad  hoc  team  in  which  a  medical  student,  on  his  first  clinical 
rotation  in  the  trauma  center  resuscitation  team,  incorrectly  communicated 
that  he  heard  breath  sounds  in  the  chest.  The  anesthesia  care  providers  did 
not  check  this  themselves,  and  a  second  communication  from  the  student 
(“it’s  also  going  in  here,  too”)  when  he  listened  over  the  stomach  was 
obscured  by  loud  conversation  and  laughter  from  a  nearby  area.  Video 
review  allowed  the  care  providers  to  see  their  failure  to  use  diagnostic 
equipment  to  detect  carbon  dioxide  (the  lack  of  which  would  have 
confirmed  esophageal  not  tracheal  intubation)  and  identified  the  fixation 
error  that  occurred  because  the  patient  appeared  stable  despite  the 
misplaced  airway  tube  because  he  had  been  given  oxygen  by  face  mask 
for  eight  minutes  before  attempts  at  tracheal  intubation.  Lack  of 
communications  among  the  team  occurred  when  there  was  the  greatest 
uncertainty  about  the  patient  status.  In  addition,  the  patient’s  vital  signs 
monitors  were  cycling  for  3  minutes  after  esophageal  intubation  without 
displaying  data. 

Subject  Matter  Expert  Review  The  SMEs  noted  the  reluctance  of  the 
surgical  and  nursing  team  members  to  intervene,  even  when  the  oxygen 
monitor  provided  a  signal  showing  very  low  levels  of  oxygen.  Five  to 
seven  team  members  were  standing  around  the  patient  for  the  6  minute 
duration  of  the  unrecognized  esophageal  intubation  and  did  not  directly 
offer  assistance  or  question  the  airway  management;  rather  they  made 
subtle  suggestions  (see  Table  1).  The  SMEs  also  noted  the  poor  error 
recovery  when  the  patient  was  not  re-oxygenated  before  re-attempting 
tracheal  intubation  when  the  patient  showed  signs  of  severe  lack  of 
oxygen,  even  though  a  nurse  can  be  seen  on  the  video  offering  the  needed 
face  mask.  Team  members  did  not  coordinate  the  recovery  efforts  after  the 
tube  was  removed  from  the  esophagus  by  protecting  the  airway  or  by 
assisting  the  repeat  tracheal  intubation. 

Comparison  to  Best  Practice  Algorithm  The  standard  operating  procedure 
recommended  by  the  expert  consensus  following  passage  of  a  tracheal 
tube  is  for  the  clinician  who  inserts  the  tube  to  listen,  first  to  the  left  and 
then  right  sides  of  the  chest,  and  communicate  to  the  team  whether  the 
breath  sounds  are  heard  and  whether  they  are  equal  on  both  sides  of  the 
chest.  Following  this  the  clinician  listens  over  the  stomach  and 
communicates  “no  breath  sounds  in  the  belly.”  The  last  check 
recommended  is  to  test  to  see  if  carbon  dioxide  is  present  in  the  exhaled 


Washington  Academy  of  Sciences 


19 


gas  from  the  airway  tube.  The  communication  “carbon  dioxide  positive” 
provides  the  definitive  confirmation  of  correct  placement  of  the  airway 
tube  in  the  trachea.  The  anesthesia  care  providers  never  listened  to  the 
chest  until  the  patient  showed  severe  oxygen  deprivation,  and  only  used 
the  carbon  dioxide  monitor  when  it  was  clear  from  other  data  that  the  tube 
was  misplaced.  The  inexperienced  medical  student  did  listen  to  the 
patient’s  chest  and  abdomen  in  the  recommended  sequence,  but 
misinterpreted  hearing  breath  sounds  in  the  chest,  due  to  air  entering  the 
stomach.  The  significance  of  air  entry  into  the  stomach  was  not 
recognized,  and  the  exam  was  not  repeated.  The  student  communication 
was  picked  up  by  the  directional  microphones  on  the  ceiling  above,  but 
was  not  heard  by  the  team  standing  three  feet  away  because  of  noise  from 
a  nearby  location. 

Comparison  with  Other  Video  Recorded  Tracheal  Intubations  Data  was 
extracted  from  each  video  record  of  tracheal  intubation  in  a  systematic 
manner  using  a  template  that  evaluated  the  completion  of  steps  in  the 
overall  task  of  tracheal  intubation  and  timed  the  duration  between  these 
steps  (Mackenzie  el  a/.  1996).  As  a  result,  a  fourfold  greater  time  interval 
was  noted  between  tracheal  intubation  and  testing  exhaled  gas  with  the 
carbon  dioxide  monitor  to  confirm  correct  tracheal  tube  position  in  those 
patients  intubated  under  emergency  conditions  in  the  resuscitation  area, 
rather  than  electively  in  the  Operating  Room.  The  same  procedures  were 
used  and  the  same  personnel  performed  the  intubations  in  each  location. 
However,  in  the  resuscitation  area  there  was  no  connection  to  allow 
carbon  dioxide  sampling  in  the  anesthesia  circuit  used  to  provide  oxygen. 
Insertion  of  a  25-cent  connector  was  recommended  to  allow  carbon 
dioxide  sampling  immediately  after  tracheal  intubation. 

Standard  Operating  Procedures  were  changed  as  a  result  of  these 
video  analyses  to  a)  ensure  that  the  clinical  exam  task  was  carried  out,  b) 
stress  communication  of  the  clinical  findings,  and  c)  advocate  conduct  of 
carbon  dioxide  testing  immediately  after  all  tracheal  intubations.  In  the  10 
years  since  this  prolonged,  uncorrected  esophageal  intubation  occurred, 
after  implementation  of  the  task/communication  algorithm  and  insertion  of 
the  carbon  dioxide  sampling  connector,  more  than  14,000  tracheal 
intubations  have  been  performed,  with  no  recurrence  of  undetected 
esophageal  intubation. 


Summer  2006 


20 


Table  1:  Specific  Video  Review  Findings  of  Esophageal  Intubation 


Pre-cursor  Events 

a)  Lengthy  pre-oxygenation  with  face  mask  before 
esophageal  intubation  delayed  recognition  of  lack  of 
oxygenation  after  tube  misplacement. 

b)  Ventilation  device  used  after  emergency 
intubation  had  no  simple  carbon  dioxide  analyzer 
connection  (positive  carbon  dioxide  confirms  lung, 
not  esophageal  ventilation). 

c)  Patient  physiological  monitors  of  oxygenation 
and  blood  pressure  failed  to  provide  signal  for 
nearly  3  minutes  after  esophageal  intubation. 

Fleeting  Events 

a)  Anesthesia  team  member  blows  down  tracheal 
tube  causing  “gurgling”  sounds  indicating  air  going 
down  esophagus  into  stomach. 

b)  Trauma  team  failed  to  assist  the  anesthesia  team 
for  30  seconds  when  misplaced  tube  removed. 

Subtle  Cues 

a)  Uncertainty  about  tube  misplacement  revealed  by 
comments  heard  on  audio  record  between  team 
members  “Should  you  pull  the  tube  out?”,  “He’s  got 
a  good  pulse”,  “We’re  in  there!”,  “Do  you  want  a 
new  tube?” 

Brief  Utterances 

a)  “It’s  also  going  in  here  too”  comment  by  medical 
student  listening  over  abdomen,  not  heard  by  team 
due  to  nearby  loud  conversation. 

b)  “Correlates  well  with  pulse  and  says  39  to  40” 
(nurse  commenting  on  both  the  accuracy  and  low 
oxygen  monitor  value  of  first  display  -  normal  level 
98-100.  Correlation  with  pulse  suggests  value  is 
accurate). 

Unsafe  Acts 

a)  Video  record  showed  neither  anesthesia  care 
provider  carried  a  stethoscope  to  listen  to  chest 
(standard  operating  procedure). 

Washington  Academy  of  Sciences 


21 


b)  Carbon  dioxide  analysis  (“gold  standard”  to 
detect  lung  ventilation)  was  delayed  for  5  minutes 
after  esophageal  intubation. 

c)  No  reoxygenation  (“holding  pattern”)  established 
before  re-attempt  intubation. 


Chest  Tube  Insertion 

Surgical  texts,  semi -structured  interviews  and  a  questionnaire 
completed  by  senior  trauma  surgeons  were  used  to  establish  best  practice 
techniques  for  chest  tube  insertion.  From  the  first  49  video  records  of 
chest  tube  insertion  we  extracted  more  than  80  short  (15  -  120  second) 
clips  showing  good  and  non-optimal  performance  of  chest  tube  insertion. 
These  were  copied  onto  a  compact  disc  and  distributed  among  15  senior 
trauma  clinicians  who  scored  statements  linked  to  each  video  clip  on  a 
Likert  scale  (1-10,  where  1=  strongly  agree  and  10=  strongly  disagree). 
Scores  aggregating  below  three  were  considered  to  represent  a  consensus. 
Several  analyses  were  conducted  of  the  video  records:  a)  examination  of 
breaks  in  sterile  technique  during  emergency  and  elective  insertion  of 
chest  tubes  together  with  practices  that  would  have  prevented  these 
breaks;  b)  task  analysis  template  data  extraction  by  SMEs  of  times, 
number  of  insertion  attempts,  instrument  tray  positioning,  etc.;  c) 
evaluation  of  whether  the  practices  reaching  consensus  among  the  senior 
trauma  clinicians  were  carried  out  in  each  video  recorded  chest  tube 
insertion;  and  d)  ergonomic  analysis  of  instrument  tray  position, 
instrument  tray  content  and  number  of  instrument  trays  used  for  chest  tube 
insertion  (both  unilateral  and  bilateral  chest  tube  insertions  occurred). 

Examination  of  Breaks  in  Sterile  Technique 

Among  the  26  emergency  chest  tube  insertions,  video  records 
showed  that  100%  had  breaks  in  sterile  technique.  All  but  one  of  these 
breaks  in  technique,  many  of  which  were  fleeting  events  (see  Table  2), 
occurred  within  one  minute  of  the  start  of  the  surgical  procedure  after  the 
skin  had  been  prepped  with  antibacterial  fluid.  One  surgical  site  remained 
sterile  for  three  minutes.  Among  the  24  elective  chest  tube  insertions  all 
but  one  surgical  site  was  contaminated  within  13  minutes.  In  one  patient 
the  site  remained  sterile  for  28  minutes  before  eventual  contamination.  All 
chest  tube  insertions,  whether  carried  out  in  emergency  or  elective 


Summer  2006 


22 


circumstances,  had  breaks  in  surgical  sterile  techniques.  In  emergencies 
there  were  113  breaks  in  sterile  technique  noted  by  video  analysis;  in 
elective  there  were  64  breaks  in  sterile  technique  (Mackenzie  et  al  2002). 
Practices  that  would  have  prevented  these  breaks  in  sterility  included 
wearing  of  sterile  gloves  and  gowns,  more  extensive  prepping  of  the  skin 
with  antibacterial  fluid,  wider  draping  of  the  surrounding  area  with  sterile 
drapes,  improved  operator  technique,  and  better  patient  analgesia. 


Task  Analysis  Template  Data 

There  was  a  wide  range  of  duration  for  chest  tube  insertion. 
Emergency  chest  tube  insertion  was  shorter  in  duration  than  elective  and 
required  fewer  unsuccessful  attempts.  Two  needle  sticks  and  one  knife  cut 
were  video  recorded  in  these  49  chest  tube  insertions.  Infection  and 
“sharp”  injury  risks  appeared  to  be  increased  by  sharing  of  instrument 
trays  and  simultaneous  invasive  surgical  procedures. 


Consensus  Practices 

Discrepancies  between  SMEs  recommended  practices  and 
observed  practices  seen  in  the  video  records  of  chest  tube  insertion  were 
prevalent.  Particular  discrepancy  was  in  the  use  of  maximum  barrier 
protection  to  prevent  contamination,  which  was  a  well  recognized  standard 
operating  procedure  included  in  all  surgical  best  practices.  However, 
among  the  first  25  video  records  of  emergency  chest  tube  insertion 
procedures,  these  were  frequently  omitted;  e.g .,  no  sterile  gown  (12/25), 
no  sterile  gloves  (5/25),  inadequate  sterile  drape  (18/25),  inadequate  skin 
cleansing  (12/25). 

Ergonomic  Analysis  of  Instrument  Tray  Position 
The  principle  of  keeping  the  instrument  tray  near  the  chest  tube 
insertion  surgical  site  was  routinely  violated.  Fifty-two  percent  of  tray 
positions  used  were  rated  as  sub-optimal  by  the  surgeons  themselves.  The 
most  common  position  (61%)  for  the  instrument  tray  was  directly  behind 
the  surgeon,  requiring  over  a  90  degree  turn.  Fifteen  of  the  chest-tube 
insertions  required  the  operator  to  walk  up  to  six  feet  from  the  surgical  site 
to  retrieve  instruments  (Seagull  et  al  2006).  Simple  ergonomic  problems 
impeded  performance  and  created  safety  risks  for  patients  and  operators. 


Washington  Academy  of  Sciences 


23 


Table  2:  Specific  Video  Review  Findings  of  Chest  Tube  Insertion 


Pre-Cursor  Events 

a)  No  preparatory  sterile  gown  and  gloves  worn  by 
team  when  notified  of  critical  emergency  patient 
admission  requiring  sterile  procedures. 

b)  Multiple  team  members  and  trainees  routinely 
perform  simultaneous  invasive  procedures  in 
emergency  patient  care. 

Fleeting  Events 

a)  Frame-by-frame  video  analysis  shows  elbow 
contaminating  instrument  tray. 

b)  Surgeon  wearing  sterile  gloves  grabs  patient’s 
arm  reaching  for  chest  tube  insertion  site  and  does 
not  change  contaminated  gloves. 

Subtle  Cues 

a)  Surgical  instrument  trays  often  placed  6  feet  from 
surgical  site  (Seagull  el  ci!  2006)  probably  increases 
contamination  occurrence  and  procedure  duration. 

b)  Video  revealed  non-sterile  gloves  were  difficult 
for  other  team  members  to  distinguish  from  sterile. 

Brief  Utterances 

a)  “This  won’t  take  long  and  we  will  numb  the  area 
so  you  won’t  feel  it”  comment  by  team  member 
before  chest  tube  insertion  in  anxious  patient  who  is 
seen  on  video  to  move,  reach  for  site  and  loudly 
complain  of  pain. 

Unsafe  Acts 

a)  Among  all  25  video  records  of  emergency  chest 
tube  insertion  there  was  an  omission  of  one  or  more 
measures  to  prevent  contamination  including  skin 
preparation,  adequate  sterile  draping,  sterile  gown 
and  sterile  gloves  (standard  operating  procedures). 

b)  Mentoring  of  trainees  during  chest  tube  insertion 
failed  to  follow  maximum  contamination  barrier 
precautions  (Guzzo  el  a 1 2006). 

Summer  2006 


24 


c)  Infection  and  “sharp”  injury  risks  of  sharing  of 
instrument  trays  and  multiple  simultaneous  invasive 
procedures. 


Discussion 

Video  has  the  advantage  over  observation  of  capturing  the 
minutest  and  briefest  particulars  of  human  interaction  while  retaining  the 
context  of  the  event  and  making  it  available  for  analyses  by  multiple  or 
independent  subject  matter  experts.  As  this  paper  shows,  video  recording 
in  the  medical  environment  makes  it  possible  for  clinicians  to  review  their 
own  activities  and  for  analysts  to  extract  qualitative  and  quantitative  data. 

Understanding  human  activities  in  real,  complex  environments  is 
important  (Klein  et  al  1993).  Many  significant  variables,  such  as 
expertise,  risk,  uncertainty,  and  composition  of  teams  are  often  difficult  to 
replicate  in  usual  laboratory  settings.  Studies  in  real  environments  and  in 
sophisticated  simulation  environments  with  experienced  practitioners  are 
required.  Although  indirect  data  such  as  recalled  past  incidents  can  be 
utilized  (Klein,  1989),  direct  collection  of  behavioral  data  is  needed  to 
overcome  potential  biases  in  retrospective  construction  of  past  events. 
Tools  for  collecting  behavioral  data  have  become  increasingly 
sophisticated. 

The  most  influential  among  these  new  tools  is  probably  video 
recording  (Dorwick  &  Biggs,  1983).  With  video  recording,  the  person 
who  was  recorded  can  provide  comments  on  his  or  her  covert  mental 
processes  cued  by  video  records.  Such  cognitive  approaches  to 
examination  of  real  medical  events  are  a  powerful  tool  to  examine 
performance  and  identify  patient  and  practitioners  safety  issues. 

Video  was  used  with  simulation  for  medical  education  (Cooper  et 
al.  2000)  and  in  the  analysis  of  crisis  resource  management  trauma 
assessment  training  debriefing  after  patient  simulation  (Gaba  &  DeAnda 
1998;  Lee  el  al  2003).  Video  by  its  nature  is  a  powerful  tool  for 
behavioral  researchers,  and  its  value  was  recognized  soon  after  its  initial 
consumer  availability  (Tardiff  et  al.  1978;  Dorwick  &  Biggs  1983).  The 
potential  utilities  of  video  recording  for  studying  performance  in  high  risk 
healthcare  settings  are  difficult  to  overstate. 

The  advances  in  hardware  and  software  have  made  video 
technology  a  routine  tool  for  research  in  individual  and  collaborative 


Washington  Academy  of  Sciences 


25 


performance.  An  increasing  number  of  research  projects  include  video 
recoding  as  a  key  data  collection  method.  How  this  tool  should  be 
exploited  methodologically  and  theoretically  is  thus  a  key  question  for 
researchers  (Xiao  &  Mackenzie,  2004). 

The  video  analysis  data  described  in  this  paper  shows  the 
advantages  compared  to  other  approaches  to  knowledge  acquisition  about 
safety  and  technical  work  in  emergency  medical  care.  Subject  matter 
expert  interviews  conducted  before  chest-tube  insertion  did  not  identify 
what  was  actually  occurring  in  the  real-life  event;  rather  the  experts 
described  an  ideal  version  of  what  they  hoped  would  occur  in  the  real 
event.  Examination  of  standard  operating  procedures,  even  those 
developed  by  consensus  for  the  two  tasks  of  tracheal  intubation  and  chest 
tube  insertion,  were  deficient  in  ensuring  safe  practices  for  the  patient  and 
the  clinician.  Medical  texts  provided  only  non-specific  assistance  to 
optimum  task  performance.  The  literature  and  evidence  based  best 
practices  identified  problems,  but  were  unable  to  articulate  solutions  that 
would  increase  patients’  safety  during  these  two  tasks. 

The  importance  of  omission  of  a  high  priority  task  was  confirmed 
by  the  critical  incident  that  resulted  in  prolonged  uncorrected 
misplacement  of  the  tracheal  tube  in  the  esophagus.  In  this  incident,  the 
anesthesia  care  providers  became  fixated  on  lack  of  information  about  the 
vital  signs  and  oxygen  levels.  They  failed  to  employ  simpler,  but  less 
technological  contingency  solutions,  such  as  listening  to  the  chest,  to 
identify  tracheal  tube  position.  They  also,  as  has  been  recognized  in  other 
critical  incidents,  failed  to  use  equipment  that  was  at  hand  (carbon  dioxide 
analyzer)  that  could  have  definitively  answered  their  concerns  about 
whether  the  tracheal  tube  was  correctly  placed. 

While  emergency  chest  tube  placement  was  almost  twice  as  rapid 
as  elective  chest  tube  placement,  there  were  no  steps  omitted  once  the  skin 
incision  started  the  procedure.  Rather,  the  task  omissions  occurred  before 
skin  incision  when  operator  gowning,  adequate  skin  preparation,  and 
surgical  draping  were  deficient.  Because  of  these  preparatory  deficiencies, 
the  opportunities  for  subsequent  contamination  of  the  surgical  site  due  to 
break  in  sterile  technique  was  magnified.  The  14-inch  long  flexible  plastic 
chest  tube  is  difficult  to  control  and  easily  became  contaminated 
unintentionally  on  the  operator’s  non-sterile  clothing  or  on  an  area  of  the 
patient  not  covered  with  surgical  drapes.  In  some  instances,  the  patient 
themselves  contaminated  the  surgical  site  because  the  hand  (positioned 
above  the  head  to  open  the  space  between  the  ribs)  on  the  side  of  the  chest 


Summer  2006 


26 


tube  insertion  was  not  held.  The  non-optimal  positioning  of  the  surgical 
instrument  tray  increases  the  likelihood  that  breaks  in  sterile  technique 
could  occur.  Large  workload  under  time  pressure  creates  challenges  not 
only  for  individuals  but  also  for  the  resuscitation  staff  as  a  whole  to 
coordinate  activities. 

For  both  the  studied  tasks,  video  record  review,  especially  of 
emergency  task  accomplishment,  provided  fine-grained  data  analyses  that 
identified  errors  of  omission  and  non-optimal  performance.  Even  the 
experienced  team  members  who  participated  in  the  care  that  was  video 
recorded  were  not  immune  from  these  deficiencies,  and  were  unaware  of 
their  performance  until  receiving  the  feedback  from  the  video  record. 
Aggregate  data  from  multiple  task  accomplishments  compared  at  two 
levels  of  task  urgency  was  a  non-pejorative  means  of  conveying  the  need 
for  procedural  changes  to  increase  patient  and  clinician  safety. 

End  Note:  This  paper  is  based  in  part  on  the  formal  presentation 
Videos  of  Emergency  Care  Show  Challenges  for  Patienl  Safely  by  Colin  F. 
Mackenzie,  MD.  It  was  presented  in  a  symposium  on  Human  Error  in 
Medicine  by  the  Potomac  Chapter  of  the  Human  Factors  and  Ergonomics 
Society  at  the  Washington  Academy  of  Sciences’  Capital  Science  2006 
Conference,  held  at  the  National  Science  Foundation,  Arlington,  VA 
March  25,  2006. 


Acknowledgements 

The  video  data  and  analyses  presented  here  would  not  have  been 
possible  without  the  human  factors  engineering  and  psychology  expertise 
(Jacob  Seagull  PhD),  technical  skills  (Peter  Fu-Ming  Hu),  and  clinical 
expertise  of  the  subject  matter  expert  (SME)  surgeons,  anesthesiologists 
and  nurses  in  the  Shock  Trauma  Center  at  the  University  of  Maryland 


Funding  from  Office  of  Naval  Research  (ONR),  National  Science 
Foundation  (NSF),  Agency  for  Healthcare  Research  and  Quality  (AHRQ), 
and  the  Army  Research  Institute  (ARI)  for  the  Behavioral  and  Social 
Sciences 


Washington  Academy  of  Sciences 


27 


References 

Caplan,  E.S..  Hoyt  NX  Rodriguez,  A..  Cowley,  R.A.  (1984)  Empyema  occurring  in  the 
multiply  traumatized  patient.  J.  Trauma ,  24,785-789. 

Caplan,  R.A..  Posher.  K.L.,  Ward,  R.J..  Cheney.  F.W.  (1990)  Adverse  respiratory  events 
in  anesthesia;  a  closed  claims  analysis.  Anesthesiology,  72.  828-833. 

Cheney.  F.W..  Posher,  K.L.,  &  Caplan.  R.A.  (1991)  Adverse  respiratory  events 
infrequently  leading  to  malpractice  suits.  A  closed  claims  analysis. 
Anesthesiology,  75,  932-939. 

Cooper.  J.B..  Barron,  D.,  Blum.  R..  Davison.  J.K..  Feinstein,  D..  Halacsz.  J..  Raemer.  D.. 

Russell,  R.  (2002)  Video  teleconferencing  with  a  realistic  simulation  for  medical 
education.  J.  Clinical  Anesthesiology,  12,  256-261. 

Dorvvick.  P.W.  &  Biggs.  S.J.  ( 1983)  Using  video  psychological  and  social  applications. 
Chichester.  UK;  Wiley  &  Sons. 

Ernest,  A.,  Silvesteri.  G.A.,  &  Johnstone,  D.  (2003)  Interventional  pulmonary 

procedures:  guidelines  from  the  American  College  of  Chest  Physicians.  Chest, 
123,  1693-1717. 

Etoch,  S.W.,  Bar-Natan,  M.F.,  Miller.  F.B.,  Richardson,  J.D.  (1995)  Tube  Thoracostomy 
factors  related  to  complications.^  rch  ives  of  Surgery.  130.  521-525. 

Gaba.  D  M.  &  DeAnda.  A.  (1988)  A  comprehensiv  e  anesthesia  simulation  environment: 
Recreating  the  operating  room  for  research  and  training.  Anesthesiology,  69. 
387-394. 

Guzzo.  J.L.,  Seagull.  F.J..  Bochicchio.  G.V..  Sisley.  A..  Mackenzie.  C.F.,  Dutton.  R.P.. 
Scalea.  T..  &  Xiao.  Y.  (2006)  Mentors  Decrease  Compliance  with  Best  Sterile 
Practices  during  Central  Line  Placement  in  the  Trauma  Resuscitation  Unit. 
Surgical  In  fection,  7.  15-20. 

Katz.  S.M..  &  Falk.  J.L.  (2001)  Misplaces  endotracheal  tubes  by  paramedics  in  an  urban 
medical  services  system.  Annals  of  Emergency  Medicine,  37,  32-37. 

Klein  GA  (1989)  Recognition-Primed  Decisions.  In  WB  Rouse  (Ed)  Advances  in  Man- 
Machine  System  Research .  5.  pp  47-92  Greenwich  CT.  JAI  Press. 

Klein  GA.  Zsambok  CE.  Thorsden  ML.  (1993)  Team-Decision  Training:  Five  myths  and 
a  Model.  Military?  Review  pp  36-42. 

Lee.  S.K.,  Pardo.  M..  Gaba.  D.,  Sowb,  Y.,  Dicker,  R.,  Straus,  E.M.  Khaw.  L..  Matabito. 
Krumpel  T.M..  &  Knudson  M.M.  (2003)  Trauma  assessment  training  with  a 
patient  simulator:  A  prospective  randomized  study.  J.  Trauma,  55.  651-657. 
Mackenzie.  C.F..  Jeffries.  N.J..  Hunter,  A..  Bernard.  W..  Xiao.  Y..  &  LOTAS  Group. 

(1996)  Comparison  of  Self  Reporting  of  Deficiencies  in  Airway  Management 
with  Video  Analysis  of  Actual  Performance.  Human  Factors,  38.  623-635. 
Mackenzie.  C.F..  Martin,  P.,  Xiao.  Y.  &  LOTAS  Group.  (1996)  Video  analysis  of 

prolonged  uncorrected  esophageal  intubation.  Anesthesiology,  84.  1494-1503. 
Mackenzie,  C.F..  Xiao,  Y..  Hu,  P„  Seagull.  F.J.,  Hammond.  C..  Bochicchio.  G..  Chiu. 

W..  O’Connor.  J.,  Gerber-Smith.  L..  Dutton.  R.  (2002)  Video  clips  as  a  data 
source  for  safety  performance.  In:  Proceedings  of  the  Human  Factors  & 
Ergonomics  Society.  46th  annual  meeting,  p  1414. 

Mackenzie.  C.F.,  &  Xiao.  Y.  (2003)  Video  techniques  and  data  compared  with 

observ  ation  in  emergency  care.  Quality  &  Safety  in  Health  Care,  121  (Suppl  1 1 ) 
ii  51-ii  57. 


Summer  2006 


28 


Mackenzie.  C.F..  Xiao.  Y..  &  Horst.  R.  (2004)  Video  task  analysis  in  high  performance 
teams.  Cognition,  Technology  &  Work,  6.  139-147. 

Reason.  J.  (1990)  Human  Error .  Cambridge,  England:  Cambridge  University  Press. 

Rogers.  Y.  (1992)  Coordinating  computer-mediated  work.  Computer  Supported 
Cooperative  Work .  1.  295-315. 

Seagull.  F.J..  Mackenzie.  C.F.,  Xiao.  Y..  &  Bochiccio.  G.V.  (2006)  Video-based 

Ergonomic  Analysis  to  Evaluate  Thoracostomy  Tube  Placement  Techniques.  J. 
Trauma.  60.  227-232. 

Tardiff.  K..  et  a/.  (1981)  A  videotape  technique  for  measuring  clinical  skills.  Three  years 
of  experience.  J  Medical  Education.  56.  187-191. 

Vincente.  K.J..  (2000)  Work  Domain  Analysis  and  Task  Analysis.  A  Difference  that 
matters.  In:  Schraagen  JM.  Chipman  SF.  &  Shalin  VL.  Eds.  Cognitive  Task 
Analysis .  p:  101-118.  Mahwah  NJ:  Lawrence  Erlbaum  Associates. 

Weinger.  M.B..  Gonzales.  D  C..  Slagel.  J.  &  Syeed.  M.  (2004)  Video  capture  of  clinical 
care  to  enhance  patient  safety.  Quality  &  Safety  in  Healthcare  13.  136-144 

Xiao.  Y.  &  the  LOTAS  Group.  (2001)  Understanding  Coordination  in  a  Dynamic 

Medical  Environment:  Methods  and  Results.  In:  New  Trends  in  Cooperative 
Activities.  Eds.  McNeese  M.  Salas  E.  &  Endslev  M.  Chap.  14.  pp  242-258. 
Santa  Monica,  CA:  Human  Factors  and  Ergonomics  Society  . 

Xiao.  Y..  &  Mackenzie.  C.F.  (2004)  Introduction  to  the  special  issue  on  video-based 
research  in  high  risk  settings:  Methodology  and  experience.  Cognition, 
Technology  &  Work .  6.  127-130. 


Washington  Academy  of  Sciences 


29 


It's  Not  Who  in  98,000  Medical  Error  Deaths,  It's  What! 

Marilyn  Sue  Bogner 

Institute  for  the  Study  of  Human  Error,  LLC 

msbo  gnerifoerols .  com 


Abstract 

In  1999  the  Institute  of  Medicine  (IOM)  reported  that  44,000  to  98,000 
hospitalized  patients  die  annually  due  to  medical  error  (Kolm. 
Corrigan  &  Donaldson.  1999).  This  caused  public  consternation. 
Following  recommendations  from  the  report  that  care  providers  be 
held  accountable  for  their  errors  and  research  should  be  focused 
primarily  on  accountability  through  error-reporting  programs,  the 
report  continued  that  the  purpose  of  such  research  was  to  reduce  the 
incidence  of  error  by  50%  in  5  years.  The  findings  from  that  $250 
million  of  U.S  government  funded  research  provided  little  if  any 
indication  of  how  the  magnitude  of  error  might  be  reduced  effectively. 
The  ensuing  background  material  may  seem  tedious  and  theoretical  for 
a  problem  that  needs  urgent  and  effective  action.  The  detail  is 
important  because  it  supports  a  paradigm  change  from  the  person,  the 
care  provider  -  the  “who”  -  being  solely  responsible  for  the 
unexpected  adverse  outcome  associated  with  an  error  to  a  paradigm 
that  errors  and  attendant  adverse  outcomes  are  the  result  of  the  systems 
of  environmental  factors  affecting  the  individual  -  the  “what”.  The 
power  of  this  systems  paradigm  for  addressing  the  interplay  of  factors 
that  induce  error  is  illustrated  by  the  discussion  of  a  case  with  an 
adverse  outcome.  The  implications  of  this  systems  paradigm  to 
effectively  reduce  health  care  error  by  considering  the  role  of  the 
“what”  as  well  as  the  “who”  are  discussed. 


The  Problem 

In  health  care  as  in  other  industries  when  an  incident  occurs  in 
which  an  act  of  one  person  results  in  harm  to  another,  directly  as  in  health 
care  or  indirectly  as  in  aviation,  that  act  typically  is  considered  an  error. 
This  attribution  of  error  is  evident  not  only  in  the  media,  but  also  in 
conversations  about  the  incident  especially  if  that  incident  involves  health 
care.  Indeed,  health  care  providers  often  blame  themselves  for  an  adverse 


Summer  2006 


30 


outcome  because  they  are  taught  they  are  responsible  for  what  happens  to 
the  patient  Blaming  the  person  associated  with  the  act  that  caused  actual 
harm  or  an  inappropriate  act  in  which  no  harm  occurred  reflects  the  Stop 
Rule  -  the  predilection  when  reviewing  or  backward-chaining  the 
conditions  to  identify  what  caused  an  error  to  stop  at  a  familiar  possible 
cause,  one  that  can  be  readily  addressed  (Rasmussen,  1990).  What  more 
familiar  and  more  easily  addressed  cause  of  an  incident  than  the  individual 
associated  with  the  act?  The  extent  to  which  the  presumption  that  the  care 
provider  is  the  source  of  medical  error  is  pervasive  in  health  care  and  its 
literature  on  error  is  evident  in  the  title  of  the  IOM  report  To  err  is  human 
(Kohn,  Corrigan  &  Donaldson,  1999)  and  in  its  primary  recommendation 
to  determine  health  care  provider  accountability. 

The  IOM  report  states  that  one  way  to  learn  from  errors  is  from  a 
reporting  program  and  that  such  programs  can  serve  two  functions  -  they 
can  “  .  .  .  hold  providers  accountable  for  performance,  or  alternatively  .  .  . 
provide  information  that  can  lead  to  improved  safety”  (Kohn,  Corrigan  & 
Donaldson,  1999,  p.  74).  According  to  the  report  those  two  functions  are 
not  incompatible  but  can  be  difficult  to  satisfy  simultaneously.  In  light  of 
that,  the  report  recommended  that  an  error-reporting  program  be 
developed  that  focuses  on  the  former  function,  that  of  provider 
accountability,  that  such  error  reporting  be  mandatory  and  that  a  national 
database  be  developed  from  the  error  reports.  The  IOM  report  also 
recommended  that  the  latter  function  of  an  error-reporting  activity  - 
providing  information  that  can  lead  to  improved  safety  —  is  in  the  domain 
of  voluntary  reporting.  The  error  reporting  for  accountability  reflects  the 
prevailing  definitions  of  error. 

Error  is  defined  in  terms  of  the  point  in  process  of  care  an  incident 
occurred  such  as  errors  of  missed  diagnosis,  mistakes  during  treatment, 
medication  mistakes,  inadequate  postoperative  care,  and  mistaken  identity 
(Gibson  &  Singh,  2003).  Definitions  of  errors  have  been  differentiated 
into  technical  errors  reflecting  skill  failures,  judgmental  errors  that  involve 
the  selection  of  an  incorrect  strategy  of  treatment,  and  normative  errors 
which  occur  when  the  larger  social  values  embedded  within  medicine  as  a 
profession  are  violated  (Bosk,  1979).  The  focus  of  most  of  the  IOM  report, 
that  of  provider  accountability,  is  in  keeping  with  those  definitions. 
Indeed,  the  recommendation  for  provider  accountability  perpetuates  the 
presumption  that  the  person  is  the  sole  cause  of  an  error  and  by  collecting 
data  only  on  the  provider  perpetuates  blaming  that  person.  This  illustrates 
the  potency  of  the  Stop  Rule  triggering  the  attribution  of  the  cause  of  an 
error  to  the  easiest  explained:  the  care  provider  caused  the  error  because 


Washington  Academy  of  Sciences 


31 


he  or  she  performed  the  act  led  to  an  unexpected  adverse  outcome. 
Because  the  explanation  stopped  with  the  provider,  the  act  is  considered  as 
solely  of  his  or  her  own  volition,  so  reporting  errors  is  of  the  form  of  “who 
did  what”  -  for  example.  Dr.  Surgeon  lacerated  Mrs  Patient’s  liver. 

To  address  the  chilling  statistics  that  44,000  to  98,000  hospitalized 
patients  die  annually  because  of  error  in  their  health  care,  the  IOM  report 
stated  that  error  related  research  and  related  efforts  were  to  focus  on 
provider  accountability  and  that  the  results  of  those  efforts  were  to  reduce 
the  incidence  of  error  by  50%  in  5  years.  Congress  appropriated  $50 
million  per  year  for  those  5  years  to  meet  that  goal.  At  the  conclusion  of 
the  5  years,  November  2004,  a  conference  was  held  to  determine  the 
extent  to  which  the  $250  million  in  research  funding  approached  the  goal 
of  50%  reduction  in  errors.  In  considering  the  findings  from  that  research 
it  was  generally  concluded  that  efforts  to  attain  the  50%  reduction  in  error 
not  only  failed  to  meet  that  goal  but  the  impact  of  those  efforts  on  error  is 
negligible  (Commonwealth  Fund,  2004)  and  that  “  .  .  .  little  data  exist 
showing  progress  and  researchers  are  still  debating  not  how  to  save  lives, 
but  what  to  measure”  (Zwillich,  2004).  Indeed  one  presenter  stated  that 
many  states  and  private  health  systems  require  health  workers  to  report 
medical  errors  or  near  misses  in  which  a  patient  is  put  at  potential  risk,  but 
researchers  still  have  not  figured  out  what  to  do  with  the  reports  once  they 
have  them  (Wachter,  2004). 

Despite  the  lack  of  empirical  support  for  error  reporting  for 
provider  accountability  as  a  means  of  reducing  error  that  approach  to  the 
problem  persists.  Rather  than  continuing  work  on  a  non-productive 
approach,  an  alternative  should  be  pursued.  The  alternative  approach  is  to 
consider  error  for  what  it  is.  The  previously  stated  definitions  of  error 
describe  an  error,  but  do  not  define  the  process  by  which  an  error  occurs  - 
that  process  is  an  action,  a  behavior.  Behavior  has  been  documented  by 
centuries  of  research  and  theory  in  psychology  and  the  social  sciences  as 
well  as  the  physical  sciences  and  millennia  of  philosophical  thought  as  the 
interaction  of  an  entity  -  for  the  purpose  of  this  discussion  a  person  -  with 
factors  in  the  environment.  In  light  of  this  evidence-based  research, 
addressing  only  the  individual  when  considering  an  error  is  misleading 
and  inaccurate. 


Summer  2006 


32 


Error  as  Behavior 

Considering  error  as  behavior  provides  an  action  oriented  approach 
to  addressing  error  by  identifying  those  factors  in  the  environment  that  by 
affecting  the  person  induce  error.  It  should  be  noted  that  it  is  only  those 
internal  and  external  factors  that  affect  the  individual  at  the  time  of 
performing  the  task  either  directly  or  indirectly  as  the  influence  of  past 
experience  or  anticipation  of  future  actions,  are  to  be  considered  when 
addressing  that  person’s  behavior  (Lewin,  1946/1964).  Often  in  discussing 
health  care  error  reference  is  made  to  the  system,  which  is  the  health  care 
system.  This  is  not  appropriate  when  considering  an  error  associated  with 
an  individual  because  all  aspects  of  that  system  do  not  affect  that  person 
(Bogner,  2004a).  To  even  consider  them  is  to  confuse  the  issue  to  the 
point  of  considering  an  un-analyzable  situation.  The  question  then 
emerges  as  to  how  to  identify  those  factors  in  the  complex  environment  in 
which  health  care  is  provided.  Lessons  learned  through  error  research  by 
other  industries  afford  a  viable  means  to  address  that  issue. 

Error  research  in  manufacturing  and  nuclear  power  (Moray,  1994, 
Senders  &  Moray,  1991,  Rasmussen,  1982)  identifies  categories  of 
interacting  factors  or  systems  of  factors  that  affect  the  person  performing  a 
task.  Those  systems  and  factors  in  terms  of  health  care  are:  the  patient  (the 
focus  of  the  task)  weight,  co-morbidity,  name;  means  of  providing  care 
(tools  for  performing  the  task)  medications,  medical  devices;  the  care 
provider  (the  person  performing  the  task)  stamina,  physical  characteristics, 
fatigue.  Those  systems  interact  in  the  context  of  five  systems  of 
environmental  factors  of:  ambient  conditions  of  illumination,  temperature, 
noise,  altitude;  the  physical  environment  with  placement  of  medical 
equipment,  room  size,  clutter;  the  social  environment  of  other  care 
providers  and  personnel,  family  members,  professional  culture; 
organizational  factors  such  as  workload,  hours  worked,  reports,  policies 
for  caring  for  uninsured  persons,  organizational  culture;  and  legal- 
regulatory-reimbursement-national  culture  factors  that  include  threat  of 
litigation,  regulatory  constraints,  reimbursement  policies,  and  national 
cultural  mores.  These  interacting  systems  as  represented  in  Figure  1 
(Bogner,  2002)  can  function  in  a  hierarchal  manner  and  often  in  a  reverse 
ripple  effect,  that  is,  impact  on  the  more  super-ordinate  system  of  factors 
impacts  those  systems  below  that  system  or  in  Figure  1  all  those  systems 
within  the  circle  impacted. 


Washington  Academy  of  Sciences 


33 


Systems  of  Influence  for 
the  Care  Provider 


Figure  1 


In  keeping  with  models  of  error  discussed  in  terms  of  food  -  Swiss 
cheese  (Reason,  1991),  an  onion  (Moray,  1994)  -  this  systems  behavior 
model  of  error  is  likened  to  an  artichoke  with  the  care  provider  at  the 
center  as  the  heart  of  the  artichoke  represented  in  Figure  2.  The  influence 
of  the  systems  of  factors  is  represented  by  the  encircling  leaves  of  the 
Artichoke  -  when  the  affect  of  those  systems  of  factors  becomes  great,  the 
provider  can  be  Artichoked  into  a  behavior  -  an  error. 


Application  of  the  Behavior  Systems  Approach 

The  value  of  the  Artichoke  systems  approach  (Bogner,  2006)  is 
illustrated  by  the  case  of  the  previously  mentioned  adverse  incident  in 
which  Dr.  Surgeon  lacerated  Mrs.  Patient’s  liver.  Conforming  to  the 
provider  accountability  requirement  of  reporting  an  adverse  outcome.  Dr. 
Surgeon  reported  the  incident.  The  typical  response  to  this  is  that  Dr. 
Surgeon  would  be  reprimanded  in  a  Mortality  and  Morbidity  session  in 
which  the  staff  discusses  cases  and  could  be  sued  for  negligence.  The 
impact  of  that  would  be  a  blow  to  Dr.  Surgeon’s  self-esteem  and 
professional  pride,  a  possible  increase  in  his  malpractice  insurance  rate  all 


Summer  2006 


34 


of  which  could  lead  to  Dr.  Surgeon  seriously  considering  leaving  the 
profession.  Alternatively,  Dr.  Surgeon  recorded  several  factors  that 
affected  him  in  the  course  of  the  procedure  on  an  incident  worksheet  based 
on  the  Artichoke  systems  approach  (Bogner,  2000).  The  names  of  each  of 
the  eight  systems  of  factors  in  the  Artichoke  model  are  listed  as  a  column 
down  the  left  side  of  the  worksheet  with  a  line  for  the  person  reporting  the 
incident  to  record  the  factors  in  the  specific  system  that  affected  him  or 
her. 


Provider  in  Context 


Figure  2 

The  names  of  each  of  the  systems  on  the  worksheet  serve  as 
memory  aides  for  the  provider  in  identifying  the  factors  that  affected  him 
or  her  and  contributed  to  the  incident.  In  Dr.  Surgeon’s  case,  he  noted  that: 
the  patient  was  morbidly  obese,  the  means  of  providing  care  was  a 
laparoscopic  also  known  as  keyhole  surgical  procedure  in  which  the 
surgeon  manipulates  instruments  with  long  shafts  viewing  the  surgical  site 
via  a  small  video  camera  inserted  into  the  site  as  illustrated  in  Figure  3. 
Dr.  Surgeon  noted  that  he  (the  care  provider)  was  short;  in  the  physical 
environment  the  operating  table  could  not  be  lowered  sufficiently  for  the 
mass  of  Mrs.  Patient’s  body  to  be  of  optimal  height  to  manipulate  the 
instruments  so  it  was  necessary  for  Dr.  Surgeon  to  stand  on  a  stool  and 
have  the  foot  peddle  that  operates  a  certain  instrument  also  placed  on  the 
stool.  The  incident  occurred  when  the  foot  pedal  fell  off  the  stool  as  Dr. 
Surgeon  reached  his  foot  to  operate  it.  This  caused  him  to  be  off  balance; 


Washington  Academy  of  Sciences 


35 


as  Dr.  Surgeon’s  body  instinctively  lurched  to  avoid  falling  off  the  stool, 
an  instrument  he  was  holding  moved  and  lacerated  Mr.  Patient’s  liver. 


Figure  3 


The  information  provided  by  the  incident  worksheet  can  be  used  to 
reduce  the  likelihood  of  such  an  incident  occurring  again.  It  can  be 
forwarded  to  a  designated  person  who:  notifies  manufacturers  of  the  need 
for  lower  operating  tables  and  more  usable  laparoscopic  instruments, 
informs  the  hospital  purchasing  agent  of  the  importance  of  considering  the 
body  mass  of  obese  patients  when  acquiring  new  equipment  particularly 
tables  and  table-like  items  such  as  gurneys  and  beds,  and  contacts 
engineering  to  install  a  means  to  secure  foot  pedals  on  stools  not  only  in 
the  operating  room  (OR)  in  which  the  incident  occurred,  but  in  all  ORs  in 
the  facility. 

The  comparison  of  the  two  approaches  underscores  the  value  of  the 
Artichoke  systems  approach.  Information  from  the  typical  error-reporting 
for  provider  accountability  -  the  “who  did  what”  approach  -addresses 
only  the  provider,  so  the  error-inducing  conditions  of  the  stress  and  fatigue 
from  working  in  an  awkward  position  with  attendant  muscle  fatigue,  and 
the  foot  pedal  falling  from  stool  continue.  With  the  Artichoke  approach. 


Summer  2006 


36 


the  error  inducing  conditions  are  identified  and  addressed  and  the  impact 
on  patient  safety  can  be  evaluated.  Thus,  the  Artichoke  systems  approach 
provides  viable  information  that  can  lead  to  improved  safety  by  reducing 
the  likelihood  of  errors.  This  approach  also  has  implications  for  preventing 
error  through  designing  the  means  of  providing  care. 


Design  Implications 


The  relationship  between  product  design  including  labeling  and 
information  presentation  and  task  performance  is  the  focus  of  the 
discipline  of  human  factors  or  ergonomics.  Since  the  inception  of  the 
discipline,  which  typically  is  considered  as  during  World  War  2,  it  has 
guided  aspects  of  the  design  of  weapon  systems,  airplanes,  and  a  variety  of 
consumer  goods;  however,  its  application  in  health  care  has  been  limited. 
A  notable  exception  is  the  study  of  medication  errors  published  in  1960 
(Safren  &  Chapanis,  1960a,  1960b)  the  findings  of  which  are  analogous  to 
those  of  the  Harvard  Medical  Practice  Study  (Leape  et  al,  1991)  -  the 
latter  findings  to  a  large  extent  served  as  the  basis  for  the 
recommendations  of  the  IOM  study  (Kohn,  Corrigan  &  Donaldson,  1999). 
This  lack  of  applications  of  human  factors  and  ergonomics  considerations 
in  health  care  could  reflect  the  focus  of  the  provider  as  sole  source  of 
error;  if  only  the  individual  is  responsible  for  the  error,  there  is  no  need  to 
address  the  design  of  the  equipment  and  other  aspects  of  the  context  in 
which  health  care  is  provided. 

The  Artichoke  systems  approach  by  identifying  error-inducing 
contextual  factors  in  each  of  the  eight  systems  of  the  Artichoke  expands 
the  focus  from  the  provider,  the  “who”,  to  contextual  factors,  the  “what”. 
Given  that  perspective,  human  factors  and  ergonomics  considerations  can 
be  applied  to  the  interaction  of  those  factors  so  the  context  might  be 
designed  to  positively  affect  provider  performance.  Thus,  this  Artichoke 
systems  approach,  which  is  practical,  problem-solving,  action  oriented, 
and  evidence  based  not  only  can  reduce  error  through  design  by  the 
application  of  human  factors  and  ergonomics  considerations,  it  also  can 
increase  reliability  in  device  use  -  the  fewer  errors,  the  more  reliable  the 
performance.  This  approach  also  provides  a  counter-argument  to  a  typical 
industry  response  to  error  involving  a  medical  device  that  attributes  the 
cause  to  the  user  -  if  the  device  had  been  used  as  intended,  the  error  would 
not  have  occurred.  There  are  conditions  such  as  the  human  constraints  of 
the  provider  that  challenge  the  “use  as  intended”  admonition.  An  example 
is  a  left  handed  anesthesiologist  writing  the  legally  mandated  log  of  the 


Washington  Academy  of  Sciences 


37 


case  while  squeezing  the  bag  ventilating  the  patient  as  illustrated  in  Figure 
4. 


Paradigm  Change  to  the  Behavior  Approach  to  Error 

By  focusing  on  the  perspective  of  the  provider  heart  of  the 
Artichoke  and  considering  health  care  as  behavior,  the  care  provider  is  a 
collaborator  in  patient  safety  rather  than  an  adversary  and  target  for  blame 
as  in  error  reporting  for  provider  accountability.  Rather  than  the  care 
providers  conforming  to  an  inappropriately  designed  device,  human 
factors  and  ergonomics  considerations  can  be  applied  to  information 
gleaned  from  applying  the  Artichoke  systems  approach  incident  worksheet 
to  conditions  that  are  considered  as  hazardous  or  accidents  waiting  to 
happen.  This  identifies  those  contextual  factors  to  be  addressed  so  that  the 
device  might  be  designed  for  use  by  the  range  of  users  in  worst-case 
context  including  lay  persons  providing  home  care. 


Figure  4 


Summer  2006 


38 


Thus,  the  design  of  devices  conforms  to  the  care  provider  in 
context  of  use  and  as  such  enhances  the  performance  of  the  provider. 

This  approach  by  identifying  error  inducing  factors  so  they  may  be 
addressed  and  changed  to  be  neutral  if  not  performance  enhancing  affords 
the  means  to  change  the  script  of  health  care  provider  performance.  For  as 
a  script  directs  the  performance  of  an  actor  whoever  may  be  in  the  role  of 
the  script,  so  do  the  contextual  factors  determine  the  performance  of  a 
health  care  provider  whoever  he  or  she  may  be  (Bogner,  2004b).  Thus, 
applying  the  Artichoke  and  changing  the  error-inducing  factors  affects  not 
only  the  provider  involved  in  the  incident,  change  affects  all  providers 
encountering  those  contextual  factors  thus  enhancing  patient  safety. 

To  effectively  reduce  the  incidence  of  error,  it  is  time,  indeed  past 
time,  to  change  the  paradigm  for  addressing  health  care  error  from  solely 
considering  the  “who”  to  a  paradigm  that  considers  the  “what”  is  involved 
so  why  an  error  occurs  can  be  determined  and  resolved,  as  represented  by 
the  Artichoke  systems  approach. 

References 

Bogner.  M.  S.  (2000)  A  systems  approach  to  medical  error.  In  C.  Vincent  &  B.  DeMol 
(Eds.),  Safety  in  medicine  (pp.  83-100).  Amsterdam:  Pergamon. 

Bogner.  M.  S.  (2002)  Stretching  the  search  for  the  ‘why”  of  error:  The  systems  approach. 

Journal  of  Clinical  Engineering,  27.  110-115. 

Bogner,  M.  S.  (2004a)  Understanding  human  error.  In  M.  S.  Bogner  (Ed.).  Misadventures 
in  health  care:  Inside  stories  (pp.  41-58).  Mahwah.  NJ:  Lawrence  Erlbaum 
Associates.  Inc. 

Bogner,  M.  S.  (2004b)  All  the  men  and  women  merely  players.  In  M  S.  Bogner  (Ed  ). 
Misach’entures  in  health  care:  Inside  stories  (pp  165-182).  Mahwah.  NJ: 
Lawrence  Erlbaum  Associates,  Inc. 

Bogner.  M.S.  (2006)  Prevention  of  medical  errors.  In  W.S.  Marras  &  W.  Karwowski 
(Eds.),  The  occupational  ergonomics  handbook,  2nd  Ed.  :  Interventions, 
controls,  and  applications  in  occupational  ergonomics  (Chapter  47,  pp.  1  -  15). 
London:  CRC  Press  Taylor  &  Francis. 

Bosk.  C.  (1979)  Forgive  and  remember:  Managing  medical  failures.  Chicago: 
University  of  Chicago  Press. 

Commonwealth  Fund  (2004)  The  end  of  the  beginning:  Patient  safety  five  years  after  To 
err  is  human.  Retrieved  July  4,  2006,  from  http ://www . cmwf.org/publications 
Newsletter  Quality  Matters:  November  Update  from  The  Commonwealth  Fund. 

Gibson.  R.  &  Singh.  J.P.  (2003)  Wall  of  silence:  The  untold  story  of  the  medical 

mistakes  that  kill  and  injure  millions  of  Americans.  Washington.  D.C.:  Lifeline 
Press. 

Kohn,  L.T..  Corrigan,  J.M..  &  Donaldson,  M.S.  (Eds.),  (1999)  To  Err  is  Human:  Building 


Washington  Academy  of  Sciences 


39 


a  Safer  Health  System.  Washington.  D.C.:  National  Academy  Press. 

Leape.  L.  L.,  Brennan.  T.  A.,  Laird.  N..  Lawthers.  A.  G..  Localio.  A.  R..  Barnes.  B.  A.,  et 
al.  (1991)  The  nature  of  adverse  events  in  hospitalized  patients.  New  England 
Journal  of  Medicine.  324,  377-384. 

Lewin.  K.  (1964)  Behavior  and  development  as  a  function  of  the  total  situation.  In  D. 

Cartwright  (Ed.),  Field  theory  in  social  science  (238-303).  New  York:  Harper  & 
Row.  (Original  work  published  1946) 

Moray.  N.  (1994)  Error  reduction  as  a  systems  problem.  In  M.  S.  Bogner  (Ed  ).  Human 
error  in  medicine  (pp.  67-92).  Hillsdale.NJ:  Lawrence  Erlbaum  Associates.  Inc. 

Rasmussen.  J.  (1982)  Human  Errors:  A  taxonomy  for  describing  human  malfunction  in 
industrial  installations.  Journal  of  Occupational  Accidents.  4.  3 11-333. 

Rasmussen.  J.  (1990)  Human  error  and  the  problem  of  causality  in  analysis  of  accidents. 
Philosophical  Transactions  of  the  Royal  Society  of  London.  337,  449-462. 

Reason.  J.  (1990)  Human  error.  New  York:  Cambridge  University  Press. 

Safren.  M.A.  &  Chapanis.  A.  (1960  a)  A  critical  incident  study  of  hospital  medication 
errors  —  part  L  Hospitals,  JA.HA.  34.  32-66  (May  1). 

Safren,  M.A.  &  Chapanis.  A.  (1960b)  A  critical  incident  study  of  hospital  medication 
errors  —  part  2,  Hospitals,  JA.HA.  34.  54-68.  (May  16). 

Senders.  J.W.  &  Moray.  N.P.  (1991)  Human  Error:  Cause,  Prediction,  and  Reduction. 
Mahwah.  NJ:  Lawrence  Erlbaum  Associates,  Inc. 

Wachter.  R.  (2004)  Analysis  in  Health  Affairs  says  health  system  has  made  insufficient 
progress  since  IOM  issued  landmark  medical  errors  report  in  1999.  Retrieved 
July  4.  2006,  from  http ://ww w . healtlmffai rs. org/press/novdec04( )3  .lit m 

Zwillich.  T.  (2004)  Little  Progress  Seen  in  Patient  Safety  Measures.  Washington:  Reuters 
Health  Information .  November  2004. 


Summer  2006 


40 


This  page  intentionally  left  blank 


Washington  Academy  of  Sciences 


41 


Fatigue,  Drowsy  Decision-Making  and  Medical  Error: 
Issues  of  Quality  Health  Care@ 

Gerald  P.  Krueger,  CPE 

Krueger  Ergonomics  Consultants 
Alexandria.  Virginia 

Abstract 

Health-care  providers,  particularly  interns,  residents,  and  nurses, 
participate  in  lengthy  workshifts  in  excess  of  10  hours,  often  work 
overtime,  work  through  the  night,  or  serve  on-call  at  the  hospital  for 
over  24  hours  at  a  stretch.  Mam  care  providers  obtain  insufficient 
sleep,  worker  fatigue  and  drowsiness  creep  in.  mood  and  attitudes  drift 
to  lower  levels,  and  performance  becomes  degraded.  Health  care 
providers  must  meet  high  performance  expectations  while  paying 
continuous  attention  during  sustained  monitoring  of  patients.  When 
drowsy,  they  may  engage  in  cntical  decision-making  while  they  are 
less  than  fully  alert.  This  paper  identifies  issues  of  "quality  of  health 
care"  pertaining  to  length}  hours  of  work,  rotating  shiftwork  schedules, 
circadian  rhythm  physiology  effects,  sleep  loss,  and  drowsiness, 
increasing  the  likelihood  of  worker  fatigue-related  error  while 
providing  institutional  health-care  services.  General  principles  for 
preparing  hospital  staffs  for  sustained  performance  are  outlined. 


Human  Error  in  Around-The-Clock  Provision  of  Health  Care 

Medical  personnel  at  hospitals,  nursing  homes,  and  extended  care 
facilities  have  always  been  in  the  forefront  of  meeting  our  societal 
expectations  for  around-the-clock  health  care,  seven  days  per  week,  365 
days  per  year  (/>.,  24/7/365).  New  patients  show  up  at  hospital  emergency 
rooms  at  any  time.  Many  hospitalized  patients  require  continuous  care, 
necessitating  full  time  health  care  staffing.  Elder  patients  stay  in  hospitals 


(v  This  paper  was  presented  in  a  symposium  on  Human  Error  in  Medicine .  sponsored  by 
the  Potomac  Chapter  of  the  Human  Factors  and  Ergonomics  Society  at  the  Washington 
Academy  of  Sciences'  Capital  Science  2006  Conference  at  the  National  Science 
Foundation,  Arlington.  VA  March  25.  2006. 


Summer  2006 


42 


or  at  nursing  homes  until  they  become  well  enough  to  return  home,  or 
until  they  succumb  to  illness  or  the  frailties  of  old  age. 

Human  error  in  medicine  As  a  prescient  soothsayer,  Marilyn 
Sue  Bogner’s  book:  Human  Error  in  Medichie  (Bogner  1994)  predated  the 
National  Academy  of  Sciences’  Institute  of  Medicine’s  (IOM)  1999 
publication  To  Err  is  Human.  Building  a  Safer  Health  System  (Kohn, 
Corrigan  &  Donaldson,  1999).  Both  books  describe  many  types  of  human 
error  that  intrude  into  provision  of  health  care,  suggesting  that  thousands 
of  patients'  lives  are  adversely  affected  or  even  shortened  by  health  care 
provider  errors.  A  subset  of  medical  errors  is  attributable  in  part  to  health 
care  worker  fatigue.  Drowsy,  sleepy,  or  fatigued  health  care  providers 
begin  to  experience  a  slackening  of  alertness,  lose  situational  awareness; 
neglect  to  monitor  a  patient’s  vital  signs  properly,  fail  to  detect  subtle 
changes  in  a  patient’s  condition,  omit  taking  some  action  they  should  have 
done,  or  make  less  crisp  and  effective  judgments  (Krueger  1994). 
Concerns  about  worker  fatigue  leading  to  medical  errors  range  from  a 
physician  making  an  inappropriate  diagnosis;  referral  of  a  patient  to 
incorrect  treatment  for  specific  illnesses;  a  surgeon  operating  on  the  wrong 
organ  or  limb;  a  surgical  team  leaving  sponges  or  instruments  inside  a 
patient’s  abdominal  cavity;  an  anesthesiologist  failing  to  monitor  a 
patient’s  vital  signs  or  administering  the  wrong  gaseous  mix  for  the 
patient’s  precarious  condition;  or  a  treatment  nurse  misreading  a  drug 
order  and  giving  the  patient  the  wrong  dose,  or  even  the  wrong  drug. 

Medical  errors  are  often  multi -factorial,  involving  human  factors 
such  as  inattention  or  poor  communication,  as  well  as  fatigue  (Cook  & 
Woods  1994).  In  terms  of  making  faulty  judgments,  medical  errors  are  not 
readily  documented,  nor  self-reported;  and  it  is  difficult  to  pinpoint  health 
care  provider  fatigue  as  a  proximal  cause  of  medical  errors.  Scant  direct 
data  shed  little  light  on  how  frequently  an  inappropriate  or  incorrect 
treatment  decision  is  made  while  a  health  care  worker,  e.g .,  an  intern  or  a 
resident,  was  overly  drowsy.  Whether  making  judgments  differently  while 
one  is  fatigued  translates  to  actual  instances  of  medical  error  or  simply 
results  in  “less  than  crisp”  medical  decisions  (not  classified  as  “errors”)  is 
not  easy  to  determine,  even  in  case  studies. 


Washington  Academy  of  Sciences 


43 


Drowsy,  Tired,  Fatigued  Health  Care  Workers  Can  Affect  Quality  of 

Health  Care 

Fatigue  and  health  care  providers 

It  is  difficult  to  pinpoint  influences  of  worker  fatigue  on  the 
incidence  of  medical  errors.  Just  what  are  the  impacts  of  fatigue  on 
provision  of  health  care?  Such  discussion  includes  questions  of:  “What  is 
fatigue,  and  what  do  we  mean  by  it?”  and  “What  is  the  role  or  contribution 
of  fatigue  in  the  workday  life  of  health  care  providers?”  Addressing  those 
questions  necessitates  pointing  out  the  importance  of  workers  obtaining 
sufficient  quantity’  and  quality  sleep  on  a  regular  daily  basis,  and  normal 
fluctuations  in  circadian  time-of-day  influences  on  drowsiness,  mood, 
attitude,  and  on  performance  (Krueger  1994). 

Short  of  witnessing  a  person  falling  asleep  on  the  job,  studying  the 
effects  of  fatigue  on  worker  performance  in  actual  workplaces  is  very 
time-consuming  and  not  easy  to  do.  Much  of  what  we  know  about  worker 
fatigue  comes  from  psychology  laboratory  studies  and  simulations 
(Hancock  &  Desmond  2001).  From  numerous  experiments  we  know  that 
tiredness  or  drowsiness  does  not  necessarily  result  in  errors  per  se.  When 
sustained  operations  test  participants  place  a  high  premium  on  accuracy  of 
their  work,  they  often  maintain  correctness  of  response  through  the  onset 
of  drowsiness  or  fatigue,  even  after  missing  significant  amounts  of  needed 
sleep.  Fatigue  primarily  affects  speed  of  thinking  and  is  almost  always 
accompanied  by  loss  of  alertness  and  measurable  degradations  in 
performance  on  cognitive  tasks  and  even  on  some  psychomotor  tasks.  In 
timed  laboratory  trials,  participants  tend  to  make  a  speed-accuracy 
tradeoff  -  that  is,  in  doing  the  best  they  can,  test  participants  preserve 
accuracy,  but  they  tend  to  slow  down  their  work,  and  therefore  accomplish 
fewer  items  of  work  over  time.  Generally  they  may  not  complete  all 
assigned  work  (Krueger  1989).  However,  fatigue  often  also  affects  a 
person’s  situational  awareness,  including  the  ability  to  incorporate  several 
sources  of  data  into  on-the-spot  problem-solving  as  well  as  many  other 
cognitive  processes.  Practical  parallels  in  many  workplaces,  including 
health  care  settings,  demonstrate  that  worker  fatigue  effects  manifest  as 
speed-accuracy  tradeoffs,  compromises  in  situational  awareness,  lessening 
of  attention  to  important  details,  and  compromises  in  judgment. 

“What  actually  happens  when  medical  care  personnel  get  tired, 
drowsy,  or  fatigued?”  A  review  of  the  published  literature  on  these 
important  questions  suggests  the  answers  are  more  qualitative  than 


Summer  2006 


44 


quantitative.  Poulton,  el  at.  (1978)  pointed  out  physicians  have 
performance  deficits  on  grammatical  reasoning  tests  after  sleep 
deprivation,  and  they  make  the  classical  speed-accuracy  tradeoff,  but 
concluded  that  physicians  can,  and  often  do,  compensate  for  effects  of 
sleep  loss  in  both  simple  and  complex  psychomotor  and  cognitive 
functions.  In  addressing  the  lengthy  hours  of  work  expected  of  interns  and 
resident  physicians,  Gaba  and  Howard  (2002)  wrote: 

. .  despite  many  anecdotes  about  errors  attributed  to  fatigue, 
no  study  has  proved  that  fatigue  on  the  part  of  health  care 
personnel  causes  errors  that  harm  patients.  Even  when 
impaired  clinical  performance  due  to  fatigue  or  falling 
asleep  has  allegedly  been  the  cause  of  specific  medical 
catastrophes,  these  incidents  have  been  viewed  as  isolated 
lapses  that  do  not  prove  that  the  safety  of  patients  was 
systematically  jeopardized. 

It  is  too  simple  to  suggest  that  health  care  providers  are  different 
from  other  workers,  able  to  ward  off  the  effects  of  fatigue  and  drowsiness 
in  their  work.  Extracting  from  several  articles  in  the  literature.  Table  1  lists 
comments  made  by  health  care  professionals  when  asked  about  their 
sensations,  thoughts,  and  experiences  with  fatigue  in  their  workplace. 

Table  1:  Health  Care  Provider  Comments  on  Experience  with  Fatigue 


I  had  difficulty  concentrating.  I  had  a  depressed  mood.  I  get  irritable,  I 
get  hostile. _ 

Feeling  hopelessness,  passivity,  lifeless,  demoralized,  pessimistic. 

As  the  night  wears  on,  we  become  irritable,  argumentative,  easily  agitated; 
sometimes  tempers  fray;  we  might  snap  at  nurses  and  fellow  workers,  pick 
on  people.  We  become  anxious,  explosive,  or  feel  very  depressed. _ 

I  verbally  snapped  at  fellow  workers,  the  staff,  even  the  patients,  in  ways 
that  tell  me  I  am  getting  overly  tired  and  ornery. _ 

When  fatigued,  I  am  not  a  cheery  friend  to  those  to  whom  1  usually  am  so. 

Reports  of  inappropriate  affect:  I  inappropriately  laughed  at  things  said 
regarding  patients  that  I  would  not  laugh  at  if  I  was  well-rested. _ 

Memory  deficit:  You  immediately  forget  what  you  or  others  just  said;  or 
in  conversation,  you  forget  what  you  want  to  say. _ 

I  didn’t  recall  whether  or  not  I  administered  the  needed  medication;  or 
whether  or  not  I  administered  the  correct  drug,  or  the  correct  dosage. _ 


Washington  Academy  of  Sciences 


45 


I  alternate  between  adrenalin-charged  highs  when  an  ambulance  arrives 
and  crushingly  weary  lows  when  patients  die  or  are  transferred. 

My  mind  slows  after  2  a  m.  and  we  all  cut  comers  on  even  the  most 
routine  procedures. 

I  give  or  gave  inappropriate  orders  or  directions  to  fellow  health  care 
workers,  i.e.  nurses  or  subordinates  that  I  would  have  likely  given 
differently  if  I  had  been  fully  alert  and  awake. 

I  miswrote  instructions  or  memos  in  the  medical  charts  or  records  of 
patients. 

Misdialing  phone  numbers,  making  mistakes  in  typing  out  orders, 
misreading  medication  prescriptions,  bottle  labels,  etc. 

I  drink  lots  of  caffeinated  coffee  in  an  attempt  to  stay  awake  or  alert. 

I  begin  yawning,  exaggerate  eye-blinking,  scratch  my  head,  hit  the  side  of 
my  face  or  head,  to  stay  awake;  feel  an  insatiable  urge  to  take  a  nap. 

I  quietly  become  apathetic,  negative,  and  don’t  give  a  damn;  give  bad  care 
to  patients. 

Sleep  deprivation  dangerously  impairs  judgment,  gives  a  sensation  of  a 
sleep-walking  nightmare. 

My  reading  attention  level  drops.  I  have  troubles  reading  medical  journal 
articles  or  references;  must  reread  passages  numerous  times;  unable  to 
comprehend  fine  points  of  prescription  drugs  in  the  Physician  \s  Desk 
Reference. 

You  recognize  you  were  about  to  administer  the  wrong  medication  or  the 
wrong  dose,  or  use  the  wrong  procedures  in  setting  up  equipment,  e.g ., 
administering  an  IV,  or  setting  up  an  infusion  pump  correctly. 

During  internship,  under  the  pressure  of  sleepless  call  nights,  my  worthy 
aspirations  as  a  medical  professional  transformed  into  cynicism. 

A  resident  in  our  program  got  so  impaired  by  sleep  deprivation  she  fell 
asleep  at  the  wheel  and  crashed  while  returning  from  a  40-45  hour 
workshift. 

(This  list  is  adapted  from  a  variety  of  sources,  with  modifications  by  this  author.) 

Interns ,  Resident  Physicians ,  House  Officers 


In  the  1890s  when  graduate  medical  education  programs  began  at 
Johns  Hopkins  School  of  Medicine,  a  resident  physician  was  expected  to 
live  at  the  hospital,  work  exceedingly  long  hours,  and  frequently  be  “on- 
call”  through  the  night  A  resident  attended  lectures  and  conferences,  was 
exposed  to  a  broad  spectrum  of  patient  cases,  and  provided  health  care  any 


Summer  2006 


46 


time  day  or  night  for  upwards  of  30-40  hours  at  a  stretch,  very  often 
without  much  sleep.  Residents  stayed  near  their  patients  to  observe  the 
sequalae  of  disease  over  successive  days  and  to  witness  the  impact  of 
medical  interventions  administered. 

Modern  residency  training  includes  3-5  years  of  long,  intensive 
work  and  study  under  supervision  of  senior  faculty  physicians  -  so  as  to 
obtain  substantial  hospital  clinical  experience  with  patients,  intermixed 
with  attending  didactic  lectures  and  participating  in  numerous  professional 
meetings  and  seminars  on  specialized  research  and  practice  topics  (Adler, 
Werner,  &  Korsch  1980).  The  combination  of  hands-on  clinical  work  and 
the  intensified  academic  training  prepares  physicians  for  practice  in 
medical  or  surgical  specialties. 

Today’s  interns  and  residents  work  dayshifts  lasting  8  to  12  hours, 
along  with  working  a  night  call  shift  every  2  to  4  nights.  Residents 
typically  do  not  sleep  much  during  night-call  shifts,  yet  are  expected  to 
continue  their  training  the  following  day.  There  are  only  168  hrs  in  a  7-day 
week;  but  some  residents  reported  working  shifts  of  from  18  to  60  hours 
duration,  with  every  other  night  on-call,  and  accumulations  of  more  than 
120-130  work  hours  per  week.  These  circumstances  can  involve  severe 
drowsiness  and  sleepiness,  with  an  accompanying  loss  of  situational 
awareness  or  alertness,  contributing  to  the  likelihood  of  medical  errors 
associated  with  fatigue. 

During  on-call  shifts,  sleep  obtained  during  brief  slack  periods 
tends  to  be  intermittent,  interrupted,  non-restorative  sleep.  When  an 
ambulance  arrives,  emergency  room  interns  who  are  temporarily  asleep, 
perhaps  napping,  are  abruptly  awakened  to  respond  to  the  arrival  of  new 
patients.  Arousing  from  a  short  sleep  in  the  middle  of  the  night  commonly 
produces  an  experience  of  sleep  inertia  -  bouts  of  severe  grogginess  and 
incomplete  arousal  attributable  to  awakening  from  the  deeper  stages  of 
sleep  (stage  3  &  4  sleep).  Sleep  inertia  can  last  10-20  minutes.  A  person 
experiencing  sleep  inertia  can  act  confused,  exhibit  poor  memory,  and 
demonstrate  inferior  decision-making  (Bruck  &  Pisani  1999).  It  is 
common  for  residents  or  interns  to  experience  bouts  of  acute  fatigue, 
accompanied  by  sleep  inertia  after  awakening  from  naps. 

Due  to  today’s  medical  advances  and  concern  for  cutting  costs, 
patient  stays  in  hospitals  are  shorter  then  a  decade  ago.  With  rotating 
workshifts,  and  more  hand-off  of  patients  among  health  care  providers, 
today’s  resident  no  longer  spends  as  much  time  watching  a  patient’s 
disease  progress.  Having  interns  work  24+hour  on-call  schedules  provides 


Washington  Academy  of  Sciences 


47 


teaching  hospitals  with  necessary  personnel -related  efficiencies  netting 
obvious  cost  savings  by  employing  residents  as  relatively  “cheap  labor’1 
(Steinbrook  2002).  Sleep  deprivation,  extreme  drowsiness,  and  resultant 
fatigue  in  the  medical  workplace  are  identified  as  some  of  the  several 
major  sources  of  stress  in  residency  (Colford  &  McFee  1989)  prompting 
interns  and  residents  to  declare  they  suffer  from  chronic  fatigue  during 
residency.  However,  many  in  the  medical  profession  argue  that  sleep- 
depriving  night-call  is  a  valid  learning  experience  and  quality  of  care  is 
not  compromised  by  sleep-deprived  physicians  (Asken  &  Raham  1983). 
Traditional  on-call  assignments  still  are  deemed  necessary  as  part  of 
resident  training.  Thus,  for  over  100  years,  not  much  has  changed 
regarding  the  exceedingly  long  duty  hours  for  medical  residents. 

Large  numbers  of  today’s  residents  are  more  likely  to  be  older  than 
was  true  fifty  years  ago;  many  residents  have  families  or  other 
commitments  outside  the  hospital.  Over  50%  of  medical  residents  in  the 
United  States  are  women,  with  unique  considerations  such  as  childcare 
needs.  These  facts  complicate  life  and  affect  how  residencies  are  managed 
in  today’s  medical  schools.  Few  studies  focused  solely  on  how  extended 
duty  hours  affect  the  home  life  of  resident  physicians;  but  we  do  know  that 
medical  workers  on  sustained  schedules,  including  residents,  are  involved 
in  significant  numbers  of  traffic  accidents  driving  home  after  lengthy 
hospital  workshifts  (Barger  el  a/.  2005).  Many  stressors  combine  to  affect 
resident  life,  but  it  is  unclear  that  extended  duty  hours  themselves  are 
responsible  for  negative  consequences  in  health  care  (Liskowsky  1991). 

Residents  and  interns  working  lengthy  schedules  invoke  a  vision  of 
drowsy  young  doctors  making  important  health  care  decisions  while  they 
are  sleep-deprived,  when  their  attention  levels  are  not  as  crisp,  and  when 
they  might  fail  to  correctly  grasp  nuances  of  some  medical  maladies  of 
their  patients’  cases.  Do  drowsy  residents  make  judgments  or  medical 
decisions  which  are  not  fully  appropriate  for  the  circumstances  at  hand? 

Sleep  deprivation  impairs  decision-making  involving  the 
unexpected  innovation  (or  involving  innovation),  revising  plans,  and 
competing  distractions,  as  well  as  interfering  with  effective 
communication  (Krueger  1994).  These  are  all  involved  in  making  accurate 
diagnosis  and  prescribing  the  best  treatments  for  tricky  medical  cases  that 
may  arise  while  resident  physicians  are  sleepily  working  their  way  through 
on-call  duty.  Although  no  studies  captured  performance  measures  with 
physicians  on  unexpected,  innovative,  plan  revision,  communication-laden 
decision-making  circumstances,  it  is  tempting  to  connect  the  intuitive  link 


Summer  2006 


48 


between  worker  fatigue  and  degradation  in  decision-making  performance. 
We  would  dread  having  sleepy  residents  in  the  middle  of  the  night  making 
life-determining  decisions  for  our  medical  case  at  the  emergency  ward. 

Most  laboratory  studies  of  performance  decrement  effects  of  sleep 
deprivation  employ  “probe  tests,”  simple  tasks  (e.g.,  vigilance,  reaction 
time,  and  short  term  memory)  known  to  be  sensitive  to  sleep  deprivation 
and  representative  of  the  attention-demanding  requirements  of  dull, 
monotonous  monitoring  tasks  in  everyday  jobs.  Studies  of  sleep 
deprivation  indicate  resident  physician  performance  is  impaired  for  some, 
but  not  all  tasks;  task  performance  decrements  are  similar  to  those  for 
psychological  tests  sensitive  to  sleep  loss  (Samkoff  &  Jacques  1991).  A 
30-year  old  study  by  Goldman,  McDonough,  and  Rosemond  (1972)  found 
substantial  decrements  in  the  performance  of  surgical  interns  after 
obtaining  too  little  sleep  on-call  nights.  Following  one  night  of  sleep  loss, 
interns  exhibited  poor  planning  skills,  inferior  surgical  technique,  and 
committed  more  errors.  After  night  on-call  duty,  sleep-deprived  interns, 
exhibit  poorer  concentration  (Robbins  &  Gottlieb  1990);  compromised 
language  and  numeric  skills  (Hawkins  et  al.  1985);  degraded  retention  of 
information  (Hart  et  al  1987);  and  fleeting  short  term  memory  (Rubin  et 
al  1991). 

After  missing  one  night’s  sleep,  surgeons  were  more  prone  to 
errors  and  performed  slower  on  a  laparoscopy  simulator  (Taffinder  et  al. 
1998).  Whereas,  Reznick  &  Folse  (1987)  found  no  performance 
differences  between  sleep-deprived  and  rested  surgery  residents  on  a 
comprehensive  psychomotor  test  battery.  Fatigue  is  common  for  members 
of  surgical  teams  who  repeatedly  become  involved  in  long,  complicated 
operative  cases  such  as  intricate  neurosurgeries  that  sometimes  take  from 
12  to  20+  hours  to  complete  (Greenberg  1997).  In  a  24-hour  performance 
study,  emergency  physicians  made  more  errors  on  a  simulated  triage  test 
and  while  intubating  a  training  mannequin  (Smith-Coggins  et  al.  1997). 

Other  studies  of  impaired  performance  by  “fatigued”  residents  and 
interns  found:  evaluation  and  interpretation  of  electrocardiograms  wanting 
(Lingenfelser  et  al  1994);  excessive  time  was  required  to  review  and 
mistakes  were  made  with  telemetered  ECG  data  (Friedman,  Bigger,  & 
Komfeld  1971);  inconsistent  response  times  monitoring  of  anesthesia 
during  simulated  surgery  (Denisco,  Drummond,  &  Gravenstein  1987); 
general  degradations  in  cognitive  performance  (Jacques,  Lynch,  & 
Samkoff  1990),  lapses  in  memory  and  slower  responses  (Hart  et  al , 
1987);  mistakes  while  ordering  medications  and  documenting  medical 


Washington  Academy  of  Sciences 


49 


histories  (Gottlieb  et  al.  1991);  compromised  clinical  problem  solving 
(Rubin  et  at.  1991;  Nelson  et  al.  1995),  and  stresses  in  radiology 
residencies  (Christensen  et  al.  1977). 

Less  easily  quantified  are  the  negative  affects  such  as  poor  mood 
or  communications  and  the  interactions  a  sleepy  care  provider  has  with 
patients  (Deary  &  Tait  1987).  Resident  physicians  in  training  readily 
report  more  negative  feelings  and  less  concern  about  their  patients  when 
they  are  suffering  from  insufficient  sleep  on-call  (Orton  &  Gruzelier 
1989).  Wallerstein,  Rosner,  and  Wallace  (1989)  reported  rested  interns 
had  better  moods,  felt  more  vigor,  less  fatigue,  and  more  elation. 
Baldwin,  Dodd  and  Wrate  (1997)  reported  numerous  concerns  for  the 
health  and  the  psychological  stresses,  even  clinical  depression  of  interns, 
resident  physicians,  and  student  nurses.  Some  stresses  noticeably  affect 
family  issues  and  in  some  cases  prompt  divorces  (Nelson  &  Henry  1978). 

Residents  Interns  Workshift  Honrs 

There  have  been  several  attempts  to  assess  and  propose  ways  to  cut 
down  on  resident  fatigue  in  training  programs  ( e.g .  Richardson  et  al. 
1996).  Although  different  traditional  workshift  lengths  have  been 
examined  for  nurses,  there  are  only  a  few  comparison  studies  for 
physicians.  In  comparing  8-  vs  12-hour  shifts  for  emergency  physicians, 
Thomas,  Schwartz  and  Whitehead  (1994)  found  only  insignificant 
differences  in  performance.  In  the  1980s  some  medical  school  training 
programs  developed  the  night-float  rotation  system ,  whereby  residents 
work  a  series  of  from  5  to  15  consecutive  nights  on-call  without  any 
daytime  work  activities  and  are  permitted  to  sleep  during  the  day.  The 
night-float  rotation  system  gives  day-residents  greater  opportunity  to  sleep 
at  night  and  produces  high  levels  of  satisfaction  among  residents;  but, 
those  on  night  float  reported  lower  sleep  quality  and  duration,  mood 
changes,  less  vigor,  slower  thought  processes,  and  some  depression, 
claiming  their  attention  levels  were  unchanged,  but  admitted  to  more 
errors  of  omission,  and  fatigue  inertia  (Cavallo,  Ris,  &  Succop  2003). 
Although  the  night-float  rotation  system  has  intuitive  appeal,  the  full 
impact  of  this  system  is  not  clear,  and  there  is  room  for  more  study  of  how 
best  to  integrate  it  into  modern  residency  training. 

Another  contributor  to  the  shortage  of  sleep  and  free  time  of 
medical  residents  is  the  large  amount  of  moonlighting  (overtime  work) 
they  undertake.  Moonlighting  shifts  are  often  at  odd  work  hours  and  are 
disruptive  to  normal  sleep  patterns.  Presumably,  the  driving  incentive  is  to 


Summer  2006 


50 


earn  the  extra  money  moonlighting  offers.  Nearly  half  of  all  emergency 
medicine  residents  in  the  United  States  partake  in  moonlighting  (Li, 
Tabor,  &  Martinez  2000);  as  many  as  65  per  cent  of  internal  medicine 
residents  and  fellows  moonlight  (McCue,  Janiszewski,  &  Stickley  1990). 
Moonlighting  is  common  among  other  residencies  and  fellowships  as  well 
(Majidian  el  al  1993). 


New  Work  Hour  Limits  for  Residents  and  Interns 

A  New  York  hospital  incident  in  1984  stimulated  public  concern 
about  fatigued  interns  and  residents,  triggering  attempts  to  limit  resident 
hours  of  work.  Ms  Libby  Zion  was  admitted  to  a  hospital  emergency 
room,  where  she  was  treated  by  an  intern  and  a  junior  resident;  but  she 
died  several  hours  later.  Both  the  intern  and  the  resident  had  been  on  duty 
for  18+  hours  prior  to  her  admission.  It  was  alleged  they  failed  to  properly 
monitor  their  patient,  and  that  they  prescribed  medications  contraindicated 
in  light  of  her  history  of  drug  and  medication  use.  Additionally,  they  were 
not  properly  supervised  by  a  senior  supervisory  physician  (Asch  &  Parker 
1988).  This  case  prompted  much  publicity  and  discussion  about  overwork 
in  medical  residency  programs  (Green  1995).  The  resulting  campaign  to 
limit  the  number  of  work  shift  hours  of  interns  and  residents  in  New  York 
caught  on  across  the  country.  (Daughtery,  Baldwin  &  Rowley  1998 
ACGME  2003). 

After  the  1999  IOM  study  of  medical  error,  sensitivity  to  patient 
safety  issues  has  been  heightened  more  generally.  The  US  medical 
profession,  and  some  state  legislatures,  are  implementing  reform  plans  to 
restrict  hospital  work  schedule  lengths  by  developing  hours  of  service 
(HOS)  limiting  rules  for  residents  similar  to  those  in  place  for 
transportation  operators  and  controllers  of  nuclear  power  plants. 

The  Accreditation  Council  for  Graduate  Medical  Education 
(ACGME),  the  accrediting  body  for  7,800  graduate  medical  education 
programs  in  118  specialties  in  the  United  States  worked  two  decades  to 
establish  limits  to  duty  hours  of  resident  physicians  and  interns.  Effective 
in  July  2003,  ACGME’s  new  approved  rules  limited  work  hours  for 
medical  students  to  no  more  than  80  hours  a  week  (averaged  over  a  4- 
week  period)  and  limited  shift  duration  to  no  longer  than  24  hours,  with  at 
least  a  minimum  of  10  hrs  off-duty  between  workshifts.  Residents  must 
have  at  least  one  full  day  (24-hours)  out  of  seven  free  of  educational  and 
clinical  care  responsibilities  (averaged  over  4-weeks).  Residents  must  not 


Washington  Academy  of  Sciences 


51 


be  assigned  in-house  call  more  often  than  every  third  night  (averaged  over 
4-weeks).  Continuous  time-on-duty  including  in-house  call  must  not 
exceed  24  consecutive  hours,  with  additional  time  (up  to  six  hours)  for 
inpatient  and  outpatient  continuity,  transfer  of  care,  educational  debriefing 
and  formal  didactic  activities.  Residents  may  not  assume  responsibility  for 
new  patients  after  24  hours.  Since  moonlighting  to  perform  other  patient 
care  activities  may  be  inconsistent  with  interns  obtaining  enough 
rest/sleep,  these  activities  require  prospective  permission  from  program 
directors  and  sponsoring  institutions,  and  resident  performance  must  be 
monitored.  ( www.acgme.org ) 

ACGME’s  new  rules  emphasize  faculty  supervision  to  ensure  safe 
patient  care  and  resident  learning.  Faculty  and  residents  are  to  be  educated 
to  recognize  the  signs  of  worker  fatigue  and  to  apply  preventive  and 
operational  countermeasures.  A  medical  school  program  director  and  the 
faculty  are  to  monitor  residents  for  the  effects  of  sleep  loss  and  fatigue, 
and  are  to  respond  when  fatigue  may  be  detrimental  to  resident 
performance  and  his/her  well-being. 

The  new  hours  limitations  have  enormous  cost  implications  as  the 
need  for  additional  hospital  staffing  increased  (Weinstein  2002; 
Steinbrook  2002).  However,  if  the  new  work-hours  rules  are  not  followed 
ACGME  threatens  to  withdraw  a  teaching  hospital’s  accreditation,  which 
could  cost  training  hospitals  millions  of  dollars  in  federal  funding.  To 
emphasize  the  seriousness  of  the  changes,  in  July  2003,  after  it  was 
determined  that  several  first-year  residents  worked  almost  90  hours  per 
week,  the  Johns  Hopkins  Hospital’s  accreditation  was  compromised  for 
five  months  until  it  restructured  workshifts  to  comply  with  the  new 
standards. 

After  two  years  experience  with  ACGME’s  common  duty  hours,  a 
confidential  Internet  survey  of  over  50,000  residents  indicates  many 
residency  programs  are  using  innovative  approaches  to  restructure  duty 
hour  schedules  for  residents  and  the  vast  majority  of  residency  programs 
are  complying  with  the  new  duty  hour  rules  (ACGME  Press  Release, 
September  2005).  However,  it  is  also  clear  that  the  rules’  several 
extenuating  circumstances,  ( e.g .,  averaging  over  4-weeks,  etc.)  have 
provided  sufficient  “wiggle-room”  for  there  to  be  internal  scheduling 
tradeoffs  that  somewhat  thwart  the  spirit  and  intent  of  the  rules  to  prevent 
intern  and  resident  fatigue. 


Summer  2006 


52 


Nurses,  Shiftwork,  &  Fatigue 

The  literature  contains  reports  of  numerous  studies  of  shiftwork 
schedules  for  nurses:  presenting  examinations  of  job  performance  via  self- 
reported  or  supervisor-rated  measures;  patient  ratings  of  quality  of  care 
received;  use  of  sick-days  and  health  services  by  nurses;  and  nurses’ 
preferences  for  particular  shift  schedules.  Just  as  it  is  with  interns  and 
residents,  the  “whole  person”  is  of  concern;  for  many  factors  play  into 
provision  of  excellent  nursing  care.  Nurses  work  a  variety  of  non-daytime 
shifts,  and  endure  circadian  rhythm  disruption  not  just  for  the  duty-time  at 
work,  but  often  chronically  over  the  duration  of  their  professional  and 
personal  lives.  While  the  likelihood  of  nursing  care  errors  attributable  to 
sleep  loss,  waning  alertness,  cognitive  fatigue,  and  circadian  disruption  are 
addressed  here,  it  must  also  be  acknowledged  that  nurses  often  experience 
physical  fatigue  while  administering  to  patients,  especially  while 
repeatedly  helping  bed-ridden  patients  move  about  in  hospital  rooms  and 
nursing  centers. 

Shiftwork  schedules  that  are  more  compatible  with  circadian 
rhythm  physiology  and  which  permit  nurses  to  obtain  more  quality  sleep, 
should  result  in  less  fatigued,  more  alert  nurses  on  the  job.  However,  many 
other  variables  involved  with  shiftwork  make  problematic  the  delineation 
of  fatigued  nurses  per  se.  Nurses’  reasons  for  agreeing  to  work  particular 
shifts  involve  personal  schedule  preferences,  the  need  for  child  care  at 
home,  worries  of  personal  security  in  dark  hospital  parking  lots,  salary 
differences,  desires  to  work  overtime  or  not,  electing  to  work  longer  hours 
over  fewer  days  in  trade  for  more  successive  days  off  from  work, 
perceived  control  over  their  jobs,  and  other  family  reasons  and  indicators 
of  worker  satisfaction  or  dissatisfaction.  Determinations  of  whether 
nurses’  work  schedules  affect  levels  of  alertness  on  the  job  or  increase 
fatigue-related  medical  error  is  difficult  for  more  than  a  case-by-case 
basis. 


Nurses’  workshift  schedules  generally  follow  six  basic  scheduling 
schemes  (Liskowsky  1991): 


(1)  Traditional  8-hour  day  shifts,  5  days  per  week,  with  15-30  minutes 
tacked  on  for  transfer/changeover  of  patient  care  to  incoming 
personnel;  rotation  to  different  start  times  is  on  a  one  week 
change-over  cycle  (start  times  usually  are  7  a.m.,  3  p.m.,  or  11 
pm.); 


Washington  Academy  of  Sciences 


53 


(2)  “4  to  40”  -  four  10-hour  shifts,  followed  by  three  days  off-duty; 

(3)  “Baylor  Plan”  -  two  or  three  12-hour  shifts  for  separate  2-3  day 
weekend  staffing;  while  regular  staff  works  traditional  five  8-hour 
shifts  the  Monday -Friday  workweek, 

(4)  “7-on,  7-off’  -  working  seven  days  or  nights  of  10-hour  shifts  on 
alternate  70-hr  work  weeks,  with  the  intervening  week  off-duty; 

(5)  “12-hour  shifts”  -  three  12-hour  shifts  one  week,  then  four  12-hr 
shifts  the  next  week; 

(6)  “Customized  schedules”  -  choice  of  many  combinations  of  shift 
lengths,  start  and  end  times. 

Scheduling  has  been  a  major  source  of  stress  among  hospital 
nurses.  In  1991,  the  Congressional  Office  of  Technology  Assessment 
estimated  one-third  of  all  RNs  worked  some  combination  of  day,  evening, 
and  night  shifts,  including  rotating  all  three  shifts;  only  7  per  cent  of  RNs 
had  every  weekend  off  (Liskowsky,  OTA  1991).  Younger,  less 
experienced  nurses  tend  to  be  assigned  to  rotating  shifts,  and  more 
experienced  nurses  are  assigned  the  more  desirable  dayshifts.  Many  US 
hospitals  pay  extra  for  evening  and  nightshift  work,  but  not  for  weekends. 
Nurses  draw  1.5  times  as  much  pay  for  overtime  work;  and  many  nurses 
like  to  amass  their  weekly  working  hours  in  12-hr  shifts  to  obtain  more 
consecutive  days  off. 

Supervisors  rate  quality  of  nursing  care;  or  it  may  be  measured  in 
terms  of  nursing  care  process  ( e.g .,  chart  audits  of  following  planned 
procedures).  Other  indicators  of  quality  include  incident  reports  of 
medication  errors,  accidents  or  injuries,  and  occasionally  patient  ratings  of 
satisfaction  with  care  received.  Nurses  themselves  may  occasionally 
experience  adverse  circumstances  in  the  workplace  (e.g.,  administering  an 
incorrect  drug  or  dose)  but  they  do  not  readily  attribute  drowsiness  or 
fatigue  with  a  contributory  role  in  the  outcome  of  some  decision-making 
or  action  they  took  or  that  they  should  have  taken  on  a  patient’s  behalf. 
Nurses’  shiftwork  schedules,  or  working  extended  duty  hours,  are  among 
several  factors  contributing  to  medical  incidents  in  hospitals  or  nursing 
homes,  but,  as  with  the  physicians,  it  is  difficult  to  partial  out  the  effects  of 
drowsiness  or  fatigue  of  nursing  care  providers  on  the  likelihood  of 
medical  errors. 

Although  there  are  numerous  studies  of  the  effects  of  shiftwork  on 
nurses,  few  studies  directly  examined  the  relationship  between  nursing 
shiftwork  schedules  and  job  task  performance,  or  match  schedule  effects 


Summer  2006 


54 


to  indicators  of  quality  of  patient  care.  In  a  pediatrics  unit,  when 
comparing  the  4-to-40  workweek  (four  10-hr  workshifts  per  week)  to  the 
traditional  8-hr,  5-dayshift  system,  process  measures  did  not  differentiate; 
but  staff  reports  showed  the  4-to-40  schedule  made  improvements  in 
quality  of  intershift  continuity  of  care  (Kent  1972).  In  a  comparison  of  12- 
hr  shifts  to  8-hr  shifts  for  nurses  in  an  intensive  care  unit  Eaton  and 
Gottselig  (1980)  found  no  significant  differences  between  types  of  shift 
for  reaction  time  measures  of  alertness,  or  for  self-ratings  of  fatigue. 
Although  reaction  times  on  the  12-hour  shift  were  faster  for  day  shift  than 
for  nightshift  nurses,  quality  of  nursing  care  was  deemed  not  to  have 
changed.  In  a  similar  study  in  a  surgical  intensive  care  unit  adopting  12-hr 
shifts,  some  nurses  reported  increased  subjective  feelings  of  fatigue,  and 
decreased  accuracy  was  reported  on  some  performance  tests;  but  evidence 
from  chart  audits  revealed  no  significant  changes  in  quality  care,  and  the 
nurses  themselves  reported  they  thought  their  performance  actually  had 
improved  (Mills,  Arnold,  &  Wood  1983).  In  another  study,  RNs  in 
intensive  care  units  on  12-hr  vs.  8-hr  shifts  volunteered  comments  about 
experiencing  more  fatigue;  but  ratings  of  patient  care  were  not 
substantially  different  (Nelson  &  Blasdell  1988). 

Most  shiftwork  studies  address  the  issue  of  whether  regular 
shiftworkers  obtain  sufficient  sleep.  All  shiftworkers  adopt  sleep  patterns 
different  from  a  “normal  daytime  worker”  and,  in  so  doing,  most, 
especially  night  workers,  experience  deficits  in  the  quality  and  the 
quantity  of  sleep,  generally  obtaining  about  an  hour  less  sleep  per  24-hr 
day  (Scott  1990). 

An  early  NIOSH  study  reported  nurses  on  rotating  shifts,  or  even 
on  fixed  nightshifts,  experienced  more  problems  with  sleep,  as  compared 
to  nurses  who  worked  dayshifts  or  fixed  afternoon  swing  shifts;  and 
rotating  shift  nurses  exhibited  higher  rates  of  digestive  disorders  than  other 
nurses  (Tasto  el  ctl.  1978).  Those  on  fixed  nightshifts  reported  obtaining 
the  least  sleep  overall.  Harma,  Ilmarinen,  and  Knauth  (1988)  reported 
nurses  on  irregular  rotating  shifts  experienced  decreased  sleep  duration 
after  nightshift  work.  Gadbois  (1981)  also  described  how  women  nurses 
on  fixed  nightshifts  reported  shorter  sleep  durations,  with  more  frequent 
sleep  interruptions  for  married  women  with  children  than  for  unmarried 
women.  After  working  the  nightshift,  mothers  with  young  children  went  to 
bed  later  in  the  day  than  did  the  unmarried  women.  Nurses  working 
rotating  shifts  and  nightshifts  involving  only  a  few  nights  on  duty  tend  to 
have  more  sleep  disturbances  than  other  nurses.  The  greatest  disruption  of 
family  and  social  life  occurs  for  nurses  on  rotating  shifts.  Nurses  who  are 


Washington  Academy  of  Sciences 


55 


also  mothers  express  dissatisfaction  with  time  available  to  be  with  their 
children  (Liskowsky,  OTA  1991). 

On  March  13,  2006,  the  American  Nurses  Association  requested 
the  National  Institute  of  Occupational  Safety  and  Health  (NIOSH)  in  its 
2006  National  Occupational  Research  Agenda  (NORA)  give  priority  to 
sponsoring  additional  research  on  nursing  care  and  fatigue. 


Advice  and  Discussion  Points 

This  article  provides  only  a  cursory  review  of  some  of  the  concerns 
for  the  risk  of  health  care  provider  fatigue  influencing  the  likelihood  of 
medical  errors.  It  should  prompt  discussion  among  managers  and 
supervisors  in  medical  institutions  about  what  to  do  to  manage  worker 
alertness  and  fatigue.  Based  upon  substantial  prior  experience  as  an 
operator  fatigue  subject  matter  expert  consultant  to  US  military  forces, 
and  to  the  long  haul  truck  driving  community,  this  author  offers  two  lists 
of  hints  for  reducing  the  risks  of  worker  fatigue  in  health  care  settings. 


Advice  for  institutional  health  care  managers  and  supervisors 

The  1994  chapter  in  Bogner’s  first  edition  of  Human  Error  in 
Medicine  (Krueger,  1994)  ended  with  basic  pointers  regarding  fatigue, 
performance,  and  medical  error.  Those  not-so-simple  hints  which  are  still 
valid  today  are  amplified  here  in  a  12-step  fatigue  management  program 
for  hospital  and  nursing  home  management  staff.  If  supervisors  want  to 
help  their  employees  reduce  risks  of  worker  fatigue,  which  can  contribute 
to  costly  human  errors  in  provision  of  health  care,  they  are  advised  to: 

1.  Allocate  adequate  staffing  for  around-the-clock  operations. 
Scrutinize  rosters  for  under-staffing  in  peak  periods;  rectify 
staffing  discrepancies. 

2.  Cross-train  several  workers  for  the  same  tasks  so  they  will  be  able 
to  spell  or  relieve  one  another,  permitting  periodic  rest  breaks. 

3.  Train  staff  to  perform  tasks  so  well  (over-learning)  that  they  are 
less  likely  to  be  subject  to  fatigue  effects.  Acknowledged,  it  is 
difficult  to  over-learn  medical  care  decision-making. 

4.  Honor  known  bodily  circadian  rhythm  principles  when  designing 
shiftwork  schedules. 


Summer  2006 


56 


5.  Schedule  rest  breaks  during  long  work  sessions,  especially  at  high 
fatigue  risk  periods  (mid-afternoon,  and  from  1  to  5  a.m). 

6.  Approve  implementation  of  fatigue  countermeasures,  including 
infusing  short  nap-taking  into  some  workshifts  ( e.g .  night  calls); 
provide  a  good  place  to  take  naps;  emphasize  recognition  of  sleep 
inertia. 

7.  Stress  importance  of  rest  and  sleep  for  an  alert  staff.  Implement  a 
training  program  on  worker  alertness  and  fatigue. 

8.  Encourage  employees  to  eat  nutritious  meals.  Ensure  ready  access 
to  healthy  food  choices. 

9.  Learn  to  recognize  signs  of  fatigue  in  medical  staff  personnel. 
ACGME’s  new  work  hour  rules  require  both  faculty  and  residents 
to  be  educated  to  recognize  the  signs  of  fatigue,  and  to  apply 
preventive  and  operational  countermeasures.  Internship  program 
directors  and  faculty  must  monitor  residents  for  effects  of  sleep 
loss  and  fatigue,  and  respond  when  fatigue  may  be  detrimental  to 
resident  performance  and  well-being. 

10.  Bring  in  a  worker-fatigue  expert  to  provide  alertness  and  fatigue 
management  assessment  and  training  for  all  employees,  especially 
those  at  high  fatigue  risk  (i.e.,  nurses,  interns,  residents). 

11.  Become  knowledgeable  of  the  circumstances  surrounding  the 
numerous  shift  schedule  issues  of  your  institution.  Do  not  tamper; 
but  offer  an  understanding  ear  to  those  who  are  making  the 
schedules  and  to  those  who  are  living  them. 

12.  Set  the  example  and  develop  a  sleep  management  /  sleep  discipline 
plan  for  yourself;  and  then  encourage  others  to  do  likewise. 

Mastering  alertness  and  coping  with  worker  fatigue  for  health  care 

providers 

Health  care  providers  (hospital  and  nursing  home  employees)  must 
learn  to  cope  with  shiftwork,  long  working  hours,  and  shortage  of  sleep. 
They  must  learn  to  recognize  waning  alertness,  the  onset  of  worker 
fatigue,  and  know  what  to  do  about  it.  The  following  pointers  may  be  of 
some  help: 

1 .  For  supervisors  and  those  at  risk  of  worker  fatigue,  a  good  place  to 
start  is  by  attending  the  Institution’s  training  course  on  mastering 
alertness  and  managing  health  care  provider  fatigue. 

2.  Rotating  one’s  shiftwork  schedule  every  couple  days  or  weeks 
(forward  or  backward  on  the  clock)  forces  our  physiology  to  make 


Washington  Academy  of  Sciences 


57 


adjustments  to  resynchronize  our  circadian  biological  timing 
system  (suprachiasmatic  nuclei). 

3.  Rapid  rotations  of  one’s  work  schedule  cause  physiological 
discomforts,  and  disrupts  ability  to  acquire  the  right  amount  of 
sleep.  Learning  principles  of  circadian  rhythm  physiology  and 
workshift  scheduling  may  help  workers  sort  out  the  best  ways  to 
cope  with  arduous  work  hours. 

4.  Generally,  working  longer  than  12  hours  in  a  row  at  almost 
anything  increases  risk  of  worker  fatigue.  Try  to  avoid  working 
double  shifts  that  go  beyond  12-hours  of  continuous  work.  (On-call 
residents  will  often  be  required  to  make  exceptions  to  this 
admirable  goal). 

5.  Workshifts  requiring  some  night  work  ( e.g .,  swing  and  midnight 
shifts)  often  result  in  the  worker  obtaining  approximately  1.2  hours 
less  sleep  per  day. 

6.  Adults  operate  reasonably  well  with  7-8  hours  of  sleep  in  every  24- 
hour  period.  Obtaining  sleep  in  long  continuous  bouts  (4+  hours)  is 
preferred  to  taking  numerous  shorter  sleeps  (naps). 

7.  It  is  important  to  augment  shorter  sleeps  with  naps  to  reach  a  goal 
of  7-8  hrs  of  sleep  in  every  24-hr  period. 

8.  With  daily  shortages  of  sleep  we  accumulate  a  sleep  debt  which 
biologically  we  must  pay  back  to  our  brain  and  body.  It  is 
critically  important  to  obtain  extra  long  sleeps  on  days  off 
(recovery  sleep),  to  make  up  for  our  accumulated  sleep  debt,  and  to 
restore  alertness. 

9.  Develop  a  list  of  signs  of  “waning  alertness,  onset  of  fatigue” 
symptoms,  to  be  able  to  recognize  the  onset  of  fatigue,  and  know 
what  to  do  about  it. 

10.  Be  knowledgeable  and  attuned  to  possible  fatigue  countermeasures 
available;  learn  to  use  them.  Education  and  trial  and  error  of 
various  techniques  can  help. 


References 


Accreditation  Council  for  Graduate  Medical  Education  (2003).  Common  program 

requirements,  resident  duty  hours  language  (final  requirements),  and  residency 
review  committee  procedures  for  granting  duty  hours  exceptions. 
lit  tp :  //www' .  ac  gme .  org 


Summer  2006 


58 


Accreditation  Council  for  Graduate  Medical  Education  (2005).  September  2005  press 

release  on  first  two  years  of  medical  school  experiences  with  the  new  duty  hours 
limitations  for  interns  and  resident  physicians,  http ://www .acgme.org 

Adler.  R..  Werner.  E.R..  &  Korsch.  B.  (1980).  Systematic  study  of  four  years  of 
internship.  Pediatrics .  66.  6.  1000-1008. 

Asch.  D  A.  &  Parker.  R.M.  (1988).  The  Libby  Zion  case:  One  step  forward  or  two  steps 
backward?  New  England  Journal  of  Medicine,  318.  771-775. 

Asken.  M.J.  &  Raham.  D.C.  (1983).  Resident  performance  and  sleep  deprivation:  A 
review.  Journal  of  Medical  Education.  58,  382-388. 

Baldwin.  P.J..  Dodd.  M..  &  Wrate.  R.W.  (1997).  Young  doctors'  health  I.:  How  do 

working  conditions  affect  attitudes,  health,  and  performance?  Social  Science 
and  Medicine,  45.  35-40. 

Barger.  L.K..  Cade.  B.E..  Ayas.  N.T..  Cronin.  J.W..  Rosner.  B.,  Speizer.  F.E..  &  Czeisler. 
C.  A.  (2005).  Extended  work  shifts  and  the  risk  of  motor  vehicle  crashes  among 
interns.  New  England  Journal  of  Medicine.  352.  125-134. 

Bogner.  M.S.  (Ed.)  (1994).  Human  error  in  medicine.  Hillsdale.  NJ:  Lawrence  Erlbaum 
Associates.  Publishers. 

Bruck.  D.  &  Pisani.  D.L.  (1999).  The  effects  of  sleep  inertia  on  decision-making 
performance.  Journal  of  Sleep  Research.  8.95-103. 

Cavello.  A..  Ris.  M.D..  &  Succop,  P.  (2003).  The  night  float  paradigm  to  decrease  sleep 
deprivation:  good  solution  ora  new  problem?  Ergonomics  46.  653-663. 

Christensen.  E.E..  Dietz,  G.  W..  Mum  ,  R.C..  &  Moore.  J.  G.  (1977).  The  effect  of 
fatigue  on  resident  performance.  Radiology .  125.  103-105. 

Colford.  J.M..  &  McFee.  S.J.  (1989).  The  raveled  sleep  of  care:  Managing  the  stresses  of 
residency  training.  Journal  of  the  American  Medical  Association.  261:  889-893. 

Cook.  R.I..  &  Woods.  D.D.  (1994).  Operating  at  the  sharp  end:  The  complexity  of 

human  error.  In:  M.S.  Bogner  (Ed.)  Human  Error  in  Medicine.  Hillsdale,  NJ: 
Lawrence  Erlbaum  Associates. 

Daugherty.  S.R..  Baldwin,  D  C.,  &  Rowley.  B.D.  (1998).  Learning,  satisfactioa  and 
mistreatment  during  medical  internship:  A  national  survey  of  working 
conditions.  Journal  of  the  American  Medical  Association,  279.  1194-1199. 

Deary.  I.J.  &  Tait,  Q.R.  (1987).  Effects  of  sleep  disruption  on  cognitive  performance  and 
mood  in  medical  house  officers.  British  Medical  Journal  Clinical  Research  & 
Education.  295,  1513-1516. 

Denisco.  R.A..  Drummond.  J.N..  &  Gravenstein.  J.S.  (1987).  The  effect  of  fatigue  on  the 
performance  of  a  simulated  anesthetic  monitoring  task.  Journal  of  Clinical 
Monitoring.  3.  22-24. 

Eaton.  P.  &  Gottselig.  S.  (1980).  Effects  of  longer  hours,  shorter  week  for  intensive  care 
nurses.  Dimensions  in  Health  Sen’ice ,  57.  (8).  25-27. 

Freidmaa  R.C.,  Bigger,  J.T..  &  Komfeld.  D.S.  (1971).  The  intern  and  sleep  loss.  New' 
England  Journal  of  Medicine.  285.  201-203. 

Gaba.  D  M.  &  Howard.  S.K.  (2002).  Fatigue  among  clinicians  and  the  safety  of  patients. 
New1  England  Journal  of  Medicine.  347.  1249-1255. 

Gadbois.  C.  (1981).  Women  on  night  shift:  Interdependence  of  sleep  and  off-the-job 

activities.  In:  Night  and  Shift  Work:  Biological  and  Social  Aspects.  Proceedings 
of  the  5th  International  Symposium  on  Night  and  Shift  Work.  Oxford:  Pergamon 
Press. 

Goldman.  L.I..  McDonough,  M.T..  &  Rosemond.  G.P.  (1972).  Stresses  affecting  surgical 
performance  and  learning.  I:  Correlation  of  heart  rate,  electrocardiogram,  and 


Washington  Academy  of  Sciences 


59 


operation  simultaneously  recorded  on  videotapes.  Journal  of  Surgical  Research , 
12.  83-86. 

Gottlieb.  D.J.,  Parenti.  C  M..  Peterson.  C.A..  &  Lofgren.  R.P.  (1991).  Effect  of  change  in 
housestaff  work  schedule  on  resource  utilization  and  patient  care.  Archives  of 
Internal  Medicine.  151,  2065-2070. 

Green,  M.J.  (1995).  What  if  anything  is  wrong  with  residency  overwork?  Annals  of 
Internal  Medicine .  1 23.  5 1 2-5 1 7. 

Greenberg.  M.S.  (1997).  Handbook  of  Neurosurgery.  Vol.  4.  Lakeland.  FL:  Greenberg 
Graphics,  Inc. 

Hancock.  P.A.  &  Desmond.  P.A.  (Ed.).  (2001).  Stress,  workload,  and  fatigue.  Mahwah. 
NJ :  Lawrence  Erlbaum  Associates.  Publishers. 

Harma,  M.,  Ilmarinen,  I,  &  Knauth.  P.  (1988).  Physical  fitness  and  other  individual 
factors  relating  to  the  shift  work  tolerance  of  women.  Chronobiology 
International.  5,  417-424. 

Hart.  R.P..  Buchsbaum,  D.G.,  Wade.  J.B.,  Hamer.  R.M..  &  Kwentuus,  J.A.  (1987).  Effect 
of  sleep  deprivation  on  first-year  residents'  response  times,  memory,  and  mood. 
Journal  of  Medical  Education.  62.  940-942. 

Hawkins.  M  R..  Vichick,  D.A..  Silsby.  H.D..  Kruzich.  &  D.J..  Butler.  R.  (1985).  Sleep 
and  nutritional  deprivation  and  performance  of  house  officers.  Journal  of 
A ledical  Education.  60.  530-535 . 

Jacques.  C.H..  Lynch.  J.C..  &  Samkoff.  J.S.  (1990).  The  effects  of  sleep  loss  on 

cognitive  performance  of  resident  physicians.  Journal  of  Family  Practice.  30. 
223-229. 

Kent.  L.A.  (1972).  The  4-40  workweek  on  trial.  American  Journal  of  Nursing.  72,  683- 
686. 

Kohn.  L.T..  Corrigan.  J.M..  &  Donaldson,  M.S.  (Eds  ).  (1999).  To  err  is  human: 

Building  a  safer  health  system.  Washington.  DC:  Institute  of  Medicine. 

National  Academy  of  Sciences  Press. 

Krueger,  G.P.  (1994).  Fatigue,  performance  and  medical  error.  In:  M.S.  Bogner(Ed.). 
Human  error  in  medicine  (Chapter  14,  pp.  311-326).  Hillsdale.  NJ:  Lawrence 
Erlbaum  Associates.  Publishers. 

Krueger,  G.P.  (1989).  Sustained  work,  fatigue,  sleep  loss  and  performance:  A  review  of 
the  issues.  Work  and  Stress.  3,  (2),  1 29- 141. 

Li.  J..  Tabor.  R..  &  Martinez.  M.  (2000).  Surv  ey  of  moonlighting  practices  and  work 
requirements  of  emergency  medicine  residents.  American  Journal  of 
Emergency  Medicine.  18.  147-151. 

Lingenfelser.  T.H..  Kaschel.  R.,  Weber,  A.,  Zaiser-Kaschel,  H..  Jabober.  B..  &  Kuper,  J. 
(1994).  Young  hospital  doctors  after  night  duty:  Their  task-specific  cognitive 
status  and  emotional  condition.  Medical  Education.  28.  566-572. 

Liskowsky.  D.R.  (Ed.).  (1991).  Biological  rhythms:  Implications  for  the  worker  (U.S. 
Congress.  Office  of  Technology  Assessment.  Report  No.  OTA-BA-463). 
Washington.  DC:  U.S.  Government  Printing  Office. 

Majidian.  A  M..  Brinker.  M  R..  Rice.  J.C.  &  Kerstcin.  M.D.  (1993).  Moonlighting  in  a 
university  surgical  training  program.  Southern  Medical  Journal.  86.  441-446. 

McCue.  J.D..  Janiszewski.  M.  &  Stickley.  W.T.  (1990).  Residents'  views  of  the  value  of 
moonlighting.  Archives  of  Internal  Medicine.  150.  1511-1513. 

Mills.  M  E..  Arnold.  B..  &  Wood.  C  M.  (1983).  Core-12:  A  controlled  study  of  the 
impact  of  12-hour  scheduling.  Nursing  Research.  32.  356-361. 


Summer  2006 


60 


Nelson.  B.J.  &  Blasdell.  A.L.  (1988).  Comparing  quality  on  eight-  and  twelve-hour 
shifts.  Nursing  Management.  19.  (11).  64A-64H. 

Nelson,  C.S..  Dell' Angela.  K..  Jellish.  W.S..  Brown.  I.E..  &  Skaiedoff.  M.  (1995). 

Residents'  performance  before  and  after  night  call  as  evaluated  by  an  indicator 
of  creative  thought.  Journal  of  the  American  Osteopathic  Association.  95.  600- 
603. 

Nelson.  F.  &  Henry.  W.  (1978).  Psychosocial  factors  seen  as  problems  by  family 

practice  residents  and  their  spouses.  Journal  of  Family  Practice.  6.  581-589. 
Orton.  D.I.  &  Gruzelier.  J.H.  (1989).  Adverse  changes  in  mood  and  cognitive 

performance  of  house  officers  after  night  duty .  British  Medical  Journal.  298. 
21-23. 

Poulton.  E.C..  Hunt.  G.M..  Carpenter.  A..  &  Edwards.  R.S.  (1978).  The  performance  of 
junior  hospital  doctors  following  reduced  sleep  and  long  hours  of  work. 

Ergonomics.  21.  279-295. 

Resnick.  R.K..  &  Folse.  J.R.  (1987).  Effect  of  sleep  deprivation  on  the  performance  of 
surgical  residents.  American  Journal  of  Surgery.  154.  520-525. 

Richardson.  G.S..  Wyatt.  J.K.,  Sullivan.  J.P..  Orav.  E.J..  Ward.  A.E..  Wolf.  M.A..  & 

Czeisler.  C.A.  (1996).  Objective  assessment  of  sleep  and  alertness  in  medical 
house  staff  and  the  impact  of  protected  time  for  sleep.  Sleep.  19.  718-726. 
Robbins,  J.  &  Gottlieb.  F.  (1990).  Sleep  deprivation  and  cognitive  testing  in  internal 
medicine  house  staff.  Western  Journal  of  Medicine.  12.  82-86. 

Rubin.  R..  Orris.  P..  Lau.  S.L.,  Hryhorczuk.  D  O..  Fumer.  S..  &  Letz.  R.  (1991). 

Neurobehavioral  effects  of  the  on-call  experience  in  housestaff  physicians. 
Journal  of  Occupational  Medicine,  33.  13-18. 

Samkoff.  J.S.  &  Jacques.  C.H.M.  (1991).  A  review  of  studies  concerning  effects  of  sleep 
deprivation  and  fatigue  on  residents'  performance.  Academic  Medicine.  66. 
687-693. 

Scott,  A.J.  (Ed.)  (1990).  Occupational  medicine:  Shiftwork,  state  of  the  art  reviews. 
Philadelphia.  PA:  Hanley  &  Belfus. 

Smith-Coggins.  R..  Rosekind.  M  R..  Buccino.  K.R..  Dinges.  D.F..  &  Moser.  R.P  (1997). 
Rotating  shiftwork  schedules:  Can  we  enhance  physician  adaptation  to  night 
shifts?  Academic  Emergency  Medicine  4.  95 1  -96 1 . 

Steinbrook.  R.  (2002).  The  debate  over  residents'  work  hours.  New  England  Journal  of 
Medicine.  347.  16.  1296-1302. 

Taffinder.  N.J..  McManus,  I.C.,  Gul.  Y..  Russell.  R.C..  &  Darzi.  A.  (1998).  Effect  of 

sleep  deprivation  on  surgeons'  dexterity  on  laparoscopy  simulator.  The  Lancet. 
352, 1191 

Tasto.  D.L..  Colligan.  M.J.,  Skjei,  E.W.,  et  al.  (1978).  Health  consequences  of  shift  work. 
Cincinnati.  OH.  U  S.  Department  of  Health  and  Human  Serv  ices.  National 
Institute  for  Occupational  Safety  and  Health. 

Thomas.  H.,  Schwartz.  E..  &  Whitehead.  D  C.  (1994).  Eight-  versus  12-hour  shifts: 

Implications  for  emergency  physicians.  Annals  of  Emergency  Medicine.  23. 
1096-1100. 

Walerstein.  S.J.,  Rosner.  F..  &  Wallace.  E.Z.  (1989).  House  staff  stress.  New}  York  State 
Journal  of  Medicine.  89,  454-457. 

Weinstein.  D.  (2002).  Duty  hours  for  resident  physicians:  Tough  choices  for  teaching 
hospitals.  New'  England  Journal  of  Medicine.  347 .  16.  1275-1278. 


Washington  Academy  of  Sciences 


The  Role  of  Academies  of  Science  in  the  Critical 
Examination  of  New  Ideas:  Looking  at  Gaia 

Frank  R.  Haig,  S.J.,  and  Peg  Kay* 


61 


Abstract 

In  science,  new  ideas  have  to  fight  for  acceptance.  The  process  is 
essential  to  ensure  the  founded  confidence  of  the  scientific  community. 
There  is  a  continuum  ranging  from  speculation  at  one  end  through 
theory  to  fact  at  the  other  end.  This  paper  deals  with  the  role  of 
Academies  of  Science  in  encouraging  the  widest  possible  discussion  of 
legitimate  theories.  The  Gaia  Theory  of  the  earth  as  a  self-regulating 
system  is  used  as  an  example  of  the  type  of  controversial  theory  that 
benefits  from  scientific  discussion.  A  description  of  the  Gaia 
Conference  follows  the  paper. 

"A  scientific  truth  does  not  triumph  by  convincing  its  opponents  and 
making  them  see  the  light,  but  rather  because  its  opponents  eventually 
die  and  a  new  generation  grows  up  that  is  familiar  with  it." 

-Max  Planck 


Facts,  Theories,  and  Speculation 

In  science  a  new  idea  has  to  fight  its  way  to  acceptance.  The  path 
may  be  long  and  conflicted.  The  opposition  may  be  intense  and  tortuous. 
The  process,  however,  is  necessary  to  ensure  the  emergence  of  a  founded 
confidence  on  the  part  of  the  broad  scientific  community. 

Perhaps  one  of  the  most  famous  examples  of  the  opposition  a  new 
theory  can  meet  comes  in  the  case  of  Alfred  Wegener  and  his  concept  of 
what  we  now  call  continental  drift  and  plate  tectonics.  Wegener  started 
presenting  his  theory  in  1912.  The  scientific  community  reacted  with 
ridicule  and  derision.  His  personal  treatment  by  colleagues  was  almost 
brutal.  His  work,  however,  is  now  considered  by  many  the  most  important 
element  of  progress  in  the  Earth  sciences  of  the  20th  century.  Nonetheless, 
Wegener  was  never  able  to  obtain  a  university  position  in  his  native 
Germany. 


*Both  authors  are  past  presidents  of  the  Academy 


Summer  2006 


62 


Wegener’s  major  problem  was  that  he  could  propose  no 
mechanism  for  continental  drift.  And  so,  this  theory  was  subjected  to 
intense  criticism.  Perhaps  also  there  was  a  feeling  that  he  was  out  of  his 
field.  After  all,  his  doctorate  was  in  astronomy,  not  geophysics  or  anything 
related. 

Science,  however,  cannot  afford  to  let  ideas  sweep  through  a  field 
the  way  social  fads  do  in  modern  society.  Consider  the  case  of  Trofim 
Lysenko,  a  Russian  biologist  who  became  the  darling  of  Joseph  Stalin.  In 
a  famous  speech  in  1929  Stalin  extolled  practical  scientists  against  the 
more  theoretical  ones  who  joyfully  spent  their  days  studying  fruit  flies 
while  a  famine  raged  all  around  them. 

Lysenko  was  a  former  country  boy  who  could  inspire  peasant 
farmers  who  had  been  largely  turned  off  by  Stalin’s  collectivization 
projects.  Lysenko  did  not  believe  in  careful  bench  work  in  agriculture.  His 
ideas  were  a  mixture  of  Lamarkianism1  and  other  half  examined  notions. 
There  was  no  control  from  the  profession.  In  fact,  he  waged  a  bitter  and 
savage  campaign  against  established  scientists  with  the  happy  support  of 
the  NKVD,  the  then  Soviet  secret  police.  It  took  some  time  after  the  death 
of  Stalin  for  Soviet  science  to  break  free  from  Lysenkoism.  Science 
requires  a  self-discipline  to  remain  out  of  the  clutches  of  charlatans  and 
ideologues.  New  ideas  cannot  and  should  not  expect  to  win  easy  victories. 


THE  ROLE  OF  ACADEMIES  OF  SCIENCE 

Academies  of  science  have  a  special  role  in  the  exposition  and 
critical  examination  of  new  ideas.  They  provide  a  willing  but  intelligent 
audience  to  which  an  innovator  can  make  a  presentation.  In  so  doing, 
academies  do  not  endorse  such  theories.  They  only  allow  them  to  be  easily 
and  widely  exhibited  and  so  begin  their  battle  to  achieve  confirmation  and 
acceptance  or  relegation  to  the  dustbins  of  history. 

Of  course,  academies  of  science  start  with  an  established  concept 
of  what  science  is.  Academies  are  not  in  the  business  of  providing 
platforms  for  mountebanks  and  crack-pots.  There  exists,  therefore,  a  filter 
that  academies  use  to  select  concepts  worthy  of  consideration. 


Washington  Academy  of  Sciences 


63 


WHAT  IS  A  THEORY? 

To  begin  with,  there  exist  certain  known  facts.  Cooling  water 
under  normal  conditions  will  result  in  its  freezing.  Under  normal 
conditions,  a  cubic  foot  of  lead  is  more  massive  than  a  cubic  foot  of 
hydrogen.  In  the  healthy  human  body,  blood  circulates  through  the 
structure.  There  is  a  body  of  statements  not  dependent  on  opinion. 

At  the  other  extreme  there  are  statements  that  are  of  the  order  of 
guesses,  speculations,  initial  conjectures  not  yet  subjected  to  verification. 
A  theory  may  start  in  such  an  environment.  Unfortunately,  many  people 
unacquainted  with  science  take  the  word  “theory”  to  mean  such  untested 
conjectures. 

But  for  scientists  a  theory  must  link  together  facts,  show  their 
interrelationships,  and  present  some  kind  of  a  model  that  makes  the 
situation  intelligible.  Even  more,  it  must  at  some  point  exhibit 
verifications  so  that  the  theory  can  be  confirmed  or  rejected.  Only  then 
does  an  idea  move  to  the  status  of  being  a  theory. 

Every  field  of  developed  science  has  such  fundamental  structures 
that  have  been  hammered  out  over  time  and  through  repeated  critical 
evaluation.  Wegener’s  plate  tectonics,  Mendel’s  genetics,  Newton’s 
explanation  of  the  tides,  Darwin’s  evolution,  big  bang  cosmology,  and 
more.  All  are  subject  to  analysis  and  revision  as  new  data  become 
available  or  new  understandings  emerge. 

A  classical  example  is  the  Michelson-Morley  experiment.  In 
Michelson’s  day.  Maxwell’s  laws  of  the  electromagnetic  field  were  a 
prized  possession.  Maxwell,  however,  used  a  concept  of  the  ether  to  hold 
his  ideas  together.  Michelson  decided  to  measure  the  motion  of  the  earth 
through  that  strange  entity.  His  experiment  was  brilliantly  conceived  and 
meticulously  carried  out.  After  all,  Michelson  was  one  of  the  greatest 
experimentalists  the  human  race  has  ever  produced.  But  the  result  was  a 
null  value.  Ether  could  not  be  shown  to  exist. 

It  is  not  totally  clear  whether  this  experiment  was  the  motivation  of 
the  re-thinking  Einstein  then  produced  in  his  1905  Theory  of  Special 
Relativity.  But  it  could  have  been.  To  this  day  it  is  easy  to  explain  the 
Theory  of  Relativity  by  starting  from  the  Michelson-Morley  experiment. 
Einstein  himself  seems  to  have  been  thinking  in  a  different  direction  in  his 
original  work  in  the  field.  The  point,  however,  is  that  the  Michelson- 
Morley  experiment  and  the  Theory  of  Special  Relativity  exhibited  a 


Summer  2006 


64 


reformulation  of  Newton’s  rich  concepts  of  space  and  time.  New  data  and 
new  insights  can  refashion  even  seemingly  established  theories. 

Constant  and  consistent  verifications  of  Special  Relativity  have 
made  it  an  accepted  part  of  our  model  of  reality,  and  part,  we  think,  of  the 
laws  of  the  universe. 

Of  course,  Einstein  went  on  to  develop  his  ideas  further  and  came 
to  include  acceleration  and  gravity  in  his  perspective  and  so  produced  the 
General  Theory  of  Relativity.  Continuing  and  ever  more  accurate 
measurements  are  part  of  the  life  of  this  theory  and  its  steady  evolution. 

It  is  worth  recalling,  however,  that  Einstein’s  work  did  not 
immediately  win  acceptance.  It  had  to  fight  to  earn  its  place.  Einstein  did 
not  receive  his  Nobel  Prize  in  Physics  for  relativity.  It  was  still  too 
controversial  in  1921.  He  received  the  prize  for  his  work  on  the 
photoelectric  effect  which  won  more  ready  acceptance  although  it,  too, 
had  to  be  verified  by  sets  of  experiments. 

Perhaps  the  great  editor  who  published  Einstein’s  1905  articles  had 
the  best  statement.  That  was  Max  Planck  who  had  himself  achieved  fame 
by  his  work  originating  modern  quantum  mechanics  in  1900.  We  have 
used  his  statement  at  the  head  of  this  article. 

THE  GALA  THEORY 

The  point  of  this  discussion  is  to  treat  a  theory  that  is  controversial 
and  even  hotly  disputed.  How  does  an  academy  of  science  act  in  such  a 
case? 


The  example  is  James  Lovelock’s  approach  to  understanding  life 
and  evolution  -  the  Gaia  theory. 

Lovelock  began  speculating  about  the  possibility  of  the  Earth’s 
being  a  self-regulating  system  in  1965  when  he  was  part  of  NASA’s 
planetary  exploration  team.  He  began  to  formulate  a  hypothesis,  namely 
that  living  organisms  regulate  the  atmosphere  in  their  own  interest.2  He 
discussed  this  hypothesis  with  the  author,  William  Golding  (Lord  of  the 
Flies),  a  discussion  that  resulted  in  Lovelock’s  accepting  Golding’s 
suggestion  that  he  name  his  fledgling  hypothesis  “Gaia”  -  a  result  that  has 
dogged  the  theory  since  its  inception.  Scientists  find  it  very  difficult  to 
take  seriously  a  theory  named  after  a  Greek  Earth  Goddess. 


Washington  Academy  of  Sciences 


65 


Over  time,  the  hypothesis  lost  its  teleological  aspect  and  the  Gaia 
Theory  as  now  set  forth  by  Dr.  Lovelock  and  his  close  collaborator.  Dr. 
Lynn  Margulis,  proposes  that  the  Earth  is  a  self-regulating  system  made 
up  of  physical,  chemical,  biological,  and  human  components.3  As  with  all 
but  the  most  simple  mechanical  systems,  sophisticated  feedback  loops  are 
at  work.  Despite  Lovelock’s  insistence  that  he  never  meant  to  imply  that 
the  earth  was  a  living,  purposeful  organism,  that  he  used  the  term  “living” 
only  in  a  metaphorical  sense,  many  fringe  scientists  adopted  what  is  now 
known  as  the  “strong”  Gaia  theory  -  i.e.,  the  Earth  is  alive  in  the 
biological  sense.  This  silliness  has  become  the  strawman  that  is  often  used 
to  discredit  the  theory.4 

The  more  mainstream  Gaia  has  made  a  number  of  striking 
predictions.  Among  them  are:  that  Mars  would  be  lifeless  (based  on 
atmospheric  evidence  and  confirmed  by  the  Viking  mission  in  1977);  that 
elements  are  transferred  from  the  ocean  to  the  land  by  biogenic  gases 
(supported  by  the  discovery  of  dimethyl  sulphide,  dimethyl  selenide,  and 
methyl  iodide  in  1973  and  2000);  that  climate  is  regulated  through 
biologically  enhanced  rock  weathering  (strengthened  by  the  discovery  that 
microorganisms  greatly  increase  the  rate  of  rock  weathering/') 

Did  all  that  convince  the  scientific  world  that  Gaia  was  a  true 
representation  of  reality?  Not  entirely.  Controversy  still  abounds,  with 
many  scientists  pointing  out  that  theories  other  than  Gaia  could  have  made 
the  same  predictions. 

In  addition,  Stephen  Schneider  observed  that  “Controversy 
occurred  for  at  least  three  reasons:  (1)  there  was  outright  hostility  to  the 
name  ‘Gaia’...;  (2)  there  was  little  or  no  shared  understanding... of  the 
‘Gaia  hypothesis;’  and  (3)  studying  Gaia  required  strong  multidisciplinary 
training  and  an  interdisciplinary  commitment  that  transcended  traditional 
scientific  approaches.”6 

Despite  the  scientific  uneasiness  with  Gaia,  the  first  American 
Geophysical  Union  Chapman  Conference  on  Gaia  was  held  in  1988  to 
discuss  the  possibility  of  active  climatic  regulation  systems  and  the 
relative  importance  of  feedback  processes  between  organic  and  inorganic 
compounds.  This  Conference  was  a  major  factor  in  stimulating 
interdisciplinary  work  as  well  as  introducing  Gaia  to  the  mainstream  of 
scientific  debate.  Many  of  the  results  of  that  debate  were  explicated  in  the 
second  AGU  conference  held  in  2000.  The  papers  given  there  were 
collected  in  Scientists  debate  Gaia1  A  partial  listing  of  the  contents  of  that 


Summer  2006 


66 


volume  illustrates  Gaia’s  success  in  fostering  the  interdisciplinary  research 
that  was  so  formidably  resisted  less  than  two  decades  ago: 


Lynn  Margulis,  “Clarifying  Gaia:  regulation  with  or  without 
natural  selection”; 

Timothy  M.  Lenton,  “Gaia  is  life  in  a  wasteworld  of  byproducts”; 

Tyler  Volk,  “Models  and  geophysiological  hypotheses”; 

J.  Scott  Turner,  “Homeostatic  Gaia:  an  ecologist's  perspective  on 

the  possibility  of  regulation”; 

David  Wilkinson,  “Phosphorus,  a  servant  faithful  to  Gaia? 
Biosphere  remediation  rather  than  regulation; 

Karl  B.  Foyllmi  [el  a/.]  “Self-regulation  of  ocean  composition  by 
the  biosphere”; 

Lee  R.  Kump,  “A  new  biogeochemical  earth  system  model  for  the 
Phanerozoic  Eon”; 

Noam  M.  Bergman,  Timothy  M.  Lenton  and  Andrew  J.  Watson  - 
Gaia  and  glaciation:  Lipalian  (Vendian)  environmental  crisis; 

Mark  A.S.  McMenamin,  “Does  life  drive  disequilibrium  in  the 
biosphere?”; 

K. M.  Nordstrom,  V.K  Gupta  and  T.N.  Chase,  “Food  web 

complexity  enhances  ecological  and  climatic  stability  in  a 
Gaian  ecosystem  model;  and 

Keith  Downing,  “On  causality  and  ice  age  deglaciations”. 

A  description  of  the  next  Gaia  Conference,  to  be  held  in  October 
2006,  follows  the  endnotes  of  this  paper. 

Whether  one  accepts  Gaia  or  not,  it  is  difficult  not  to  admit  that  (1 ) 
without  those  conferences  the  interdisciplinary  work  necessary  to  the 
Earth  sciences  would  not  have  occurred  in  this  timeframe  and  (2)  the  on¬ 
going  debate  has  fostered  greatly  increased  understanding  of  how  life  and 
our  planet  work.  For  these  reasons,  the  Washington  Academy  of  Sciences 
is  pleased  to  co-sponsor  the  Conference  scheduled  for  October  of  this  year 
(see  http://www.gaiatheorv.orq). 

To  repeat  our  earlier  statement: 

Academies  of  science  have  a  special  role  in  the  exposition  and 
critical  examination  of  new  ideas.  They  provide  a  willing  but  intelligent 
audience  where  an  innovator  can  make  a  presentation.  In  so  doing, 
academies  do  not  endorse  such  theories.  They  only  allow  them  to  be  easily 


Washington  Academy  of  Sciences 


67 


and  widely  exhibited.  They  can  then  begin  their  battle  to  achieve 
confirmation  and  acceptance  or  relegation  to  the  annals  of  history. 


End  Notes 

1  A  theory  of  biological  evolution  holding  that  species  evolve  by  the  inheritance  of  traits 

acquired  or  modified  through  the  use  or  disuse  of  body  parts. 

2  Lovelock,  James  "The  Living  Earth",  Nature .  426.  pp.  769-770.  Dec,  2003. 

3  This  statement  is  virtually  identical  with  the  first  bullet  point  of  the  Amsterdam 

Declaration,  issued  by  a  joint  meeting  of  the  International  Geosphere  Biosphere 
Programme,  the  International  Human  Dimensions  Programme  on  Global 
Environental  Change,  the  World  Climate  Research  Programme,  and  the  International 
Biodiversity  Programme  on  July  13.  2001.  It  was  that  Declaration  that  helped  to 
elevate  Gaia  from  the  status  of  hypothesis  to  that  of  a  generally  accepted  theory 
(although  not  necessarily  by  the  name  of  "Gaia”). 

4James  Lovelock.  “Reflections  on  Gaia”,  Scientists  debate  Gaia:  the  next  century  edited 
by  Stephen  H.  Schneider.  Cambridge.  Mass.  MIT  Press.  2004.  p.  2. 

5  op  cit.  "The  Living  Earth.” 

6  Preface  to  Scientists  debate  Gaia,  op.  cit. 

Scientists  debate  Gaia,  op.  cit. 


Summer  2006 


68 


CONFERENCE: 

The  Gaia  Theory  -  Model  and  Metaphor  for  the  21st  Century 

AMONG  THE  SPONSORS  ARE: 

Washington  Academy  of  Sciences 
Northern  Virginia  Regional  Park  Authority 
George  Mason  University, 

ESRI  (Environmental  Systems  Research  Institute), 

Arlington  Public  Schools, 

Arlington  County  Department  of  Environmental  Services, 

Virginia  Tech  Department  of  Science  and  Technology  in  Society, 
Northern  Virginia  Conservation  Trust, 

Arlingtonians  for  a  Clean  Environment, 

Earth  Force, 

My  Organic  Market  (MOM)  Grocers, 

Gaia  International, 

Audubon  Naturalist  Society 


DATE: 

Saturday  and  Sunday,  October  14-15 
LOCATION: 

George  Mason  University  Law  School  -  Arlington,  Virginia 
KEYNOTE  SPEAKER: 

Dr.  Lynn  Margulis,  Distinguished  University  Professor, 

Department  of  Geosciences,  Univ.  of  Massachusetts  - 
Amherst 

Other  speakers  are:  Robert  W.  Corell,  Lee  Kump,  Robert  Artigiani,  Eileen 
Crist,  Lloyd  Pinkham,  Menas  Kafatos,  Tyler  Volk,  Donald  Aitken,  Dick 
Richardson,  James  Strick,  Dan  Zimble,  Michael  Zito,  H.  Bruce  Rinker, 
Oran  Sandel,  Joel  Salatin,  Thomas  I.  Ellis,  Scott  Turner,  J.  Baird  Callicott, 
David  Schwartzman,  Thomas  Lovejoy 

Contact  person:  Martin  Ogle:  Chief  Naturalist 

Northern  Virginia  Regional  Park  Authority 
Potomac  Overlook  Regional  Park 
2845  Marcey  Road,  Arlington,  VA  22207 
potomac@nvrpa.org.  703-528-5406 


Washington  Academy  of  Sciences 


69 


For  registration  and  schedule  see: 


http :  //www .  gai  ath  eory .  org/ 


GOALS  OF  THE  CONFERENCE 

•  To  promote  awareness  and  understanding  of  the  Gaia  Theory 
among  a  diverse  audience  including  scientists,  educators,  policy 
makers  and  the  general  public. 

•  To  explore  the  broad  implications  of  the  Gaia  Theory  and  the 
connections  it  reveals  between  science,  culture,  economics, 
politics,  education  and  other  aspects  of  human  life. 

•  To  explore  &  celebrate  artistic  and  literary  significance  of  the 
metaphor,  Gaia. 

•  To  inspire  the  implementation  of  ongoing  interdisciplinary 
thinking  and  actions. 


Summer  2006 


DELEGATES  TO  THE  WASHINGTON  ACADEMY  OF  SCIENCES 
REPRESENTING  AFFILIATED  SCIENTIFIC  SOCIETIES 


Acoustical  Society  of  America 

Paul  Arveson 

American/Intemational  Association  of  Dental  Research 

J.  Terrell  Hoffeld 

American  Association  of  Physics  Teachers 

Frank  R.  Haig,  S.J. 

American  Ceramics  Society 

VACANT 

American  Fisheries  Society 

Ramona  Schreiber 

American  Institute  of  Aeronautics  and  Astronautics 

David  W.  Brandt 

American  Institute  of  Mining,  Metallurgy  &  Exploration 

Michael  Greeley 

American  Meteorological  Society 

Kenneth  Carey 

American  Nuclear  Society 

Steven  Arndt 

American  Phytopathological  Society 

Kenneth  L.  Deahl 

American  Society  for  Cybernetics 

Stuart  Umpleby 

American  Society  for  Microbiology 

VACANT 

American  Society  of  Civil  Engineers 

Kimberly  Hughes 

American  Society  of  Mechanical  Engineers 

Daniel  J.  Vavrick 

American  Society  of  Plant  Physiology 

Mark  Holland 

Anthropological  Society  of  Washington 

Marilyn  London 

ASM  International 

Toni  Marechaux 

Association  for  Women  in  Science  (AWIS) 

Emanuela  Appetiti 

Association  for  Computing  Machinery 

Lee  Ohringer 

Association  for  Science,  Technology,  and  Innovation 

F.  Douglas  Witherspoon 

Association  of  Information  Technology  Professionals 

Barbara  Saffanek 

Biological  Society  of  Washington 

VACANT 

Botanical  Society  of  Washington 

Alain  Touwaide 

Chemical  Society  of  Washington 

James  J.  Zwolenik 

District  of  Columbia  Institute  of  Chemists 

James  J.  Zwolenik 

District  of  Columbia  Psychology  Association 

David  Williams 

Eastern  Sociological  Society 

Ronald  W.  Mandersheid 

Electrochemical  Society 

Robert  L.  Ruedisueli 

Entomological  Society  of  Washington 

F.  Christian  Thompson 

Geological  Society  of  Washington 

Bob  Schneider 

Historical  Society  of  Washington,  DC 

VACANT 

History  of  Medicine  Society 

Alain  Touwaide 

Human  Factors  and  Ergonomics  Society 

Douglas  Griffith 

Institute  of  Electrical  and  Electronic  Engineers 

Sajjad  Durrani 

Institute  of  Electrical  and  Electronic  Engineers 

Murty  Polavarapu 

Institute  of  Food  Technologies 

Isabel  Walls 

Institute  of  Industrial  Engineers 

Russell  Wooten 

Instrument  Society  of  America 

Hank  Hegner 

Marine  Technology  Society 

Judith  T.  Krauthamer 

Mathematical  Association  of  America 

Sharon  K.  Hauge 

Medical  Society  of  the  District  of  Columbia 

Duane  Taylor 

National  Capital  Astronomers 

Jay  H.  Miller 

National  Geographic  Society 

VACANT 

Optical  Society  of  America 

Jim  Cole 

Pest  Science  Society  of  America 

VACANT 

Philosophical  Society  of  Washington 

Vary  T.  Coates 

Society  of  American  Foresters 

G.  Foster 

Society  of  American  Military  Engineers 

VACANT 

Society  of  Experimental  Biology  and  Medicine 

Darren  Roesch 

Society  of  Manufacturing  Engineers 

VACANT 

Soil  and  Water  Conservation  Society 

Bill  Boyer 

Technology  Transfer  Society 

Clifford  Lanham 

Washington  Evolutionary  Systems  Society 

Jerry  L.R.  Chandler 

Washington  History  of  Science  Club 

Albert  G.  Gluckman 

Washington  Chapter  of  the  Institute  for  Operations 

Research  and  Management  Science 

Russell  Wooten 

Washington  Paint  Technology  Group 

VACANT 

Washington  Society  of  Engineers 

Alvin  Reiner 

Washington  Statistical  Society 

Michael  P.  Cohen 

World  Future  Society 

Russell  Wooten 

Washington  Academy  of  Sciences 
Room  637 

1200  New  York  Ave.  NW 
Washington,  DC  20005 
Return  Postage  Guaranteed 


Medi 
U.S.  i 

P J 

Mem  field.  VA 
Permit  NO.  1 502 


ERNST  MAYR  LIBRARY 
MUSEUM  COMP  ZOOLOGY 
HARVARD  UNIVERSITY 
26  OXFORD  ST 
CAMBRIDGE.  MA  02138-2902 


_ 


<33'eA 


MCZ 

library 

DEC  1  8  2006  Volume  92 

Number  3 

HARVARD  Fall  2006 

UNIVERSITY 

Journal  of  the 

WASHINGTON 
ACADEMY  OF  SCIENCES 


Contents 


The  Editor  Comments  . . . .  . .  i 

Affiliated  Institutions  . . . . . .  i 

Instructions  to  Authors  . .  . . . .  ii 

Luciano  Battocchio,  Mission  Support  to  the  Moon  Explorations  .........................  1 

G.  Giacomelli,  D.E.  Lynch,  F.  Piccolo,  P.Sadler,  C.  Severini,  From  Alaska  to  Moon  Base  ...  11 

P.  Spillantini,  Moon  Base:  Scientific  Opportunities  for  Astroparticle  Physics  ...............  15 

Stefano  Lagrasia  and  Cosimo  La  Rocca,  Positioning  and  Navigation  on  the  Moon  . .  .  29 

P.  Magnani,  B.  Midollini  and  B.  Papalia,  Robotic  Aid  to  Moon  Base  .  . . 61 

Roberto  Varassi,  M.D.  and  Roberto  Revelli,  M.D.,  Jane  and  John  Born  in  Lunar  Jamestown, 
2020  . 69 

Gabriele  Peraldo  Bertinet,  Rodolfo  Guzzi,  Bruno  Ratti,  Anna  Rebecchi,  A  Moon  Base 
Knowledge  and  Imagination  Portal . 75 

News  of  Members  and  Affiliates . 87 


Affiliated  Societies 


Inside  back  cover 


^asJjington  Stabemp  of  Sciences 

Founded  in  1898 


Board  of  Managers 

Elected  Officers 

The  Journal  of  the  Washington  Academy  of 
Sciences 

President 

William  Boyer 

President  Elect 

Alain  Towaide 

Treasurer 

Harvey  Freeman 

Secretary 

James  Cole 

Vice  President,  Administration 

Rex  Klopfenstein 

Vice  President,  Membership 

Thomas  Meylan 

Vice  President,  Junior  Academy 

Paul  L.  Hazan 

Vice  President,  Affiliated  Societies 
Mark  Holland 

Members  at  Large 

Sethanne  Howard 

Donna  Dean 

Frank  Haig,  S.J. 

Jodi  Wesemann 

Vary  Coates 

The  Journal  is  the  official  organ  of  the  Academy. 
It  publishes  articles  on  science  policy,  the  history  of 
science,  critical  reviews,  original  science  research, 
proceedings  of  scholarly  meetings  of  its  Affiliated 
Societies,  and  other  items  of  interest  to  its  members. 
It  is  published  quarterly.  The  last  issue  of  the  year 
contains  a  directory  of  the  current  membership  of 
the  Academy. 

Subscription  Rates 

Members,  fellows,  and  life  members  in  good 
standing  receive  the  Journal  free  of  charge. 
Subscriptions  are  available  on  a  calendar  year  basis, 
payable  in  advance.  Payment  must  be  made  in  U.S. 
currency  at  the  following  rates. 

US  and  Canada  . .  $25.00 

Other  Countries  .  30.00 

Single  Copies  (when  available)  .  10.00 

Claims  for  Missing  Issues 

Claims  must  be  received  within  65  days  of  mailing. 
Claims  will  not  be  allowed  if  non-delivery  was  the 
result  of  failure  to  notify  the  Academy  of  a  change 

Peg  Kay  of  address. 

Past  President:  F.  Douglas  Witherspoon 

Notification  of  Change  of  Address 


AFFILIATED  SOCIETY  DELEGATES: 
Shown  on  back  cover 

Address  changes  should  be  sent  promptly  to  the 
Academy  Office.  Notification  should  contain  both 
old  and  new  addresses  and  zip  codes. 

Editor  of  the  Journal 

Vary  T.  Coates 

Associate  Editors: 

Alain  Touwaide 

Sethanne  Howard 

Elizabeth  Corona 

POSTMASTER: 

Send  address  changes  to  WAS,  Rm.631, 

1200  New  York  Ave.  NW 

Washington,  DC.  20005 

Journal  of  the  Washington  Academy  of  Sciences 
(ISSN  0043-0439) 

Academy  Office 

Washington  Academy  of  Sciences 
Room  63 1 

1200  New  York  Ave.  NW 
Washington,  DC  20005 

Phone:  202/326-8975 
email:  was@washacadsci.org 

Published  by  the  Washington  Academy  of  Sciences 
202/326-8975 

website:  www.washacadsci.org 

MCZ 

LIBRARY 


I 


THE  EDITOR  COMMENTS: 


DEC  1  8  2006 

HARVARD 

UNIVERSITY 


THE  ACADEMY’S  ANNUAL  AWARD  for  Excellence  in 
Physical  Sciences  (May  2006)  went  to  John  C.  Mather  of  the  Goddard 
Space  Flight  Center;  he  had  also  been  a  featured  speaker  at  our  CapSci  06 
conference  in  March.  On  October  3  the  Nobel  Prize  Committee  confirmed 
our  prescience,  naming  John  Mather  as  co-winner  of  the  2006  prize  in 
Physics  (along  with  George  F.  Smoot  of  the  University  of  California)  for 
identifying  radiation  from  the  Big  Bang.  Congratulations  to  Dr.  Mather, 
and  also  to  the  Academy’s  Awards  Committee  for  their  astute  recognition 
of  outstanding  work. 

A  PROPOSAL  TO  BUILD  a  condominium  of  observatories  on 
the  Moon,  presented  at  the  Academy’s  Capital  Science  2004  conference 
led  to  a  series  of  papers  and  conferences  on  this  exciting  topic,  including  a 
MoonBase  Conference  sponsored  by  the  Academy  in  Washington  in 
March  2005;  a  second  in  Venice,  Italy,  in  May,  2005,  co-sponsored  by  the 
Academy,  the  Italian  National  Academy  of  Astrophysics,  and  High 
Frontier,  Inc.;  and  a  third,  with  the  same  co-sponsors,  in  Washington  in 
October  2005.  Several  papers  by  American  and  British  participants  in 
those  conferences  were  subsequently  published  in  the  Journal.  In  this  issue 
we  are  pleased  to  present  a  series  of  papers  by  Italian  participants  in  the 
conferences,  and  as  always  we  invite  comments  by  readers. 


AFFILIATED  INSTITUTIONS 

The  National  Institute  for  Standards  and  Technology 
Meadowlark  Botanical  Gardens 
The  John  W.  Kluge  Center  of  the  Library  of  Congress 
Potomac  Overlook  Regional  Park 


Fall  2006 


II 


INSTRUCTIONS  TO  AUTHORS 

1 .  Manuscripts  should  be  in  Word  or  WordPerfect,  and  not  pdf. 

2.  They  should  be  5,000  words  or  fewer  (exceptions  may  be  made  by 
the  Editor).If  there  are  7  or  more  graphics,  the  number  of  words 
should  be  reduced. 

3.  Graphics  (photographs,  drawings,  figures,  tables)  must  be  in  black 
and  white  only,  and  should  be  easily  resizable  by  the  editors  to  fit 
the  Journal’s  page  size.  Do  not  wrap  text  around  the  graphics. 

4.  References  (and  bibliography,  if  included)  may  be  in  the  format 
generally  acceptable  for  the  disciplinary  or  professional  field 
represented  by  the  manuscript.  They  must  of  course  be  accurate, 
complete,  and  consistent  in  format  throughout  the  paper. 

5.  Manuscripts  should  be  sent  by  e-mail  as  an  attachment,  or  on  a 
CD,  to  Joumal@washacadsci.oro:  or  directly  to  the  editor, 
vcoates@mac.com.  Hard  copy  can  not  be  accepted. 

6.  Be  sure  to  include  both  an  e-mail  address  and  a  postal  address  for 
the  author  (or  primary  author)  including  title  and  institutional 
affiliation  if  any. 


1 


MISSION  SUPPORT  TO  THE  MOON  EXPLORATIONS 


Luciano  Battocchio 


Abstract 

A  mission  to  the  Moon  will  require  extensive  technical  and  logistical 
support  from  Earth.  Technical  support  will  differ  from  current  forms 
since  it  must  support  real  time  decisions  made  on  the  Moon.  Logistic 
activities  must  from  the  start  be  based  on  Life  Cycle  Cost  Analyses  in 
order  to  define  appropriate  maintenance  and  spares  policies. 


Introduction 

A  mission  to  the  Moon  requires  extensive  support  from  Ground 
Centers.  The  astronauts  on  the  Moon  will  have  to  face  an  unknown 
environment  and  will  depend  (at  least  for  the  first  missions)  on  the 
resources  that  Mother  Earth  will  provide  them. 

These  resources  will  cover  the  elements  absolutely  necessary  for 
survival,  but  also  those  items  that  will  guarantee  proper  functioning  of  the 
equipment  necessary  to  survive  and  to  perform  mission  activities,  in 
primis.  For  Moon  exploration,  in  other  words.  Ground  Base  has  to  provide 
consumables,  like  air,  water  and  food,  and  hardware,  like  spare  parts. 
These  elements  will  not  cover  all  the  support  that  the  Moon  explorers 
need.  Additional  support  that  is  also  extremely  important  is  based  on 
information,  which  could  be  technical  or  geographical,  including 
information  on  the  functioning  of  the  equipment  or  related  to  the  Moon’s 
surface. 

In  other  words,  the  support  provided  by  the  Ground  Base  will 
include  information  as  well  as  hardware  or  consumables,  and  the  support 
will  be,  in  general,  of  two  kinds: 

•  Technical  support,  and 

•  Logistic  support. 


Fall  2006 


2 


Technical  Support 

Technical  Support  to  a  Manned  Space  Mission 

Technical  support  to  a  manned  space  mission  is  based  on  provision  to 
Mission  Control  of  technical  information,  including  a  set  of  data  related  to 
the  functioning  of  the  various  systems,  subsystems  and  equipment,  not 
only  during  their  operations,  but  also  operational  procedures,  data 
resulting  from  trend  analyses,  etc. 

Technical  support  could  be  divided  into  three  phases  based  on  mission 
timing: 

•  Pre-mission  analyses; 

•  Technical  support  during  the  mission; 

•  Post-mission  technical  support. 

Pre-mission  analyses:  During  the  definition  of  a  mission,  it  is 
necessary  to  evaluate  the  performance  of  the  various  systems,  subsystems 
and  equipment,  given  the  mission  constraints,  in  order  to  verify  their 
capability  to  support  applicable  mission  steps. 

In  order  to  perform  this  evaluation,  it  is  necessary  to  conduct 
analyses,  simulations  and  tests  based  on  the  available  models  of  the 
systems,  subsystems  and  equipment.  It  is  mandatory  that  these  models  be 
updated  based  on  the  evolution  of  the  system,  subsystems  and  equipment 
configuration. 

Some  of  the  most  important  activities  identified  are: 

•  Inputs,  review  and  assessment  of  the  mission  plan,  mission  rules 

and  mission  procedures; 

•  Design,  development  and  manufacturing  of  the  modifications 

related  to  the  specific  mission; 

•  Technical  support  to  the  mission  manifest  and  integration  cycle; 

•  Technical  support  to  the  resolution  of  flight  manifest  issues; 

•  Participation  in  the  mission  integration  process; 

•  Mission-specific  software  build-up  and  configuration; 


Washington  Academy  of  Sciences 


3 


•  Dedicated  mission  analyses  and  tests  (structural,  thermal, 
environmental,  electrical,  EMC); 

•  Support  to  system  level  mission  specific  verification; 

•  Support  to  end-to-end  tests  and  integrated  mission  simulations. 

Technical  support  during  the  mission:  Ground  support  to  Mission 
Control  is  also  necessary  during  the  performance  of  mission  activities,  in 
order  to  assess  the  correct  performance  of  the  systems,  subsystems  and 
equipment,  and  in  order  to  support  re-planning  and  anomalies  resolution. 
With  respect  to  the  pre-mission  technical  support,  it  has  to  be  noted  that 
the  most  important  part  of  the  technical  support  is  not  pre-planned  and  has 
to  be  given  in  near  real  time,  being  correlated  with  the  above  mentioned 
anomalies  resolution. 

Also  in  this  case  it  is  possible  to  identify  the  most  important  activities, 
which  are: 

•  Systems,  subsystems  and  equipment  monitoring  and  status; 

•  Assessment; 

•  Re-planning  support  and  flight  products  (flight  notes)  assessment; 

•  Real  time  manifest  change  evaluation; 

•  Support  anomaly  resolution; 

•  Off  line  support  to  real  time  team  for  specific  technical  support, 
flight  products  preparation  (i.e.  S/W  PPL)  and/or  anomaly 
resolution. 

Post-  mission  technical  support:  The  technical  support  given  after  the 
completion  of  a  mission  is  typically  related  to  the  analysis  of  the 
performances,  as  well  as  the  behavior  of  the  operative  parameters,  of  the 
systems,  subsystems  and  equipment  during  the  various  mission  phases. 

In  particular,  the  real  operative  parameters  will  be  analyzed  in  front  of 
the  parameters  obtained  with  the  analytical  prediction,  in  order  to 
understand  if  the  behavior  of  the  systems,  subsystems  and  equipment  was 
as  predicted. 

From  the  results  of  this  first  level  analysis,  trend  analyses  are 
performed,  in  order  to  analytically  define  the  behavior  of  the  systems, 
subsystems  and  equipment  for  the  future  operative  developments  and 
missions. 


Fall  2006 


4 


Based  on  the  results  of  the  analyses,  several  actions  could  derive. 
Some  of  the  most  important  derived  activities  are: 

•  Updating  of  planning,  including  logistic  planning  and  models; 

•  Corrective  actions,  such  as  removal  and  replacement  of 
units/equipment; 

•  Manufacturing  of  new  spare  parts  based  on  new  logistic  needs; 

•  Definition  of  changes  in  the  Flight  Unit  configuration; 

•  Design  and  development  of  new  units; 

•  Updating  of  the  configuration  of  the  various  hardware  models, 
simulate etc. 


Technical  Support  to  Moon  Missions 

In  order  to  properly  understand  the  main  differences  between 
technical  support  to  a  “standard”  manned  space  mission  and  the  Moon 
base,  it  has  to  be  noted  that  the  main  differe  e  is  related  to  the  Mission 
Control  models. 

In  a  “standard”  manned  space  Mission,  the  control  is  al  ken 

by  Earth  and  the  decisions  that  could  be  taken  by  the  crew  are  very 
limited. 


On  the  Moon,  however,  the  explorers  must  have  an  increasing 
responsibility  for  the  control  of  their  activities,  since  it  is  absolutely  clear 
that  they  have  to  face  unknown  situations  and  therefore  pre-planning 
could  be  somehow  useless.  It  is  clear  that  the  amount  of  responsibility 
delegated  to  the  explorers  will  increase  over  time:  the  crew  of  the  first 
missions  will  have  a  limited  autonomy,  while  the  crew  of  the  following 
missions  will  have  more  and  more  autonomy,  until  the  entire  Mission 
Control  responsibility  is  left  to  the  Moon  explorers. 

With  this  scenario  of  increasing  autonomy,  and  of  increasing 
responsibility  of  the  Moon  explorers,  it  is  clear  that  the  role  o  hmcal 
support  will  change,  in  particular  for  those  aspects  related  u.  support 
during  the  mission. 

Also  with  respect  to  the  “mission”  concept,  Moon  missions  have  a 
different  approach.  In  “standard”  manned  space  missions  the  mission 
covers  the  entire  time  from  lift-off  to  return  to  ground.  In  this  case,  it 
could  be  assumed  that  a  “mission”  covers  al  ^  a  specific  set  of  activities  to 


Washington  Academy  of  bo  dices 


5 


be  performed  externally  to  the  Moon  Base.  In  other  words,  a  long 
“mission”  (long  term  permanence  on  the  Moon)  will  be  divided  into 
different  short  “missions,”  based  on  a  set  of  activities  performed  on  the 
Moon,  e  g.  on  a  daily  basis. 

Given  these  conditions,  it  is  clear  that  the  most  important 
modifications  to  technical  support  will  be  in  the  area  of  “technical  support 
during  the  mission.”  As  a  matter  of  fact,  the  technical  support  team  must 
give  real  time  support  to  the  Moon  explorers,  in  order  to  immediately 
support  decisions  that  have  to  be  taken  during  their  activities. 

Progressive  transfer  of  the  Mission  Control  would  therefore  have 
heavy  impacts  on  the  Moon  Base  architecture,  in  particular  in  three 
important  areas: 

•  Communications  Earth-to-Moon.  Real  time  support  from  the  Earth  will 
result  in  heavy  requirements  to  the  Earth-to-Moon  communications 
system. 

•  Communications  Moon-to-Moon:  Progressive  transfer  of  Mission 
Control  to  the  Moon  Base  will  require  a  Moon-to-Moon  communications 
system. 

•  Data  Base  on  the  Moon :  Progressive  transfer  of  Mission  Control  to  the 
Moon  Base  will  require  that  the  Mission  Control  data  base,  including 
relevant  technical  information,  be  progressively  transferred  to  the  Moon 

Logistic  Support  to  a  Manned  Space  Program 

Modem  approaches  to  system/mission  support  are  based  on 
Logistic  Support  Analysis  (LSA)  and  Integrated  Logistic  Support  (ILS) 
models,  developed  by  the  U.S.  Department  of  Defense  (DOD);  such 
models  provide  a  powerful  approach  to  logistic  support  definition, 
planning,  management,  implementation  and  acquisition  by  the  user. 
Although  they  normally  apply  to  a  Defense  context,  such  models  are 
successfully  adopted  by  space  programs,  and  in  particular  manned  space 
programs,  to  effectively  manage  applicable  logistic  support  requirements 
and  implementation. 

In  Europe,  the  European  Cooperation  for  Space  Standardization 
(ECSS)  space  standards  explicitly  refer  to  USA  DOD  MIL-STD-1388-1 A 
(LSA)  and  MIL-STD-1388-2B  (LSA  Record).  The  ECSS  also  endorse  and 
tailor  fundamental  methodologies  issued  by  DOD  to  standardize  specific 


Fall  2006 


6 


logistic  disciplines  which  are  closely  related  to  LSA  and  ILS  activities, 
namely: 

•  RAMT  (Reliability,  Availability,  Maintainability,  Testability); 

•  FMECA  (Failure  Modes  Effect  Criticality  Analysis); 

•  Safety  and  Hazard  Analysis; 

•  Configuration  Management; 

•  PHST  (Package,  Handling,  Storage  and  Transportation). 

Further  studies  and  techniques  must  be  integrated  with  the  above 
mentioned  disciplines,  namely: 

•  Human  Factor  Analysis,  and 

•  Support  Facilities  Analysis. 

Basic  guide-lines  to  effectively  manage  LSA/ILS  processes  are: 

•  They  must  be  initiated  in  the  early  phase  of  the  mission  concept 
definition,  in  order  to  influence  design  concepts  when  they  are  not 
yet  consolidated. 

•  They  must  be  integrated  within  the  system  design  process  to  fully 
achieve  mission  objectives. 

•  They  must  take  into  account  user  needs,  operating  environment, 
constraints,  capabilities  and  resources; 

•  They  must  cover  the  complete  system  life  cycle,  up  to  the  disposal 
phase. 

A  key  point  for  logistic  support  success  is  logistic  management, 
which  must  be  established  in  the  early  phase  of  Mission  design  It  must  be 
based  on  a  continuous  data  acquisition  process  and  correlated  analysis 
phases,  since  LSA/ILS  processes  efficiency  closely  depends  on  the 
capability  to  follow  system  requirements  evolution  and  timely  track 
system  design  upgrades. 

Modifications  Required  by  Moon  Missions 

The  Moon  Base  program  implies  a  complexity  that  mankind  never 
faced  during  past  space  missions.  This  complexity  is  obviously  reflected 


Washington  Academy  of  Sciences 


mi 


7 


in  several  issues  related  to  logistics,  and  in  particular  some  requirements 
increase  their  importance,  such  as: 

•  The  design  must  comply  with  serious  safety  impacts  and  system 
reliability  requirements; 

•  A  reliable  rescue  strategy  must  be  defined  to  react  in  a  timely  way 
to  unpredictable  emergency  conditions,  notwithstanding  the  Moon 
Base’s  distance  from  the  Earth; 

•  An  extensive  approach  is  required  to  properly  select,  prepare, 
manage,  transport,  deploy,  start-up,  and  maintain  the  Moon  Base  in 
working  order,  for  a  long  period  of  time. 

Unfortunately,  the  modification  or  increasing  importance  of  some 
requirements  cannot  cover  the  substantial  modification  necessary  to 
implement  efficient  logistic  support  in  the  Moon  Missions  environment. 

The  Moon  Base  program  then  not  only  requires  a  sophisticated 
system  design,  but  a  specific  additional  effort  must  be  spent  for  logistic 
engineering  and  logistic  support  definition,  development  and  management. 

Complex  analyses  are  required  to  identify  mission  support 
objectives  and  priorities,  to  coherently  define  the  Moon  Base  expected 
independence  (personnel  skill,  workload,  tools  and  spares)  and  external 
support  strategies. 

An  assessment  process  is  required  to  optimize  support  objectives, 
dependability  requirements  and  mission  targets  according  to  economical 
and  technical  constraints. 


The  Logistic  Support  Scenario  for  Moon  Missions 

The  logistic  scenario  associated  with  a  Moon  Mission  is  very 
similar  to  one  associated  with  a  ship  that  is  on  a  mission  far  away  from  the 
coasts.  In  both  the  cases,  it  will  be  strictly  necessary  to  trust  in  proper 
resources,  since  it  will  be  very  difficult  to  get  support  from  the  base. 

This  means  that  the  system  design  must  implement  requirements, 
in  terms  of  safety,  reliability,  etc.,  that  will  guarantee  to  the  system  a  very 
high  probability  of  correct  functioning. 

Extremely  important  is  the  maintenance  policy,  since  it  is  strictly 
associated  with  the  keeping  of  high  reliability  levels  and  with  the 
operational  readiness  of  the  systems. 


Fall  2006 


8 


With  respect  to  the  three  traditional  maintenance  levels 
(organizational,  intermediate  and  depot),  some  special  considerations  are 
necessary. 

As  a  first  approach,  in  u<e  beginning  the  applicable  maintenance 
level  will  be  the  organizational  level,  and  in  particular  the 
removal/replacement  operations  will  be  largely  utilized. 

This  will  immediately  bring  up  a  different  issue,  that  is,  the  policy 
to  be  followed  for  failed  units:  they  cannot  be  disposed  of,  for  obvious 
reasons,  on  the  Moon,  and  have  therefore  to  be  transported  to  Earth  for 
depot  repair  or  disposal,  or  destroyed  during  re-entry. 

With  the  increase  of  the  Moon  Base  autonomy,  it  is  reasonable  to 
assume  that  intermediate  level  maintenance  capability  would  increase, 
permitting  the  repair  of  failed  units  directly  on  the  Moon,  provided  that 
relevant  tools,  test  equipment,  etc.,  are  available  on  the  Moon.  Increase  of 
intermediate  maintenance  capabilities  will  also  allow  the  possibility  of 
performing  on  condition  maintenance. 

The  key  element  to  be  considered  for  the  decision  of  the 
maintenance  policy  is  obviously  the  cost. 

The  decision  to  perform  depot  repair  or  to  dispose  of  a  unit  will  be 
taken  considering  the  recovery  costs  from  the  return  flight  and  repair 
costs,  in  front  of  the  costs  on  a  new  spare  unit.  In  the  same  way,  the  repair 
costs  at  the  intermediate  level  on  the  Moon  have  to  include  also  the  costs 
to  have  available  on  the  Moon  the  tools,  test  equipment,  etc.,  necessary  to 
perform  that  maintenance,  and  have  to  be  verified  in  front  of  the  costs  of 
procuring  a  spare  unit  on  Earth  and  transferring  it  to  the  Moon. 

In  other  words,  the  definition  of  maintenance  and,  in  general, 
logistics  policy  depends  on  one  of  the  most  important  logistics 
engineering  tools:  Life  Cycle  Cost  Analysis. 

Life  Cycle  Cost  Analysis  is  a  logistics  tool  utilized  in  various 
programs,  but  there  will  be  a  significant  difference,  since  costs  of  the  units 
will  be  largely  lower  with  respect  to  transportation  and  storage  costs  on 
the  Moon.  Therefore,  these  models  have  to  be  modified  in  order  to  face 
the  new  situation. 

With  respect  to  other  logistics  issues,  such  as  spares  policies,  the 
approach  will  be  more  traditional;  the  decisions  on  spare  units,  and  spare 
units’  availability  on  the  Moon,  will  be  based  on  the  results  of  safety  and 
reliability  analyses. 


Washington  Academy  of  Sciences 


9 


Exploitation  of  logistic  programs  will  be  based  also  on  software 
tools,  such  as: 

•  Inventory  management, 

•  Maintenance  data  collection, 

•  Configuration  control, 

•  Limited  life  items  monitoring,  etc. 

Also,  in  the  case  of  logistic  support,  these  software  tools  and  the 
associated  database  will  be  resident  on  Earth;  transfer  of  this  database  to 
the  Moon  will  not  be  so  urgent,  even  if  the  Moon  Base  has  increased 
autonomy,  since  the  logistic  support  could  be  organized  from  the  Earth. 


Conclusions 

This  paper  presented  the  support  that  has  to  be  given  from  Earth  to 
the  explorers  of  the  Moon. 

The  identified  support  is  either: 

■  Technical  support,  or 

■  Logistic  support. 

Technical  support  could  be  divided  into  three  phases,  based  on 
mission  timing: 

■  Pre-mission  analyses; 

■  Technical  support  during  the  mission; 

■  Post-mission  technical  support. 

Technical  support  during  the  mission  will  drastically  change  with 
respect  to  present  technical  support,  since  it  will  be  mandatory  to  support 
real  time  decisions. 

With  respect  to  logistic  support,  the  most  important  result  is  that 
all  the  logistic  activities  (LSA,  etc.)  must  be  supported  from  the  beginning 
by  Life  Cycle  Cost  Analyses  in  order  to  define  maintenance  policies, 
spares  policies,  etc. 

In  order  to  accomplish  these  tasks,  several  infrastructures  must  be 
implemented,  such  as  Earth-to-Moon  and  Moon-to-Moon 


Fall  2006 


10 


communications,  and  technical  and  logistics  databases  implemented  in  the 
Moon  Base. 


Acknowledgements: 

This  paper  lias  been  prepared  with  the  support  of: 

C.V.  P.  Forlani,  Defence  General  Staff,  Roma. 
Dr.  M.  Canzonetta,  SETEL  Group.  Roma. 

Dr.  M.  Vescovo.  FASER  Sri.  Torino. 

Dr.  E.  Vittone.  ALTEC.  Torino. 


Washington  Academy  of  Sciences 


11 


From  Alaska  to  Moon  Base 

Prof.  G.  Giacomelli 
(Univ.  Arizona.  Dept,  of  Plant  Sciences) 

D.  E.  Lynch 
(ASRC  AeroSpace) 

F.  Piccolo 

(Aero  Sekur) 

P.  Sadler 
(Sadler  Machine  Co.) 

Prof.  C.  Severini 
(Univ.  Foggia.  Agricultural  Faculty) 

Reference  Scenario 

The  reference  scenario  for  the  Space  Greenhouse  Project  is 
President’s  Bush  “Vision”  presented  on  January  14th,  2005,  which  made 
the  Moon  a  fundamental  step  towards  Solar  System  exploration,  and  a 
potential  base  for  power  production,  observation  and  a  logistic  base  for 
interplanetary  journeys.  Such  a  journey,  or  the  target  of  building  up  a 
permanent  base  on  the  Moon,  requests  synergic  and  coordinated  efforts, 
concentrated  on  enabling  technologies  pointed  out  and  analyzed  in  the 
framework  of  the  first  Moon  Base  workshop. 


In  the  framework  of  the  NASA  exploration  plan,  and  among  primary 
objectives  targeted  by  ESA,  Life  Support  stands  as  one  of  the  most 
interesting  and  promising  topic  for  an  inter- Atlantic  cooperation. 


Fall  2006 


12 


A  two  year’s  journey  makes  re-supply  of  food,  oxygen  and  water 
not  feasible,  mainly  for  economic  reasons.  In  this  light,  a  synergic 
cooperation  between  Italy  and  U.S.  would  prove  very  promising. 
Competences  will  be  aggregated  with  plant  technology  on  the  U.S.  side, 
food  conservation  and  treatment  on  the  Italian  side,  through  the 
development  of  inflatable  equipment  and  infrastructures. 

Moon  Base  Growth  Plant 

Testing  in  an  extreme  environment  will  take  an  important  place  in 
the  future  Development  Plan.  The  extreme  environment  for  the  test  can  be 
Antarctica  or  Alaska;  but  it  will  be  the  first  opportunity  to  test  a  closed 
loop,  controlled  atmosphere  system,  like  the  one  that  will  be  used  in 
Space.  At  first  the  main  objective  will  be  to  replicate  the  terrestrial 
environment  (in  terms  of  temperature,  humidity,  pressure  and  atmosphere 
composition,  plus  water  resources  control)  capable  of  allowing  growth  of 
plants  on  the  lunar  soil,  and  producing  all  the  food  that  will  be  the  primary 
source  of  astronauts’  diet,  through  reasonable  photosyntetic  conditions. 
Food  production  facility  will  be: 

-  light  and  reconfigurable 

-  modular 

-  integrated  in  a  Life  Support  management  system  capable  of 

completely  recycling  air,  water,  and  non  edible  residual  parts  of 
cultivated  plants. 


Inflatable  equipment 

Food  treatment  equipment  will  make  extensive  use  of  the  enabling 
inflatable  technologies  (whose  primary  characteristics  will  be  lightweight, 
stowed  to  deployed  volume  ratio)  that  will  be  a  real  result  of  the 
forthcoming  international  cooperation. 

The  University  of  Arizona  with  its  CEA  Centre  has  been  involved 
for  20  years  in  detailed  research  on  cultivars  selection  and  environment 
parameters  optimization  (light,  temperature,  biological  cycles,  etc.)  with 
the  aim  of  increasing  the  automatic  level  of  the  system.  In  the  recent  past, 
the  University  of  Arizona  has  deployed  a  Test  Chamber  Unit  at  the  U.S. 
South  Pole  Station,  as  a  preliminary  test  bed  before  a  lunar  endeavour. 
Now  is  the  time  for  a  joint  inflatable  growth  chamber  (or  greenhouse). 


Washington  Academy  of  Sciences 


13 


Additional  equipment  that  will  be  fundamental  for  closing  the  loop 
is  the  composter  designed  by  Sadler  Machine  Company  in  Phoenix,  AZ, 
capable  of  transforming  residual  non  edible  parts  of  plants  partly  in 
nutrient  solution,  partly  in  carbon  dioxide,  necessary  for  the  life  and 
growth  of  plants.  The  next  generation  will  be  an  inflatable  composter. 

But  the  most  intriguing  device  is  a  Sadler  Machine  design  thermal 
well,  designed  to  take  water  out  of  the  ice  inside  lunar  regolith  at  high 
lunar  latitudes. 


A  preliminary  demonstrator  has  been  manufactured  and  presented 
at  the  Habitation  2006  workshop  in  Orlando,  FL. 

And  a  brand  new  set  of  devices  for  food  treatment  and 
conservation  will  be: 

•  Inflatable  masher, 

•  Inflatable  blancher,  and 

•  Inflatable  cutter, 

designed  by  the  University  of  Foggia  and  developed  by  Aero  Sekur  to  be 
capable  of  treating  food  products  and  transforming  them  into  rations  for 
astronauts,  to  be  stowed  eventually  in  an  additional  external  stowage  rack 
or  module  of  the  station. 


Fall  2006 


This  page  intentionally  left  blank 


Washington  Academy  of  Sciences 


15 


Moon  Base:  Scientific  Opportunities  for  Astroparticle 

Physics 

P.  Spillantini 

University  and  INFN.  Firenze  (Italy) 

Abstract 

The  establishment  of  a  permanent  base  on  the  Moon  in  a  not  too  far 
future  will  be  an  important  opportunity  for  astronomical  and 
astrophysics  observations.  Since  the  electromagnetic  component  has 
been  already  treated  in  a  previous  workshop,  the  discussion  here  is 
restricted  to  the  opportunities  offered  for  the  observation  of  the  cosmic 
ray  component.  In  this  field  the  most  relevant  open  questions  are:  (1) 
precise  determination  of  the  single  fluxes  of  heavy  nuclei  (including 
actinides):  (2)  rare  nuclei  and  isotopic  separation  up  to  several  tens  of 
GeV/nucleon;  (3)  spectra  of  antiparticles  up  to  the  TeV  region  and  the 
hunt  for  possible  anti-nuclei  at  energies  that  allow  them  to  diffuse 
towards  us:  (4)  elemental  composition  up  to  several  tens  PeV  (well 
over  the  knee):  (5)  Ultra  High  Energy  (UHE)  Cosmic  Rays  (CR)  up  to 
a  few  ZeV.  At  these  high  energies  we  can  conceive  of  an  efficient  UHE 
neutrino  Observatory,  capable  of  opening  a  new  observation  window  to 
the  observation  of  the  space  and  time  edges  of  the  Universe.  A  few 
concepts  of  possible  moon  based  detectors  are  described. 

Introduction 

The  renewed  interest  of  Space  Agencies  and  Industrial 
Associations  in  Lunar  exploration  and  use  offers  the  unique  opportunity  of 
profiting  from  the  facilities  that  will  be  there,  and  allows  us  to  conceive  of 
Moon  based  experiments  for  studying  the  most  important  problems 
presently  faced  in  astronomy  and  astrophysics. 

The  initiative  to  study  possible  facilities  on  the  Moon  was  taken  by 
a  working  group,  promoted  by  High  Frontier,  Inc  (USA)  and  the 
'Solidarieta  e  Sviluppo  Association’  (Italy),  representing  a  group  of 
professionals  working  in  Research  Centers  and  in  Space  Industries.  In 
2003  the  High  Frontier  Inc.,  in  its  final  report  to  the  NASA  Office  of 
Space  Flight,  assessed  the  current  technology  base  and  recommended  a 
comprehensive  program  to  reaffirm  the  US  Human  Space  Flight  Program 
and  reach  a  succinct  goal:  “Columbia,  the  First  Lunar  Base  within  a 
Decade”  [1], 


Fall  2006 


16 


In  2003  a  few  months  in  advance  of  the  US  President’s 
announcement  (in  January  14,  2004)  of  a  major  program  aimed  at  the 
human  exploration  of  the  Solar  System  using  the  Moon  as  a  starting  point, 
the  Promoting  Committee  of  the  MoonBase  initiative  set  a  program  for  an 
intense  study  of  the  problem,  involving  the  main  space  agencies  in 
partnership  with  industries  and  scientific  organizations.  The  program  was 
implemented  in  different  studies.  The  program  and  these  studies  were 
made  public  by  the  International  Conference  “Moon  Base:  A  Challenge 
for  Humanity”,  realized  by  a  series  of  dedicated  workshops.  The  first 
workshop  was  held  in  Venice  in  May  2005,  and  the  second  in  Washington. 

The  US  Presidential  Commission,  set  in  January  2004  just  after  the 
US  President’s  announcement,  recommended  in  its  final  report  [2]  to 
engage  the  scientific  community  in  a  “re-evaluation  of  priorities  to  exploit 
opportunities  created  by  the  space  exploration  vision”.  Endorsing  this 
recommendation,  in  addition  to  the  political  and  technical  themes, 
particular  attention  was  given  by  the  MoonBase  initiative  to  the  possibility 
of  using  the  Moon  for  scientific  observations  in  astronomy  and 
astrophysics.  The  electromagnetic  component  was  treated  in  the  Venice 
workshop  by  the  director  of  the  European  Southern  Observatory  [3],  while 
in  the  Washington  workshop  I  handled  the  particle  component. 


Motivations 

Several  motivations  for  using  the  Moon  as  a  suitable  platform  for 
Cosmic  Ray  (CR)  observations,  as  well  as  for  astronomical  and 
astrophysics  observations,  are  mentioned  in  the  “Lunar  Observatory  for 
Cosmic  Ray  Physics”  [4],  in  response  to  the  Cosmic  Vision  2015-2025 
ESA  call.  I  recall  some  of  them  here: 


•  Cosmic  rays  experiments  are  presently,  and  for  the  next  decade, 
carried  out  on  the  ground  or  in  low  Earth  orbit  (LEO).  Ground-based 
apparatus  can  only  register  the  characteristics  of  the  atmospheric 
shower,  a  greatly  altered  remnant  of  the  primary  CR  (PCR).  Also  even 
with  very  large  equipment  and  using  long-term  observations  they 
cannot  reach  the  extreme  energies  where  PCR  mainly  need  to  be 
studied.  Balloon-  or  space-borne  experiments  can  detect  and 
adequately  study  PCR,  but  they  are  limited  in  size,  weight,  and 
lifetime,  so  that  at  the  highest  energies  the  number  of  detected  events 
is  not  very  large. 


Washington  Academy  of  Sciences 


17 


•  The  presence  of  the  Earth’s  atmosphere  and  magnetic  field  has  an 
overwhelming  impact  on  detecting  PCR,  especially  at  lower  energies. 
The  magnetic  field  prevents  charged  particles  from  reaching  the 
Earth’s  surface,  except  near  the  geomagnetic  poles. 

•  A  lunar  base  that  includes  several  experiments  with  a  common 
calorimeter  can  reach  most  of  the  research  items  in  the  PCR  field 
through  a  comprehensive  approach  from  astrophysics  to  fundamental 
physics.  Using  a  common  calorimeter  as  the  main  component  helps 
also  to  save  costs.  Most  of  its  mass  can  be  provided  by  the  lunar 
facilities  that  will  handle  lunar  resources,  such  as  water,  and  provide 
constructive  elements. 

•  Technical  and  technological  developments  achieved  during  last 
decades  in  particle  detectors,  in  space  systems,  and  in  handling  huge 
quantity  of  data  make  lunar  bases  feasible  and  promise  a  significant 
jump  forward  beyond  the  present  programs,  owing  to  the  important 
discovery  potential  that  is  typical  of  particle  and  astro-particle 
research. 

•  As  was  the  case  in  the  seventies  for  the  Great  Observatories  (Hubble 
telescope,  CGRO,  AXAF,  SIRFT)  in  view  of  the  shuttle  operations,  a 
complete  program  at  the  forefront  of  space  science  and  technology 
should  include  a  set  of  Moon  based  Observatories  to  explore  any 
aspect  of  the  Universe.  A  Moon  based  CR  Observatory  would  be  part 
of  this  plan  to  expand  our  knowledge  to  the  extreme  Universe  at  higher 
energies. 


Possible  observations 

The  main  unsolved  problems  in  the  observation  of  CR  are 
represented  in  Fig  1.  The  vertical  lines  represent  the  energy  limit  that  can 
be  reached  in  next  10-15  years  for  the  corresponding  observations  when 
the  observed  fluxes  and  the  technical  limitations  on  the  Earth  surface  and 
in  the  Low  Earth  Orbit  satellites  are  taken  into  account.  In  the  boxes  the 
present  experiments  are  indicated  by  their  acronyms  at  different  stages  of 
running,  or  planning,  or  designing. 


Fall  2006 


18 


High  Z 

[ENTICE,  ECCO] 


Light 
elements 
and  Isotopes 

[ACE] 


Antiparticles 
and  Antinuclei 

[BESS.PAMELAAMS] 

1 - 


Elemental 

Composition 

[CREAM.  ATIC.  BEAR.  NUCLEON. 
ACCESS?,  INCA?,  PROTON-5?] 


Extreme  Energy  CR 

[AUGER,  EUSO,  TUS.  KLYPVE?,  OWL??] 


Fluxes  of  Cosmic  Rays 


1  particle  per  rn  second 


Knee 

1  particle  per  m^year 


Ilndirect^d^tectipriJEAS 
|  ’[arfa,ys  &  foresee  nee] 

|Ankle 

1  particle  per  km^  year 


*tGeV 


TeV 

•| . i 


|PeV 

i,., |  i. 


EeV 


io17  I  id2 

Energy  (eV) 


Fig.  1  -  Global  energy  spectrum  of  primary  CR  -  The  main  unsolved 
problems  are  mentioned  in  the  boxes  [present  experiments  and  projects 
are  indicated  by  their  acronyms]. 


Direct  observations  of  PCR  cannot  reach  the  most  interesting 
energy  region  around  the  knee  at  a  few  PeV.  At  these  energies  and  beyond 
we  can  only  rely  on  indirect  observations  that,  in  spite  of  the  huge  efforts 
and  resources  investments  dedicated  in  the  last  forty  years,  supply  greatly 
altered  information.  Only  when  the  energy  of  the  PCR  reaches  its  extreme, 
beyond  1019  eV,  where  the  florescence  of  the  shower  in  the  mosphere  is 
intense  enough  to  allow  us  to  register  its  longitudinal  development,  does  it 
offer  some  chance  for  the  PCR  identification  and  a  better  determination  of 
its  energy.  However  the  PCR  flux  becomes  so  tiny  that  significant 
statistics  will  hardly  be  reached. 

Let  me  mention  in  more  detail  the  limitations  to  the  CR 
observation  in  the  different  energy  regions  for  the  most  relevant  physics 
problems  mentioned  in  Fig.  1 . 


High! 

By  measuring  the  single  fluxes  of  the  high  Z  (beyond  iron)  and 
ultra-high  Z  (actinides)  nuclei  we  can  learn  about  their  formation  in 
violent  processes,  as  in  supemovae,  and  evaluate  the  rate  of  these 


Washington  Academy  of  Sciences 


19 


processes  in  the  universe.  The  low  abundances  (10'n  of  hydrogen  for  the 
actinides)  require  an  enormous  detection  surface  to  be  deployed  outside  of 
the  atmosphere.  Present  projects  (ENTICE  and  ECCO)  are  still  not 
funded,  and  in  any  case  promise  to  collect  not  more  than  a  few  tens  of 
actinides  in  several  years.  Furthermore,  for  the  registration  of  the  heaviest 
elements  only  passive  techniques  can  be  used  (as  in  the  ECCO  project), 
implying  the  recovery  from  space  of  the  detector  for  analysis.  On  the 
Moon  these  limitations  are  overcome:  many  tens  of  square  meters  of  the 
(relatively  light)  detector  elements  can  be  installed  on  the  lunar  surface 
and  can  be  recovered  for  the  analysis;  furthermore  the  total  absence  of 
magnetic  field  allows  a  significant  increase  (by  one  order  of  magnitude 
compared  with  a  LEO)  of  the  collection  rate.  Many  thousand  of  actinides 
can  be  collected  in  a  few  years.  The  rates  of  nuclei  that  have  different 
decay-times  make  possible,  inter  alias,  the  determination  of  the  rate  of 
their  production  processes. 

Elemental  composition 

Let  now  consider  the  flux  of  the  dominant  components  of  CR,  i.e. 
the  nuclei  that  can  be  synthesized  in  stellar  processes,  from  helium  to  iron. 
The  global  PCR  elemental  composition  and  the  energy  spectra  of  the  most 
abundant  ones  can  be  adequately  studied  only  by  detecting  the  PCR  before 
its  interaction  with  the  terrestrial  atmosphere,  i.e.  in  balloon  borne  or 
satellite  borne  experiments.  The  unsurpassable  mass  limits  of  these 
experiments  and  the  limitations  in  transport  to  space  do  not  allow  one  to 
reach  and  pass  the  so  called  knee  energy  (about  3  x  1015  eV),  where  the 
spectral  index  changes,  and  new  phenomena  must  be  invoked  to  explain 
the  continuation  of  the  energy  spectra  at  higher  energy.  It  is  the  central 
problem  of  the  CR  physics  that  cannot  be  solved  by  detecting  them  on  and 
around  the  Earth.  At  higher  energies  CR  can  be  detected  by  the  shower  of 
particles  they  produce  in  the  atmosphere,  but  the  characteristic  of  the 
initial  particle  cannot  be  easily  extracted.  Only  on  board  a  satellite  as  big 
as  the  Moon,  is  it  possible  to  reach  the  needed  capture  area  of  hundreds  of 
m2sr.  Twenty  years  ago  John  Linsley,  in  his  contribution  to  the  NASA 
workshop  “Future  Astronomical  Observations  on  the  Moon”[5]  suggested 
equipping  light  inflatable  gas  detectors  on  the  thick  roof  of  lunar  shelters. 
We  believe  now  that  on  the  Moon  surface  there  is  water  in  macroscopic 
quantity  (with  concentration  between  1%  and  10%  in  the  regolith  in  the 
polar  craters,  if  the  hydrogen  signatures  seen  by  the  Clementine  mission 
are  water).  Huge  water  quantities  could  be  extracted  mechanically  or 
thermodynamically.  A  very  efficient  method  by  collecting  water  vapour 


Fall  2006 


20 


from  microwave  heating  of  the  regolith  is  presently  under  study  in  the 
framework  of  the  Moon  Base  initiative.  More  than  1000  tons  of  water  can 
be  collected  by  employing  120  kW  of  electric  power  in  less  than  5  years 
[6],  The  availability  of  large  quantities  of  water  is  surely  a  prerequisite  for 
the  establishment  of  a  permanently  inhabited  Moon  Base  [7].  The 
detection  of  ultra  high  energy  PCR  could  usefully  profit  from  this 
possibility.  Their  charge  could  be  measured  on  the  top  of  a  large  water 
volume  equipped  by  Cherenkov  light  sensors  that  could  act  as  a 
calorimeter  and  measure  their  energy.  A  capture  area  of  several  tens  m2sr 
gives  an  observation  rate  that  goes  well  beyond  the  knee  energy  (see  Fig. 
2).  Such  a  measurement  will  finally  allow  us  to  clarify  the  confused 
situation  of  elemental  composition  as  deduced  by  EAS  experiments  on 
Earth  surface. 


Fig.  2  -  PCR  rates  as  a  function  of  the  energy. 

Light  elements  and  isotopes 

Let  consider  now  the  rarest  CR  components.  The  important  and 
well  performing  ACE  device  is  already  operating  at  a  very  large  distance 
from  the  Earth,  outside  the  effect  of  its  magnetic  field.  It  collected  a  large 
amount  of  PCR  up  to  a  few  GeV/nucleon,  and  it  continues  to  work.  The 
quality  of  data  is  excellent  and  the  rare  elements  and  isotopes  ratios  allow 
us  now  to  clarify  many  astrophysics  questions.  It  is  important  to  extend 
this  kind  of  observation  up  to  several  tens  of  GeV/nucleon,  because  the 
evolution  with  energy  of  rare  elements  and  isotopic  ratios  is  very  sensitive 
to  different  models  of  synthesis  in  the  stars,  injection  into  the  Galaxy  and 
of  diffusion  through  the  Galaxy,  and  therefore  sensitive  to  different  stellar 
and  Galaxy  models.  The  most  important  attempt  of  an  experiment 
dedicated  to  this  problem  (LISA  on  the  ASTROMAG  facility)  was  never 
realized  because  of  the  cancellation  of  the  Freedom  Space  Station 
program.  Several  years  later  a  large  area  balloon  borne  device  equipped  by 


Washington  Academy  of  Sciences 


21 


a  superconducting  magnet  spectrometer  (ISOMAX)  was  lost  in  a  flight 
accident.  No  more  projects  are  in  view  for  the  next  10  and  more  years.  The 
(by-product)  data  from  the  BESS-Polar  long  duration  balloon  experiment 
and  from  the  PAMELA  and  AMS  satellite  experiments  will  not  give  a 
final  answer  to  the  main  unsolved  questions.  On  the  Moon  it  will  be 
possible  to  install  detectors  of  several  m2sr  of  geometrical  area,  and  the 
absence  of  magnetic  field  will  allow  us  to  extend  the  spectra  and  isotopic 
ratios  measurements  up  to  several  tens  of  GeV/nucleon,  sufficient  to 
definitively  set  reliable  Nucleosynthesis  and  Galaxy  models. 


Antiparticles  and  antinnclei 

Antiparticles  are  somewhat  rare  CR  components.  For  the 
antiproton  and  positron  elementary  particles  in  addition  to  the  hope  of 
observing  their  secondary  origin  in  interactions  of  particles  with  cosmic 
matter,  is  the  hope  of  observing  in  their  energy  range  effects  such  as  their 
primordial  existence,  or  production  from  steady  sources,  or  signals  of  the 
so  called  new  physics,  that  increase  with  energy  and  became  significant 
beyond  several  hundred  GeV.  The  observed  fluxes  for  their  secondary 
origin  are  enough  high  to  push  the  detection  up  to  several  hundred  GeV, 
the  limitation  being  the  punch-through  of  the  much  more  abundant 
particles  of  opposite  charge  sign  in  the  sample  of  the  detected  events.  The 
PAMELA  and  AMS  experiment  represent  the  maximum  effort  that  can  be 
afforded  on  LEO  experiments,  promising  the  accurate  study  of  the 
antiproton  and  positron  spectra  up  to  this  limit.  It  would  be  important  to 
reach  the  TeV  region  in  this  study  also  considering  that  the  diffusion  time 
from  distant  regions  of  the  Universe  can  be  less  than  the  Universe  age 
only  at  such  high  energies,  and  the  effect  of  the  galactic  wind  in 
preventing  particles  from  entering  our  Galaxy  is  totally  unknown. 
Obviously  this  is  much  truer  for  the  possible  arrival  from  very  distant 
regions  to  our  point  of  observation  of  antinuclei,  an  unambiguous  signal  of 
the  presence  of  antimatter  at  the  astronomical  level.  The  antiproton  rate  for 
different  trends  of  the  antip/p  ratio  is  reported  in  Fig.  3  for  a  capture  area 
of  100  m2sr. 


Fall  2006 


22 


log(Rates/(sr*1 00m2*year)) 


primordial  antip 
source' - 


secondary  antip 


10 


>lne,rflyiIeV) 
100  1000  10000 


Fig.  3  -  Antiproton  rates  for  different  possible  trends  of  the 
antiproton  proton  ratio. 


On  the  Moon  very  rigid  particles  can  be  bent  by  a  relatively  modest 
magnetic  field  extended  over  large  volumes,  which  is  not  allowed  by 
present  launch  capabilities  in  LEO.  Such  a  magnet  could  be  inserted  in  a 
larger  detection  system  devoted  to  the  detection  of  the  tiny  flux  of  ultra 
high  energy  CR.  (Fig.  4) 


5-MOm 


5-MOm 


Direction  and  ionization  measurements 

- 1 

Spectrometers 

identification 

_ 1 

Elc 

ictromagnetic  calorimeter 

Hadronic  c 

calorimeter 

Fig.  4  -  Insertion  of  a  specialized  device  in  the  large  HECR  detection 

system 

However  it  must  be  also  noted  that  the  location  of  a  large  capture 


area  device  on  the  Moon  capable  of  a  very  good  angular  resolution  and 
good  calorimetric  measurement  of  the  released  energy  could  use  the  Earth 
magnetic  field  as  a  magnetic  spectrometer  (see  Fig.5). 


Washington  Academy  of  Sciences 


, 


23 


Fig.  5  -  Deflection  of  a  PCR  passing  at  different  distances  from  the  Earth 

magnetic  axis 

In  the  hypothesis  of  the  existence  of  large  scale  antimatter  domains  in  the 
Universe  (Symmetric  Universe),  an  angular  resolution  better  than  1  mrad 
allows  a  significant  rate  for  antiproton  up  to  about  10  TeV  (Fig.  6). 


Fig.  6  -  Antiproton  rate  for  a  Matter-Antimatter  Symmetric  Universe  in  a 
device  that  uses  the  Earth  magnetic  field  in  the  spectrometer 
Extreme  energy  CR 

At  energies  of  the  primary  CR  exceeding  1017  -  1018  eV  the 
fluoresce  emission  of  the  shower  in  the  terrestrial  atmosphere  becomes 
intense  enough  to  be  detectable  by  suitable  devices.  This  fact  allows  us  not 
only  to  measure  the  total  energy  released  by  the  CR  in  the  shower  but  also 
to  follow  its  longitudinal  development,  giving  information  on  the  nature  of 
the  primary  CR,  and  allowing  us  to  distinguish  the  different  CR 


Fall  2006 


24 


components.  At  high  enough  energies,  exciding  a  few  EeV,  the 
fluorescence  light  emission  is  intense  enough  to  be  observed  and  measured 
at  a  large  distance.  A  huge  air  volume  can  be  monitored  by  a  few  devices, 
especially  if  they  can  be  operated  from  far  away  on  an  Earth  satellite. 

However  the  limitations  in  mass  and  dimensions  of  the  transport 
systems  into  orbit  will  not  allow  us  in  the  foreseeable  future  to  go  very  far 
in  energy,  and  most  of  the  region  beyond  the  GZK  will  be  out  of  reach. 
The  observation  of  the  fluorescence  light  from  a  very  high  altitude 
satellite,  as  the  Moon  is,  could  increase  by  two  orders  of  magnitude  the 
observable  atmospheric  volume,  but,  due  to  the  three  orders  of  magnitude 
of  the  distance  from  the  terrestrial  surface,  it  requires  a  huge  diameter  of 
the  optical  system  to  maintain  a  not  too  high  energy  threshold  for  the 
detection  (see  Fig.  7). 


Fig.  7  -  Rate  of  PCR  for  a  device  installed  on  the  Moon  surface  and  based 
on  the  observation  of  the  florescence  light  emitted  in  the  terrestrial 
atmosphere.  The  evaluation  is  based  on  the  -2.6  value  of  the  index  of  the 
differential  energy  spectrum  and  does  not  take  into  consideration  the  GZK 
effect.  The  energy  threshold  is  indicated  for  several  diameter*  of  the  area 
of  the  optical  system.  The  rates  for  the  EUSO  and  OWL  projects  are  also 
reported  for  comparison 

The  observation  of  Extreme  Energy  Neutrinos 

Optical  systems  with  the  diameters  indicated  in  the  Fig. 7  likely  are 
not  a  goal  for  the  first  generation  of  lunar  experiments,  and  in  any  case 
diameters  of  the  optical  area  exceeding  100  m  cannot  presently  be  easily 
conceived.  However  diameters  in  this  range  could  be  taken  into 
consideration  if  in  the  meantime  the  need  arises  for  an  ‘Extreme  Energy 


Washington  Academy  of  Sciences 


25 


Neutrino  Observatory’;  neutrinos  become  the  fundamental  actor  in  the 
astronomy  of  the  extreme  space  and  time  Universe  and  of  the  extreme 
energy  astrophysics  when  they  are  by-products  of  the  decay  of  the  super¬ 
heavy  elementary  particles  hypothesized  in  the  Grand  Unification 
Theories  of  the  Elementary  Particle  Physics. 

It  must  be  observed  that,  besides  the  Tess  improbable’  cosmogenic 
neutrinos  that  become  relevant  beyond  1019  eV  and  could  still  be  abundant 
beyond  1021  eV  (and  should  be  already  measured  at  the  time  when  a  Moon 
based  Neutrino  Observatory  could  be  planned),  several  models  foresee 
significant  neutrino  fluxes  at  1022  eV  and  beyond.  The  results  from  the 
neutrino  experiments  performed  in  the  meantime  will  indicate  their 
importance  and  will  drive  the  parameters  of  the  possible  Moon  based 
Neutrino  Observatory. 

An  interesting  perspective  for  the  observation  of  extreme  energy 
neutrinos  is  that  proposed  by  the  Lebedev  Institute  of  the  Russian 
Academy  of  Sciences  (LORD  and  LORD  10  projects).  It  is  based  on  the 
detection  by  a  lunar  satellite  of  the  ‘Cherenkov  light’  emitted  (in  radio 
frequencies)  by  the  shower  produced  by  the  neutrino  on  the  limb  of  the 
Moon.  The  monitored  target  volume  increases  with  energy  and  becomes 
competitive  for  energies  beyond  102°  eV  (Fig.  8). 

Such  a  device,  even  if  not  installed  on  the  Moon  surface,  could 
usefully  profit  from  the  facilities  of  a  future  Moon  Base,  and  be 
considered  a  Moon  based  experiment. 


Fall  2006 


Fig.  8  -  Ob  sellable  target  volume  for  different  experiment  -  LORD  and 
LORD  100  projects  are  based  on  the  detection  on  board  of  a  Moon 
satellite  of  the  radio  signal  emitted  (as  Cherenkov  light)  by  the  Ultra  HE 
show  er  produced  by  PCR  on  the  limb  o  f  the  Moon. 

Conclusions 

As  a  conclusion  let  me  present  the  scheme  reported  in  Fig.  9, 
where  the  achievement  that  could  be  obtained  by  Moon  based  CR 
experiments  are  schematically  summarized. 

High  Z  HNeXpiorer (HNX)  jexp  ENTICE  +  ECCO]  r  stand  by 

pcssoe  orvy  on  the  Moon  surface 

Isotopes  (E>GeV/YV)  on  Earth  oft*  =60  are  accessible  txi  no  pians  exist 

iig tt  solcoes  from  BESS  PAMELA  AMS  m  next  years 
nqri  rale  assured  on  the  Moon  up  io  very  heft  E 

Rare  components  art#4/N  upto<10*(AMS) 

artp  up  to  a  >200  GeV  (PA^LA  ed  AMS) 

electrons  up  to  >3  TeV  (PAMELA  A MS  CALET) 

0  TeV  reqen  or,  reach  or  *ne  Moor;  suface 

Elemental  composition  up  to  100TeVbyba#oomng  (gomgon) 

up  to  1  PeV  n  ortxt  (several  projects  and  concepts ) 
up  to  1  CO  PeV  I  weft  behind  the  knee;  or  the  Moor 

Ultra  High  Energies  up  to  fe*  *  IGGEeV  on  Earth  surface  (gotig  on) 
up  to  1000  EeV  tom  ortxt  (but  EUSO  in  stand  by} 
x  :c  a  fe*  '  C  ZeV  fror  tne  Mocr  sjtqcs 
a  JHE  Neutrino  Observatory  1  C,,g  s  feasbte 

Fig.  9  -  Summary'  of possible  achievement  by  future  Moon  based 

experiments. 

It  must  be  emphasized  that  there  are  some  important  measurements 
that  can  be  conducted  on  the  Moon  surface  Each  of  these  measurements 
could  take  advantage  from,  and  be  the  target  of.  a  specific  project  for  a 
dedicated  Moon-based  experimental  facility  However,  the  combination 
of  all  of  them  in  a  single  base  represents  the  very  challenging  and  really 
advanced  program,  because  of  the  synergy  of  different  detection  systems 
and  measurements. 


References 

1  Heiss  K.P..  ^Columbia:  A  Permanent  Lunar  Base".  Final  report  of  High  Frontier  Inc. 
to  NASA  Office  of  Space  Flight.  December  17. 2003 

2  "A  Journey  to  Inspire.  Innovate,  and  Discover",  report  of  the  President's 
Commission  on  Implementation  of  US  Space  Exploration  Policy.  June  2004 


Washington  Academy  of  Sciences 


27 


3  Gilmozzi  R..  “Moon  Base:  Scientific  Opportunities  (in  Astronomy)”.  International 
Workshop  Moon  Base.  Venezia.  May  27.  2005 

4  Pace  E.  and  Spillantini  P..  “Lunar  Observatory  for  Cosmic  Ray  Physics”,  in  response 
to  the  ESA  call  for  ideas  ‘Cosmic  Vision  2015-2025'.  May  30.  2004. 

5  Linslev  J..  “Cosmic  Ray  Detectors  on  the  Moon".  Workshop  of  the  American 
Astronomical  Society  and  of  NASA  on  Future  Astronomical  Observatories  ob  the 
Moon’,  Houstoa  Texas.  January  10.  1986.  proceedings  NASA  Conference 
Publication  2489.  1988.  pag.55. 

6  Heiss  K.P..  Ignatiev  A.  and  Van  Susante  P..  “IRSU-Based  Development  of  a  Lunar 
Water  Astroparticle  Observ  atory”,  report  prepared  for  the  Planetary  and  Terrestrial 
Mining  Sciences  Symposium.  NORCAT.  Sudbury.  ON.  2006. 

7  The  Jamestown  Group  LLC,  “Components  of  an  Economical  Development  Scenario 
for  the  Moon”,  report  in  press. 


Fall  2006 


This  page  intentionally  left  blank 


Washington  Academy  of  Sciences 


29 


POSITIONING  AND  NAVIGATION  ON  THE  MOON 

Stefano  Lagrasta1 

Telespazio  S.p.A. 


Cosimo  La  Rocca 
Galileo  Industries  Italia  S.p.A.; 


ABSTRACT 

The  following  article  provides  a  “tutorial"  overview  of  the  navigation 
problem,  with  details  on  the  related  math  modelling  and  viable 
solutions,  depending  upon  the  available  infrastructure,  translating  the 
experience  gained  on  Earth  to  the  Moon  environment.  To  this  purpose, 
localization  with  Pseudolites  and  Satellite  Constellations  is  considered, 
with  emphasis  on  different  equipment  and  related  application  notes, 
from  the  point  of  view  of  system  design.  Alternatives  are  analyzed  for 
2D  and  3D  determination  of  coordinates.  A  finalization  of  the  study 
should  be  part  of  a  4'phase  A”  activity  specifically  targeted  to  the  Moon 
localization  problem. 


Introduction 

After  over  30  years  of  successful  experience  with  GPS  (Global 
Positioning  System)  and  GLONASS  (GLObal  NAvigation  Satellite 
System),  satellite  radio-assisted  navigation  is  now  at  the  beginning  of  a 
new  era. 

There  is  a  push  originating  with  the  demand  of  increased  accuracy 
and  availability,  not  simply  and  not  only  targeted  to  military  purposes,  but 
involving  new,  growing  communities  -  those  of  civilian  users. 

The  SOL  (Safety  of  Life)  applications,  such  as  air  services  to  assist 
aircraft  approach  and  landing,  needed  for  precision  and  a  certified 
reliability  require  an  implementation  through  the  delivery  of  appropriate, 
real-time  Integrity  information. 

The  first  response  to  these  issues  was  the  introduction  of  Satellite 
Based  Augmentation  Systems  (SBAS),  such  as  the  American  WAAS 
(Wide  Area  Augmentation  System)  and  European  EGNOS  (European 
Geostationary  Navigation  Overlay  System). 

At  the  same  time,  “Selective  Availability”  intentional  degradation 
was  suppressed,  and  a  modernization  program  was  scheduled  for  the 


Fall  2006 


30 


existing  systems.  Twelve  new  satellites  of  the  GPS  HR  block  will  provide 
the  new  civilian  band  “L2C”  and  new  “M  code”  military  signals.  A  third 
GPS  civilian  signal  (L5)  will  be  made  available  after  the  launch  of  the  first 
6  IIF  block  satellites. 

GLONASS  is  also  undergoing  a  complete  renovation  that  is 
expected  to  achieve  a  full  operational  capability  -  based  upon  24  space 
vehicles  -  by  2009. 

Europe  is  developing  the  Galileo  system,  designed  to  be  fully 
compatible  and  interoperable  with  existing  GPS  and  GLONASS,  although 
self-standing.  From  the  beginning  Galileo  will  modulate  10  circular 
polarized  navigation  signals,  using  spread-spectrum  CDMA  (Code 
Divisional  Multiple  Access)  technique  for  multiplation,  on  three  different 
carriers  and  bands:  LI,  E5,  and  E6.  It  plans  to  achieve  excellence  in  user 
positioning  by  allowing  multi -frequency  terminals  to  apply  autonomous 
on-the-fly  ambiguity  resolution  for  “real  time  kinematics”  based  on  carrier 
phase  measurements.  At  the  same  time,  Galileo  will  broadcast  its  own 
Integrity  data,  without  the  need  of  any  complementary  system. 

In  this  scenario,  new  applications  and  technology  enablers  come 
from  the  vision  of  engineers  and  scientists,  able  to  implement  their  own 
solutions. 

In  the  past,  “GPS  like”  signal  generators  were  used  only  on  the 
ground  within  laboratories  to  develop  and  test  navigation  receivers. 
However,  since  early  ‘80,  it  was  understood  that  new  ideas  about  their  use 
had  brought  benefit.  Engineers  started  to  “fix”  signal  generators  outdoors, 
and  experiment  with  the  so  called  pseudo-satellites:  emitters  similar  to  the 
satellite  navigation  payload,  capable  of  allowing  autonomous  positioning 
or  incrementing  availability,  by  complementing  with  “ground  resources” 
the  original  signals  from  constellations.  The  time  for  exploiting  the  so 
called  “pseudolites”  in  all  their  potential  has  just  started  notwithstanding 
the  difficulties  related  to  their  use  (< e.g the  “near  -  far”  problem). 

All  of  this  experience,  as  well  as  technological  and  industrial 
capability,  seem  mature  enough  to  be  exported,  in  order  to  provide  a 
viable  solution  to  the  positioning  problem  on  planets.  The  difference,  with 
respect  to  the  application  on  Earth,  is  the  lack  of  pre-existing 
infrastructures.  We  cannot  forget  that  global  positioning  is  based  upon  the 
concept  of  a  worldwide  applicable  reference  system  and  time  scale;  all  of 
this  must  be  properly  established  in  advance. 


Washington  Academy  of  Sciences 


31 


Fortuitously,  a  satellite  based  navigation  system  is  a  twofold 
“toolset,”  on  one  side,  it  allows  one  to  determine  the  user  coordinates  and, 
on  the  otlher  side,  it  is  a  precise  “measurement  instrument”  to  support  the 
maintenance  of  time  and  global  frame  references. 


Code  and  Carrier  Range  Models  For  a  Navigation  Constellation 

Around  a  Planet 

Let  us  denote  with  1  a  time  instant  on  a  uniform,  reference  time 

scale. 

We  need  a  reference  time  to  mark  the  occurring  events;  as  a  matter 
of  fact,  all  the  equipment  involved  in  the  “navigation  system”  is  made  up 
of  several  clocks,  each  generating  its  own  time  measurement,  T,  each 
different  with  respect  to  the  others. 

The  raw  read-outs  from  a  navigation  receiver  unit  consist  mainly 
of  the  so  called  code  range  (p  )  and  carrier  range  (<j> )  observables.  As  we 
shall  see,  elapsed  time  measurements  are  converted  into  distance,  or 
“ranging”  information. 

sv 

The  ET  (end  terminal)  and  SV  (space  vehicle)  clock  scales  T£r  T 

aim  at  reproducing  the  previously  mentioned  absolute  time  reference  t\ 
however,  they  depend  upon  accuracy  of  local  (receiver  and  navigation 
payload)  clock  oscillators,  and  relativistic  effects. 

A  time  difference  between  instants  in  the  two  scales  7™  T  is  at 

ET 

the  basis  of  the  first  “navigation  observable”  to  be  considered:  the  raw 
code-range,  p . 

p  is  a  length  measurement,  achieved  after  correlating  the  “local 
replica”  of  a  PRN  (Pseudo  Random  Noise)  code  with  the  signal  from  a 
Space  Vehicle  (SV)  or  Pseudolite  (PL)  and  multiplying  the  relative  time 
shift,  needed  to  align  the  codes,  by  the  speed  of  light,  c. 

p  can  be  modeled  as  follows: 

P(7r)  =  c-(7r  -  7^)  +  (high)  noise  +  multipath  (1) 

where  TR  is  the  End  Terminal  (ET)  receiver  clock  time  measurement  when 
p  is  “sampled”;  when  ET  time  reads  TR. t  the  reference  time  is  tR. 


Fall  2006 


32 


Tr=TM-  (2) 

T  is  the  SV  clock  time  measurement  at  the  emission  of  the  RF  wavefront 
that  reaches  the  ET  at  tR,  when  the  navigation  payload  clock  reads  the 
absolute  time  is 

E  SV 

T  =T  (tE).  (3) 

The  difference: 


r=  *R  ~  *E 


(4) 


is  the  effective  “light  travel  time”  of  the  navigation  signal.  One  has: 


T  *E  ^REL-SV +  ^ 
Tr  -  *R  +  A* REL-ET  +  €ET 


(5) 

(6) 


where  e£T  and  £ 1  are  the  “time  errors”  of  the  ET  and  SV  clocks,  while 
A t rel-et  anc^  A ^rel-si'  are  r^tivistic  terms,  due  to  the  fact  that  both 
satellite  and  receiver  are  moving  and  “embedded  “in  a  gravitation  field. 


The  signal  propagation  time  depends  upon  the  real  distance  p 
between  SV  and  ET,  as  well  as  on  delays  due  to  both  the  atmosphere,  if 
existing,  and  the  gravity  gradient.  In  the  language  of  satellite  navigation, 
the  variable  refractive  index  caused  by  the  presence  of  free  electrons 
represents  the  “ionospheric”  delay  (A tJONO);  the  excess  path  caused  by  a 
non-ideal  refractivity  in  the  neutral  atmosphere  is  the  “tropospheric”  effect 
(with  delay  AtTROPO)-  The  gravity  gradient  implies  a  new  relativistic  effect 

with  associated  lag  denoted  as  A t 
One  can  write 

T  ~  ((r  “  *£■)  ~  P I c  +A tREi  +  A tJONO+  At TROPO  (7) 

with  p  =  ||  AA  •  rsl  (tR  -  T)  -  rEJtR)  ||  (8) 

so  that  T  appears  implicitly  defined,  being  in  both  terms  of  the  previous 
equation.  Vectors  r1^,  r_Er  denote  the  respective  positions  of  SV  and  ET 

antennae  in  the  established,  planet  centered  and  fixed  (non-inertial) 
reference  coordinate  system. 

Operator  AA  is  an  “attitude”  matrix,  accounting  for  planet  motion 
over  the  time  interval  r.  In  the  case  of  the  Earth  and  WGS  84  or  GTRF 


Washington  Academy  of  Sciences 


33 


Earth  Centered  and  Fixed  (ECF)  Cartesian  frame,  AA  results,  with  good 
approximation,  to  be  a  “pure”  rotation  about  z-axis,  of  the  angle 
accumulated  by  the  Earth  rotation  in  the  time  interval  r: 


AA  = 


cos(x  •  co0 ) 
-sin(x  •  cd0) 
0 


sin(x-  co0) 
cos(t  •  0)0  ) 
0 


0 

0 

1 


(9) 


It  is  worthwhile  to  establish  a  planet  centered  and  fixed  global  reference 
frame  with  z-axis  close  to  the  effective  rotation  axis  of  the  celestial  body, 
so  that,  in  the  absence  of  significant  precession  and  nutation  motions  over 
the  interval,  r,  equation  (9)  is  still  valid.  In  the  case  of  Earth,  one  has: 

C0o>  0,  (Oo  =  7.29212  x  10  ~  rad/s 

while  in  the  case  of  the  Moon 

(Ocj>  0,  (Oo=  2.66167  x  10”6  rad/s. 

The  Moon  rotates  much  slower,  about  27  times  slower  than  Earth  does. 

Substituting  the  previous  equations  into  (1),  the  following  final 
expression  is  found: 

p(TR)  =  p  +c-(£et-  /')  +...  (10) 


+  C  '  (  ^REL-SV  +  ^ 


REL-ET 


^ REL  + 


^ IONO  + 


^TROPc)  +' 


+  noise  +  multipath  +  eEm 

The  carrier-phase  observable,  (p,  is  the  second  fundamental  raw 
output  measurement  from  a  navigation  receiver.  Given  in  units  of  cycles, 
it  can  be  converted  from  the  very  basic  read-out  into  units  of  length, 
multiplying  it  by  the  wavelength.  A;  it  can  be  demonstrated  that  the 
resulting  model  for  “carrier  range”  5>  is  very  similar  to  the  one  of  code 
range  ,  being  modeled  as  follows: 

0(7^)  =/t-  $  =p  +c-(eET-  (11) 

+  C  ’  (  ^ REL-SV  +  ^ REL-ET  +  ^ REL  ~  ^ IONO  +  ^TROPc)  +  ^  '  N  +. . . 

+  (negligible)  noise  +  (low)  multipath  +  e£pH 

where  N  -  the  so  called  “initial  ambiguity”  -  is  an  integer  term,  with 
relative  sign,  not  a  priori  known,  that  remains  constant,  until  carrier 
tracking  is  lost.  N  equals  the  integer  number  of  wavelengths  along  the 


Fall  2006 


34 


path  between  signal  generator  and  user  receiver,  counted  at  the  time  of 
signal  phase  lock;  such  an  ambiguity  parameter  remains  constant,  until 
any  fail  of  phase  tracking  by  the  receiver. 

Carrier  range  measurements  obtained  from  navigation  signals 
modulating  different  carrier  frequencies  do  have  different  ambiguity 
terms.  When  changing  the  satellite,  a  distinct  ambiguity  value  is  to  be 
accounted  for. 

Note  that  the  ionospheric  effect  appears  in  the  expression  of  <I> 
with  a  negative  sign. 

Both  in  (1)  and  (11),  the  term  £EPIP  known  as  “ephemeris  error”,  is 
not  caused  by  a  physical  source.  It  has  to  be  accounted  for,  whenever  an  a 
priori  model  for  i_  (t)  is  used,  in  order  to  solve  for  user  position  vector, 
r_ET  In  other  words,  eEpH  comes  out  due  to  the  poor  accuracy  in  the 

knowledge  of  the  emitter  location.  This  is  true,  unless  the  problem  under 
consideration  is  the  dual  one,  i.e .,  if  r£T  is  well  known,  and  the  aim  is  to 

nrr 

solve  for  r  ,  as  it  happens  in  the  constellation  Orbit  Determination. 

Solution  of  Navigation  Equations 

We  do  not  put  down  here  all  the  “processing  details”  necessary  to 
solve  for  user  position;  however,  some  basic  equations  are  given,  which 
allow  one  to  understand  the  fundamental  issues  and  problems 

The  standard  solution  provided  by  a  navigation  receiver  uses  code 
range  raw  observables.  The  approach  consists  of  an  iterative  process, 
assuming  that  a  first  “guess”,  is  available  for  ET  coordinates,  as  well  an 

initial  value  £**  for  the  ET  clock  error  and  for  signal  travel  time,  r*. 

Based  on  such  initial  rough  estimates,  and  assuming  one  knows  an 
evaluation  can  be  obtained  for  “signal  emission”  time  instant  as  well  as 

a  better  estimate  for  r we  get 

tE  tR  ~  T  =  Tr  -  £*  -  r*  (12) 

r<-  p/C  -  e.  +  isv(tE).  (13) 

Here  £SI  is  an  estimate  of  the  SV  clock  error,  built  up  by  using  the  “clock 
correction”  engineering  information  transmitted  by  navigation  systems 
within  the  so  called  “navigation  message”. 


Washington  Academy  of  Sciences 


35 


Given  t  the  position  of  navigation  satellite  is  achieved 

through  the  «  precision  ephemeris  »  of  the  navigation  message.  Mainly 
from  r*  and  rSl ,  plus  additional  elements  of  the  navigation  message  ( e.g 
the  « ionospheric  correction  parameters »)  and  meteorological  data 
(pressure,  humidity,  temperature),  the  following  overall  set  of  estimates 
can  be  achieved: 


£  1  ^  REL-SV*  ^  REV  ^  IONCT  ^  TROPO- 

The  aim  is  to  produce  a  “corrected”  code  range  pc ,  i.e.. 

Pc  -  P  +  £  “  C  '  (  ^ REL-SV  +  ^ R£L  +  ^  IONO+  ^  TROPc)  ’ 

Assimilating  At  ET  within  the  clock  error  e  one  has 

PC  =  P  +*'eET  +€ 

where  e  «  conglobates  »  noise,  multipath,  as  well  as  the  uncompensated 
terms,  to  be  intended  as  all  undesired  signal  components  that  do  not  cancel 
out,  after  application  of  the  feed-forward  compensation  formula  ( 1 5). 

Equation  (16)  is  linear  in  the  unknown  receiver  clock  offset  e£V 
whilst  it  is  non-linear  with  respect  to  ET  coordinates,  r£T  . 


(14) 

(15) 

(16) 


A  linearization  about  point  r *  provides 


1  T  A 

p  =  p* - p  Ar 

p*  “* 

(17) 

where 

rET  =  L,  +  Ar 

(18) 

£>,  =  AA  ■  if1  -  rt 

(19) 

P.  =11 A  II  =  II  AA  -r,  ||. 

(20) 

Substituting  (17)  into  (16)  yields: 

Pc  -A=  -T-p/-Ar  +c-eET+e 

(21) 

P* 


where  £>*  is  a  1  x  3  row  vector. 


Fall  2006 


36 


From  ni  distinct  and  “corrected”  code  range  measurements 
{  p J,  pc? p"1 },  all  affected  by  the  same  clock  offset  e£r  one  can  build 
up  a  linear  system 


=  M 


Ar 

C  ■  £et 


+  8 


M  = 


-4  <pDt  • 

pi 

pi  - 


-•(pm)T  i 

.m  'll*  ' 


(22) 


Note  that  M  is  an  m  x  4  matrix;  a  least-squares  solution  for  Ar,  eET  (that  is: 
4  scalar  unknowns),  is  provided  by 


Ar 

C  •  8et 


=  (Mt  •  M)'1  MT  • 


(23) 


assuming  that  m  >  4  ,  and  that  4x4  matrix  (M  •  M)  is  nonsingular. 
After  applying  (23),  letting 

£*  £  ET  (24) 

t*  =  L*+  Ar  (25) 


allows  us  to  start  with  a  new  iteration  of  the  algorithm,  performing  all 
computation  steps  indicated  by  (12),  (13),  (15),  (19),  (20),  and  finally  (23). 

Iterations  are  stopped  when  ||  Ar  ||  comes  out  to  be  of  negligible 
dimension. 


Error  Budget  and  System  Features 

Whatever  the  user  positioning  algorithm  is  that  estimates  r£r  e£T 

from  the  set  of  “corrected”  code  range  measurements  {p^  ,  pc2  p^11 }, 
the  linear  relationship  (23)  is  always  the  proper  one  to  describe  how  the 


Washington  Academy  of  Sciences 


37 


uncompensated  error  £  that  still  affects  observables 
positioning  and  timing  error 

converts  into 

'  Sr  ‘ 

c  •  ST 

=  (Mt  •  M)"1  -Mt  -e 

(26) 

where 

positioning  error 

dr  —  v  —  f 

-  -ET  -ET 

(27) 

timing  error 

a? 

II 

1 

m> 

Let  us  assume  that  an  ellipsoid  is  defined  to  approximate  the 
surface  profile  of  the  planet.  Several  applications  require  the  error  Sr  to  be 
converted  into  “local  coordinates”,  i.e.,  projected  in  a  frame  {xL,yL,  zl) 
co-located  with  the  end  terminal,  the  axes  {xl, yi]  defining  a  plane  tangent 
to  the  ellipsoid  at  the  user  position,  whilst  zl  is  in  the  zenith  direction.  If 
the  orientation  matrix  L  converts  from  planet  centered  to  local 
coordinates,  one  has 


"  §Il  " 

Ol 

Sr 

c  ST 

_0T  1 

c-ST 

(28) 


L  = 


-  sin(?t) 

-  sin((|))  •  cos(^) 
cos((|))  •  cos(^) 


cos(A,)  0 

-  sin((j))  •  sin(A,)  cos(<})) 
cos(<|))  •  sin(^)  sin((j)) 


(29) 


On  Earth,  {A,  (f)}  are  the  well  known  geodetic  longitude  and  latitude  of  the 
user. 


Let  us  assume  now  that  residual  measurement  error  e  is  purely 
stochastic  and  characterized  by  the  covariance:  P  =  E{e  •  £}  =  ap2  •  I  , 
that  means  all  error  components  affecting  code  range  measurements  are 
intended  to  have  the  same  variance,  ap2.  If  this  is  the  case,  the  following  is 
achieved 


C2(xl) 

C2(xl) 

o2(xl) 


o2(T) 


(30) 


Fall  2006 


38 


LT  0 
0T  1 


and  the  “Dilution  of  Precision”  (DOP)  parameters  defined  as: 


GDOP  cp2 

=  (Trace  of  W  )  •  gp2 
=  a\xL)  +  <r(yL)  +  a\zL)  +  g2(7) 

(31) 

TDOP2  cp2 

=  W(4,4)  •  ap2  =  t AT) 

(32) 

PDOP2  •  Cp2 

=  (  GDOP2  -  TDOP2 )  •  Gp2 

=  g2{xl)  +  a2  (yL)  +o2  (zL) 

(33) 

VDOP2  cp2 

=  W(3,3)  •  Gp2  =  g2(z/) 

(34) 

HDOP2  Cp2 

=  (  PDOP2  -  VDOP  y* 

=  g2(xl)  +  G^i) 

(35) 

where  GDOP  means  “Geometric  DOP”,  PDOP  “Position  DOP”,  TDOP 
“Timing  DOP”,  HDOP  “Horizontal  DOP”,  and  VDOP  “Vertical  DOP”. 

Even  if  £*is  not  purely  stochastic  or  Gaussian,  the  DOP  parameters 
provide  a  crucial  indication  on  what  happens  in  a  given  navigation  satellite 
or  pseudolite  configuration.  As  a  matter  of  fact,  DOP  coefficients  imply  an 
amplification  (or  de-amplification,  in  some  lucky  cases)  of  the  residual 
measurement  error  components.  Given  a  budget  for  £  DOP  elements 
explain  what  will  be  the  final  impact  on  positioning  and  timing,  further 
“splitting”  the  positioning  into  horizontal  and  vertical  localization 
accuracies. 

Obtaining  a  limiting  budget  for  gp2  (the  so  called  HERE,  User 
Equivalent  Range  Error)  is  one  aspect  of  the  navigation  mission  study. 

For  instance,  given  a  target,  and  maximum  allowed  3D  positioning 
error  ||  5 r  ||  max,  a  “prudent”  mission  design  suggests: 

PDOP  <  ||  &  _  ||  max  /(3op)  (36) 

As  already  said,  any  configuration  of  satellites  or  pseudolites  must 
be  accompanied  by  a  careful  DOP  analysis  in  the  area  of  service  coverage 
(that  is  the  overall  planet  in  the  case  of  satellite  navigation). 


L  O' 

oT  7 


(Mt-  M)' 


L  0 

\T  1 


(Mt  •  M)-1 


Lt  0 


o2  =  W-o2 


Washington  Academy  of  Sciences 


39 


Laser  Ranging  Techniques 


Laser  Ranging  (LR)  technology  is  based  upon  the  emission  of  a 
pulsed  laser  beam,  transmitted  from  ground  to  a  space  target.  The  returned 
laser  photons  will  be  collected  by  a  telescope,  and  the  timing  between 
issue  and  receipt  of  each  single  laser  pulse  converted  to  an  accurate 
measure  of  the  distance  between  ground  equipment  and  the  space  mirror. 

As  explained  in  [5],  Lunar  Laser  Ranging  (LLR)  was  one  of  the 
first  space  geodetic  techniques,  providing  observations  a  short  time  after 
the  first  manned  mission  to  the  Moon  in  1969  (Apollo  11). 

As  a  matter  of  fact,  Apollo  1 1  and  a  number  of  subsequent 
missions  to  the  Moon  deployed  passive  retro-reflectors  on  its  surface;  LLR 
allows  one  to  perform  ranging  measurements  of  the  distance  between 
points  on  Earth  and  on  lunar  surface. 

In  the  following  picture  from  [6],  the  history  of  precision  attained 
on  measurements  is  shown  over  a  time  period  of  35  years 


Historical  Accuracy  of  Lunar  Laser  Ranging  Data 
Weighted  RMS  Least  Squares  Residual  [cm] 


1970  1975  1980  1985  1990  1995  2000  2005 


Fall  2006 


40 


When  the  laser  signal  returns  after  bouncing  off  of  a  spacecraft  target,  the 
technique  is  called  Satellite  Laser  Ranging  (SLR). 

Telespazio  operates  the  Matera  Laser  Ranging  Observatory 
(MLRO),  a  joint  SLR/LLR  station;  its  telescope  is  shown  in  the  figure 
below.  It  employs  a  1.5  m  astronomical  quality  reflector.  The  laser  is  a 
hybrid  that  produces  a  100  MHz  pulse  train  with  a  pulse  length  that  is  less 
than  50  picosec. 

The  analytical  expression  of  range  measurement  p  is  required,  in 
terms  of  the  elements  to  be  estimated,  to  follow  a  well  assessed  physical 
and  geometrical  model.  Simplifying,  let  x  be  the  vector  listing  the  items  to 
be  determined,  and  p/=  h(x ,  /)  the  model  for  the  ith  two-way  (laser  ray 
round-trip)  distance  measurement. 


Washington  Academy  of  Sciences 


41 


Partial  derivatives  of  the  range  are  computed  relative  to  the 
components  of  x.  Assuming  that  an  initial  guess  x*  is  available  for  the 

unknowns,  the  linearization  of  h(x ,  /')  about  x*  provides 


fi  =  h(x^  /)  +  H 1  •  Ax  , 


H1  = 


dh(  x .  i ) 


dx 


x=x* 


(37) 


so  that,  after  collection  of  a  large  number  m  of  measurements  from 
different  sites  and  at  different  time  instants,  the  correction  Ax  to  x*  is 

obtained  from  pseudo-inversion  of  a  linear  model,  where  the  design  matrix 
M  has  lines  made  up  of  row  elements  {H1,  /  =  1,2,  . m} 


’  p‘ -h(x,l)  " 

H1 

Ax  =  (Mt  M)-1  -Mt  • 

p2  -h(x,2) 

M  = 

H2 

pm  -h(x,m ) 

H'” 

The  observability  of  items  improves  with  variableness  in  the  structure  of 
Jacobian  rows  {H1}. 

LLR  data  analysis,  through  least-squares  solution  of  large  models 
with  respect  to  their  coefficients  correction  “deltas”,  provides  a  number  of 
numeric  parameter  values.  More  precisely,  according  to  [5],  two  groups  of 
parameters  (170  in  total)  are  determined  by  a  weighted  least-squares  fit  of 
the  observations;  the  first  group  comprises,  among  others,  these  five 
parameters: 

-  geocentric  coordinates  of  three  Earth-based  LLR  stations  and 
their  velocities; 

-  a  set  of  Earth  Orientation  Parameters  (EOP); 

-  selenocentric  coordinates  of  used  retro-reflectors; 

-  rotation  of  the  Moon  at  one  initial  epoch  (physical  librations); 

orbit  (position  and  velocity)  of  the  Moon  at  this  epoch. 

Thus,  LLR  contributes  to  the  establishment  of  both  the  Terrestrial  and 
Moon  global  reference  frames  as  they  result  from  the  station  coordinates 
and  velocities;  the  IERS  technical  notes  of  [7]  provide  a  deep  insight  on 
definition  of  the  International  Terrestrial  reference  Frame. 

Compared  to  SLR,  LLR  has  the  advantage  of  following  targets 
with  stable,  highly  accurate  “orbit”  lack  of  non-conservative  forces  from 


Fall  2006 


42 


the  atmosphere,  which  (on  the  contrary)  perturbs  significantly  satellite 
orbits. 

The  definition  of  ground  points  on  the  Moon,  with  respect  to 
selenocentric  coordinates,  is  the  first  step  to  achieve  localization 
capabilities  on  it.  For  instance,  using  LLR,  precise  coordinates  of 
pseudolite  emitters  and  reference  receivers  can  be  achieved,  following  the 
techniques  described  in  the  beginning. 

Complementary  calculation  of  the  lunar  gravitational  field  and  of 
other  parameters  of  physical  interest  can  be  achieved  by  means  of  SLR 
observations  made  from  the  Moon  towards  a  spacecraft  put  in  circumlunar 
orbit. 

Local  Positioning  With  Pseudolites 

Pseudolites  (PL)  are  ground-based  transmitters  broadcasting 
GPS  or  Galileo-like  signals.  In  principle,  pseudolites  may  complement  or 
even  fully  replace  a  constellation  of  satellites  for  radio-assisted  navigation. 
PL  instruments  can  be  easily  monitored,  managed,  and  also  maintained, 
whilst  this  is  not  applicable  for  navigation  payload  on-board  of  space 
vehicles. 

We  will  not  treat  the  specific  RF  problems  coupled  with  using  PL 
equipment,  which  can  be  faced  by  a  proper  design  of  the  single  component 
(PL  or  receiver).  On  the  contrary,  system  design  issues  will  be  addressed. 

It  is  just  recalled  here  that  one  major  impediment  in  the  use  of  PL’s 
is  known  as  the  “near-far”  problem  -  when  the  signal  originates  from  a 
constellation  of  navigation  satellites,  the  average  power  at  the  end  terminal 
site  has  a  relatively  small  range  of  variation.  When  using  Pseudolites  the 
situation  is  completely  different.  When  the  receiver  approaches  the  “near” 
distance  limit  close  to  a  transmitter,  that  PL  causes  an  undesired  jamming 
to  signals  from  the  other  emitters.  When  at  the  “far  limit”,  the  terminal 
captures  a  signal  power  that  stays  just  a  little  bit  above  noise  level. 

The  use  of  pulsed  RF  emission  operation,  with  a  low  duty-cycle 
( e.g .,  10%),  allows  one  to  increase  the  ratio  between  far  and  near 
distances,  up  to  a  “far  limit”  on  the  order  of  tens  of  nautical  miles, 
compared  to  a  “near  limit”  which  reduces  to  below  a  few  hundred  meters. 

In  what  follows,  system  design  issues  will  be  considered, 
depending  upon  the  technology  adopted  to  implement  the  PL;  emphasis 
will  be  given  to  the  core  aspects  of  related  math  models  that  are  explicitly 


Washington  Academy  of  Sciences 


43 


written,  in  order  to  understand  how  to  achieve  the  solution  of  user 
positioning  problem. 


Free-running  Pseudolite  s  (FPL) 

When  the  signal  source  to  the  ET  is  a  pseudolite,  then  in  the  case 
of  a  “free  running”  pseudolite  equipment  (FPL)  the  modeling  of  code 
range  and  carrier  range  observables  is  similar  to  that  one  from  a  navigation 
constellation. 

Assuming  operation  on  the  Moon,  neglecting  atmospheric  delays, 

one  has 


Pet  =  P+c-{eET-  £PL)  +... 

(39) 

+  noise  +  multipath  +  e£pH 

with 

1 

('ll 

(40) 

p|j  denotes  the  code  range  measurement  attained  by  “ET”,  which  is  a 
receiver  (symbol  is  “down”),  when  signal  is  emitted  by  the  “PL”,  that  is  a 
generator  (symbol  is  “up”). 

It  is  immediately  clear  that,  in  order  to  use  the  observable  p£j, 
some  ground  infrastructure  has  to  estimate  and  communicate  the  clock 
offset  / L  to  the  receiver;  it  is  an  unknown  that  cannot  be  autonomously 
solved  for  by  the  ET. 

One  can  also  guess  that,  if  the  pseudolite  “fixed”  position  vector 
is  not  precisely  known,  the  “ephemeris  error”  eEpH  will  be  a  constant 

bias.  This  is  much  worse  than  for  navigation  satellites,  where  e£pH  is  time 
dependent  and  changing  polarity  within  the  validity  window  of  Navigation 
message  (Precise  Ephemeris)  parameters. 

To  implement  a  system  configuration  based  on  FPL  equipment,  a 
relatively  easy  approach  is  based  upon  the  adoption  of  a  “master”  station 
(MS),  capable  of  collecting  code  range  measurements  and  broadcasting 
them  to  the  ET,  via  a  digital  communication  RE  channel.  The  concept  is 
illustrated  in  the  following  figure. 

Measurements  by  ET  are: 

Pet  =  Pet +c' (%“  (41) 


Fall  2006 


44 


with  pJET  =||  £  -  r_E1{lR) 


(42) 


whilst  the  set  of  observables  collected  by  the  MS  receiver  is 

Pms  =  Pms  +  c '  (£ms ~  *4  7=1,2,  .  ..,m  (43) 

with  Pms  =11  Zf  —  LMS  ||.  (44) 

If  now  the  MS  “sends”  the  ET  its  measurements,  the  latter  will  be 
able  to  form  the  m  differences 

Apj  =  Pet  -  Pms-  Pet  ~  Pms  +  c  '  Ae »  j=h2,...,m  (45) 

Af-  ( £et~  £ms )■ 

Assuming  that  the  position  of  MS,  rMS,  and  those  ones  of  all 
pseudolites  {/^,  j=  1,2,  ...,  m)  are  known  a  priori ,  then  all  distances 
{  pJMS,  j=  1,  2,  m}  can  be  calculated. 


Washington  Academy  of  Sciences 


45 


A  linearization  brings  us  to  the  solution  of  the  positioning  problem, 
where  the  unknowns  to  be  determined  are  the  increment  A r£T  to  improve 
an  initialization  value  r*  for  ET  coordinates  and  the  clock  offset  difference 
Ae 


Ap1  -  pi  +Pms 

AP2-P.2+Pms  (46) 

Ap  -p*  +Pms  J 
Synchronies  (SL) 

There  are  several  alternative  implementations  for  a  “pseudolite 
transceiver”,  a  unit  able  to  receive,  process,  and  re-generate  a  navigation 
signal. 

A  first  design,  appropriate  for  application  on  a  planet,  is  depicted 
in  what  follows.  The  equipment  receives  a  navigation  signal  from  a  single, 
“master”  generator  (a  GPS/Galileo  satellite  or  a  FPL),  then  it  re-modulates 
such  a  signal,  with  the  same  carrier,  but  a  different  PRN  code. 

The  ET  will  receive  both  the  “direct”  (master)  source,  as  well  as 
the  output  of  all  synchrolites  (SL). 

The  processing  needed  for  positioning  will  not  foresee  any 
“master”  station.  The  intended  configuration  is  shown  in  the  following 
figure. 

The  “direct”  FPL  signal  generates  the  code  range  observable 

Pet  —  Pet  +  c  '  A^  (47) 

Ae  ={eET-  e). 

It  can  be  shown  that  the  measurement  at  the  ET  side,  from  the  jth  SL,  is 

Pet  =  pj°  +  pJET  +  c-Ae  +  c-  8  ,  j  =  1,  2,  m  (48) 


Af 
c  Ae 


(MT  •  M)'1  -Mt 


with 

Pj  =\\LJ  -  £  II,  7=  1,2,  m 
Pet  =11  £  -  Let  II*  7=  1,2,  ...,  m 


Fall  2006 


(49) 

(50) 


46 


where  {  r',  j  =  0,  1,2,  .  .  ,m)  are  the  “a  priori  known”  positions  of  all 
pseudolites,  that  is,  the  “master”  FPL  and  all  synchrolites. 


The  new  addictive  terms  {8\  j  =  1,  2,  ...,  m}  denote  the  time 
delay  for  the  synchrolite  to  re-transmit  the  incoming  signal. 

The  linearized  model  for  calculating  ET  coordinates  is 


Washington  Academy  of  Sciences 


47 


Ar 

c  •  Ae 


Pet  ~ P?  +  P*  -c  •  81 

=  (Mt  •  M)_l  ■  Mt  •  p|T-p"+p?-c  S2  ,  (51) 


and  it  is  readily  seen  that  estimates  are  assumed  to  be  available  for  each 
SL  operational  delay.  S'.  This  undesired  bias  can  be  effectively  monitored 
and  communicated  to  the  ET,  by  providing  the  SL  of  a  receiver  the  ability 
to  demodulate  its  own  signal  as  seen  in  the  figure  below. 

As  a  matter  of  fact,  if  the  new  receiver  “embedded”  within  the  SL 
is  clocked  by  the  same  timing  source  which  feeds  the  signal  generator,  its 
“self-measurement”  of  generated  signal  reads  out  as 


Pj  =  C-  S'  ,  j=  1,2,  ...,m 


(52) 


and  this  value  can  be  communicated  to  the  ET. 


i 


DIGITAL 

LINK 


Fall  2006 


48 


Differlites  (DL) 

The  term  “Differlite”  was  introduced  by  the  Aerospace  Robotics 
Laboratory  at  Stanford  University  to  describe  a  new  pseudolite  system, 
suitable  for  positioning  on  planets  (Self-Calibrating  Pseudolite  Array 
(SCPA),  see  LeMaster). 

The  DL  belongs  to  the  class  of  pseudolite  transceivers,  in  the  sense 
that  it  comes  equipped  with  a  signal  generator  and  a  receiver. 

In  its  simplest  realization,  the  DL  is  made  ;p  of 
transmitter/receiver  components  that  are  fully  “separate”  (witi  ^erent 
clocks);  the  only  feature  that  couples  the  two  elements  is  the  cat  y  by 
the  receiver  to  demodulate  the  signal  produced  by  the  emitter,  ^ich  is 
materially  a  (simple)  FPL.  Lr  is  consider  the  configuration  that  is  shown 
in  the  following  picture: 


Only  a  couple  of  DL  are  provided;  this  will  not  be  sufficient  to  s  e  for 
ET  coordinates,  but  is  necessary  to  demonstrate  that  DL  units  a  .  to  be 
used  in  “pairs”. 


Washington  Academy  of  Sciences 


49 


Note  that,  assuming  a  separation  between  receiver  and  emitter 
hardware  resources,  two  distinct  clock  offsets  are  to  be  considered  within 


each  unit  and  the  following  primary  equations  hold 

Pet  —  Pet  +  c  '  ( £et~  £ 

Pet  -  Pet  +  c  '  ^£et~  e  ) 

with  pJET  =||  rj  -  rET{tR)  ||,  j  =  1,2  (54) 

where  d ,  d  are  the  clock  errors  of  the  FPL  element  in  each  DL. 

Further,  one  has 

Pi2  =  Pi2+c-(£;-  ?)  ,  Pi1  =c-(et  -  d)  (55) 

pj  r  P|2+c-(e,-f')  ,  p l  =  C  ■  (e2 -  e)  (56) 

where  €  e ,  are  the  clock  errors  of  the  receiver  element  in  each  DL,  and 
where  p,2  =  ||  r1  -  r  ||  (57) 


assuming  (for  the  sake  of  simplicity)  that  the  emitter  and  receiver  antennae 
are  co-located. 

In  each  DL  a  computation  capability  allows  one  to  perform  the 
following  differences  (which  justify  the  term  “Differlites”) 

Ap2  =  Pi2  -  Pi1  s  Pi2  -  c  •  (e  -  e1)  (58) 

AP2  =  P2  -  P2  2  Pi2  +  c  ■  {e  -  d).  (59) 

The  ET  can  calculate  as  well  the  difference 

APet  =  Pet  —  Pet  —  Pet  —  Pet  —  c  —  d).  (60) 

Now,  assume  the  two  Differlites  are  able  to  communicate  the  differences 
(58),  (59)  to  the  ET;  one  sees  that  the  ET  will  be  able  to  extract  the 

unknowns  pf  and  Ae2-1  =  (d  -  d)  from  linear  system 


Api2 

"1  -f 

P? 

=> 

Pi2 

_  1 

1  f 

Ap,2 

_Ap2_ 

_Ap2  _ 

1  1 

c  •  Ae2-1 

c- Ae2-1 

”  2 

-1  1 

(61) 


Fall  2006 


50 


After  detection  of  A£,  given  that  r7,  r  are  known,  the  linearized  model  for 
calculating  ET  coordinates  is: 


Ar  =  (Mt  •  M)"1  -Mt  • 


^Pet  “P*  +  P*  +  c- Ae2-1 
Ap£f*  -  p*  +  p*  +  c  •  Ae4-3 
^Pet5  -  p*  +  p*  +  c  •  Ae6-' 
APet7  -  p*  +  p l  +  c  •  Ae8-7 


(62) 


where  “coupled”  pairs  of  DL  are  considered  to  be  in  each  other’s  “line  of 
sight”,  and  no  “master  station”  (MS)  is  needed;  note  that  a  MS  would 
impose  the  (severe)  requirement  of  receiving  signals  from  all  emitters 
simultaneously. 

Using  the  DL  architecture,  techniques  have  been  developed  as  in 
[11  PI  [3],  so  that  the  array  of  emitters  is  capable  of  “self-surveying”  the 
relative  locations,  creating  a  Self-Calibrating  Pseudolite  Array  (SCPA). 
The  approach  can  be  extended  to  the  carrier  range  measurements  to 
achieve  centimeter  accuracy  [4], 


Sensitivity  of  DOP  Figures  to  Geometry 

When  using  pseudolites  on  the  surface  of  a  planet,  DOP  figures 
that  approach  the  classical  solution  of  user  3D  coordinates  show  dramatic 
singularities. 

As  an  example,  assume  that  four  PL  are  placed  at  the  corners  of  a 
box,  centered  at  Apollo  11  Mission  landing  base:  selenodetic  longitude: 
Ao=  23.5°  East,  latitude:  </>0=  0.7°  North. 

Let  the  box  sides  span  ±0.16°  in  latitude  and  longitude,  with  a  fifth 
PL  placed  exactly  at  the  center  of  the  area.  All  PL  are  assumed  to  be 
placed  at  50  m  of  altitude  above  the  surface  of  the  Moon,  to  compensate 
for  its  curvature  and  to  grant  optical  visibility. 

The  PDOP  figure  for  such  a  configuration  is  shown  below,  where 
black  diamonds  denote  the  antennae  of  signal  generators: 


-L 


Washington  Academy  of  Sciences 


51 


3D  Position  Dilution  of  Precision  (PDOP)  factor 


300 

250 


200 


100 

50 


0 

23.1 


Ion  [deg]  la*  1*41 


Values  are  too  high  and  cannot  be  accepted.  The  fact  is  that  there  is  a 
difficulty  in  evaluating  the  vertical  component  of  the  ET  position. 

To  overcome  it,  one  may  define  a  “local  horizon  fixed”  coordinate 
system,  for  instance  related  to  the  “central”  PL  ,  using  matrix  L  defined  by 
(29)  to  convert  vectors  {r\  j  =  1 ,  ...,  5},  denoting  PL  coordinates,  from 
planet-centered  axes  to  the  local  frame  {*1,  yu  zl}- 

Then,  the  ET  position  can  be  solved  neglecting  the  zL  coordinate 
(approximately  the  altitude,  in  a  little  area),  thus  obtaining  a  2D  solution 
of  the  navigation  problem.  In  this  case,  PDOP  values  are  given  below. 

Within  the  area  having  PL  at  corners,  the  PDOP  value  is  of  the 
order  of  0.95,  which  is  perfect.  An  equivalent  performance  is  obtained  as 
well  by  eliminating  the  5th  PL  emitter  at  the  origin  Ao,  (/>o  of  local 
coordinates. 


Fall  2006 


52 


2D  Position  Dilution  of  Precision  (PDOP)  factor 


lat  [deg] 

1.1-i 

1.0- 

0.9- 

0.8- 

0.7- 

0.6- 

0.5- 

0.4- 

0.3- 
23.1 


— I — 
23.2 


23.8 


23.9 


2D  Position  Dilution  of  Precision  (PDOP)  factor 


— i - 1 - 1 - 1 - 1 - 1 - 1 - 1 - r~ 

23.3  23.4  23.5  23.6  23.7 


Global  Moon  Positioning  With  Constellations 


Ion  [deg] 


Washington  Academy  of  Sciences 


53 


The  basic  equation  set  (22),  with  solution  (23),  is  the  standard 
receiver  approach  to  determine  ET  coordinates  in  geocentric  as  well  as 
selenocentric  fixed  coordinates. 

The  navigation  equations  (22),  (23)  imply  the  need  to  determine 
four  scalar  magnitudes  (the  three  components  of  position  adjustment,  A r, 
and  the  ET  clock  offset  eET). 

A  necessary  condition  is  to  have  ni  >  4  navigation  space  vehicles  in 
view.  To  this  purpose,  several  studies  analyze  the  (minimum)  number  of 
visible  satellites  at  the  nodes  of  a  grid,  over  the  surface  region  of  the  Moon 
that  is  meant  to  be  covered  by  a  “positioning  service”. 

However,  this  is  not  sufficient;  in  fact,  the  solution  of  (23)  must  be 

T 

well  posed  and  the  inversion  of  the  4x4  matrix  (M  •  M)  far  from 
singularities. 

In  other  words,  the  geometric  configuration  of  satellites,  as  seen 
from  the  user  receiver,  has  to  be  favorable,  with  DOP  figures  illustrated  in 
the  beginning  characterized  by  promising  values. 

If  four  satellites  are  in  view,  but  (for  instance)  all  pertaining  to  the 
same  orbit  plane,  the  ET  will  not  be  able  to  solve  for  its  position.  DOP 
values  will  jump  to  very  high  values  in  the  proximity  of  singularity 
conditions. 

One  should  also  take  into  account  a  “masking  angle”  on  elevation 
of  visible  satellites.  On  Earth,  navigation  space  vehicles  that  are  still  in 
view,  but  below  a  minimum  elevation  of  5°  with  respect  to  the  ET  local 
horizon,  are  commonly  excluded  from  positioning  computations,  due  to 
the  large  portion  of  atmosphere  that  is  passed  through  by  signals.  The 
Moon  does  not  exhibit  such  a  huge  variation  of  the  refraction  index; 
however,  a  not-null  masking  angle  is  to  be  considered  when  performing 
“volume”  simulation  of  performances,  to  account  for  natural  obstacles  to 
the  propagation  of  rays.  If  a  “trial”  constellation  appears  sensitive  to  small 
masking  angles  (say  2°),  then  its  design  should  be  re-examined. 

In  order  to  “save”  the  number  of  satellites  needed  for  positioning, 
or  to  overcome  temporary  singularities  for  the  3D  position  determination, 
one  may  assume  a  simplification  and  treat  a  2D  problem  instead. 


Fall  2006 


54 


The  problem  of  imposing  long-term  orbits  is  an  open  issue  as  well, 
due  to  the  intrinsic  instability  of  Moon  orbit  profiles,  which  will  not  be 
treated  here  in  detail. 


2D  Positioning  with  Navigation  Constellations 
This  can  be  accomplished  by  projecting  the  unknown  correction 
Arto  local  horizon  coordinates,  {xL,yL,  -l},  then  neglecting  the  “delta” 
along  zi  and  assuming  one  estimates  only  its  components  about 

Let  us  consider  again  the  basic  equation  set  (22),  and  split  matrix 
M  according  to  the  3  x  1  structure  of  Ar  and  of  the  scalar  (c-  £ET),  so  that 


=  H  Ar 


+ 1  •  (c  •  ) 


+  e, 


H  = 


i  (P*) 
p* 


— L.(p”)T 

m  —  * 

P  * 


Now,  using  the  transpose  of  matrix  L  defined  by  (29) 


Axi 

"1 

0" 

"0" 

T  T 

L  -ArL  =  L  ■ 

<1 

T 

=  L 

0 

1 

Axl' 
Ay  t 

+ 

0 

_Azl  _ 

0 

0 

►  L  _ 

1 

(63) 


(64) 


Azl  =  0  =>  Ar  =  W 


Axl 

Avl 


with 


"1 

0" 

-sin(?i*) 

-cos(^*)sin(<|)*) 

0 

1 

= 

cos(>,* ) 

—  sinf^t^ )  -  sintfj)^ ) 

0 

0 

cos((|)*) 

L, 


(65) 


(66) 


Substituting  (66)  into  (63)  and  rearranging  terms,  one  obtains  the  desired 
2D  formulation  of  the  positioning  problem 


Washington  Academy  of  Sciences 


55 


~1  1 

Pc  ”P* 

~2 

i — 

_! 

<1 

Pc  "P* 

=  M 

Avl 

~  m  _  m 

_c  •  £ex  _ 

|_Pc  “P*  J 

+  £ 


M  = 


pi 


(P*)TW 


-4-(p;)T  w 

p* 


pr 


(Pl”)Tw  1 


(67) 


with  the  solution 


Axl 

Ayl 

=  (Mt  •  M)_l  Mt  • 

Pc  -  pi 

Pc  -P* 

_C-£Et_ 

_Pcm  -P*m. 

(68) 


Walker  Constellations 

After  having  assigned  a  nominal  semi-major  axis,  eccentricity, 
inclination,  common  to  all  space  vehicles,  a  Walker  satellite  constellation 
is  characterized  by  three  (constant)  integer  parameters:  Tl  PI  F,  where 
T=  total  number  of  satellites,  P  =  number  of  orbit  planes,  F=  “phase 
factor”,  with  0  <  F  <  (P  -  1 ) . 


From  Tl  PI  F,  the  following  derived  parameters  are  obtained: 

D  Number  of  satellites  per  plane:  S  =  T I P 

°  Pattern  unit  [deg]:  PU  =360  IT 

a  In-Plane  spacing  angle  [deg]:  IPS  =  PU x  P 

n  Angular  spacing  of  ascending  nodes  [deg]:  NS  =  PU  x  S 

°  Phase  delta  angle  [deg]:  PHD  =  PU x  F 

The  PHD  is  the  angular  distance  of  a  satellite  with  respect  to  the 
ascending  node  of  its  orbit,  evaluated  at  the  time  when  the  companion 
satellite  in  the  next  most  Western  plane  achieves  its  ascending  node. 

A  navigation  constellation  may  be  designed  as  the  “union”  of 
several  distinct  constellations,  each  one  being  of  Walker  type.  For 
instance,  the  Galileo  navigation  system  baseline  is  a  Walker  27/3/1 


Fall  2006 


56 


constellation,  with  an  additional  3/3/1  set  of  “spare”  vehicles,  sharing  the 
same  orbit  planes  of  the  former. 


Halo  Orbits 

There  are  no  stable  Lunar  orbits;  thus,  in  order  to  maintain  a 
spacecraft  in  a  planned  trajectory  path  around  the  Moon  over  a  long  time, 
frequent  station-keeping  maneuvers  are  to  be  executed. 

Fuel  budget  is  a  critical  issue  for  settling  and  keeping  up  a 
constellation  on  the  Moon. 

A  spacecraft  in  the  Earth-Moon  system  is  the  “third  body”  (of 
negligible  mass)  in  a  configuration  with  two  additional  large  primaries. 

The  singularities  of  the  manifold  of  the  states  of  motion  are 
equilibrium  points  for  the  dynamical  system,  named  Lagrangian  or 
libration  points.  A  good  tutorial  is  provided  by  [8]  on  this  topic. 

There  are  three  “collinear”  (LI,  L2  and  L3)  and  two  “triangular” 
(L4  and  L5)  points;  in  the  Earth-Moon  system,  the  Earth  is  the  primary 
with  bigger  mass  and  it  is  possible  to  demonstrate  that  this  implies  the  two 
“interesting”  points  LI  and  L2  are  close  to  the  Moon 

The  possibility  of  taking  advantage  of  the  nature  of  the  libration 
points  for  useful  spacecraft  orbits  has  been  analyzed  in  the  literature.  In 
more  detail,  R.W.  Farquhar  discovered  that  full  3D  periodic  “halo”  orbits 
can  occur  around  LI  and  L2.  He  proposed  the  use  of  a  communication 
satellite  in  a  halo  orbit  about  L2,  then  complementing  it  with  a  second 
relay  satellite,  placed  at  the  cislunar  libration  point  LI. 

The  application  of  halo  orbits  for  navigation  purposes  is  an 
interesting  perspective.  It  has  been  estimated  that  a  cost  of  about  100 
m/s/year  is  enough  to  counteract  the  solar  gravity  force  and  radiation 
pressure  that  tends  to  interrupt  the  periodicity  of  the  halo  orbits. 

The  drawback  is  that  halo  orbits  are  difficult  to  design  and 
implement  with  real  missions;  the  problem  is  highly  non-linear,  so  that 
small  changes  in  the  initial  conditions  compromise  the  possibility  to 
achieve  or  to  maintain  the  desired  orbit  profile,  as  explained  in  [8]. 


Washington  Academy  of  Sciences 


57 


Sample  Lunar  Constellation 

Even  with  the  lack  of  a  rigorous  demonstration,  it  is  seen  that  at 
least  18  satellites  are  needed  to  solve  for  3D  global  positioning  on  the 
Moon,  without  an  augmentation  performed  by  ground  pseudolites. 

There  is  a  number  of  alternative  ways  to  arrange  18  satellites  to 
form  a  navigation  constellation,  aiming  at  covering  both  its  polar  and 
equatorial  regions  with  a  suitable  positioning  service. 

A  proposed  example  of  constellation  with  18  space  vehicles  is 
made  up  of: 

a  polar  Walker  sub-constellation,  with  T=  12/  P  =  3  / F=  2 

an  equatorial  sub-constellation  of  6  equally-spaced  additional 
satellites 

all  with  null  nominal  eccentricity  and  an  orbit  semi-major  axis  about  5 
times  the  Moon  radius. 

The  overall  arrangement  is  shown  in  the  following  picture: 


z 


The  minimum  number  of  visible  satellites  is  shown  hereafter,  assuming  a 
masking  elevation  angle  of  2° 


Fall  2006 


lat  [deg) 


Minimum  number  of  visible  satellites 


0  50  100  150  200  250  300  350 


Ion  [deg] 


The  worst  case  PDOP  figures  are  as  follows 


lat  [deg]  3D  Position  Dilution  of  Precision  (PDOP)  factor 


0  50  100  150  200  250  300  350 


Ion  [deg] 


One  sees  that  there  are  “spots”  where  temporary  performances  degrade  at 
a  maximum  4.6  times  the  1 -sigma  of  residual  error  affecting  code  range 


Washington  Academy  of  Sciences 


59 


measurements;  all  the  rest  of  the  surface  experiments  have  satisfactory 
DOP  values,  including  the  polar  regions. 


Conclusion 


Local  and  global  navigation  on  the  Moon  are  seen  as  an  achievable 
objective  with  current  technologies,  especially  with  Pseudolites  and 
Satellite  Navigation  Constellations,  providing  that  all  possibilities  and 
combinations  of  these  “tools”  are  carefully  examined. 

A  strong  effort  on  the  study  phase  will  maximize  the  attainable 
results,  with  a  proper  trade-off  between  available  alternatives.  Especially 
considering  the  opportunity  of  complementing  the  navigation  space 
vehicles  with  “advanced”  ground  emitters,  it  will  drive  the  design  of  very 
promising  equipment  and  related  localization  systems. 

When  dealing  with  constellations,  a  number  of  features  are  to  be 
accounted  for,  which  include  the  difficulty  of  “bringing”  and  maintaining 
the  satellites  on  target  orbits,  due  to  the  lack  of  stable  Moon  trajectories. 

Issues  range  from  geometric  DOP  figures  to  the  cost  for  reaching 
and  keeping  over  time  a  desired  orbit  profile  by  each  navigation  vehicle. 

Special  care  is  required  in  developing  mission  studies,  due  to  the 
complexity  and  inter-discipline  skills  related  to  the  matter.  However,  the 
preliminary  feasibility  analyses  demonstrate  that  all  of  this  is  not  science 
fiction,  but  a  real  opportunity  for  a  joint  adventure  between  Europe  and 
America. 


REFERENCES 

[1]  E.A.  LeMaster,  S.M.  Rock:  "Self-Calibration  of  Pseudolite  Arrays  Using  Self- 
Differencing  Transceivers”.  Institute  of  Navigation  GPS-99,  Nashville,  TN, 
September  1999. 

[2]  E.A.  LeMaster.  S.M.  Rock:  "A  Local-Area  GPS  Pseudolite-Based  Mars 
Navigation  System”.  IEEE  10th  International  Conference  on  Advanced  Robotics. 
Budapest.  Hungary',  August  2001. 

[3]  E.A.  LeMaster.  S.M.  Rock:  "An  Improved  Solution  Algorithm  for  Self- 
Calibrating  Pseudolite  Arrays”.  Institute  of  Navigation  National  Technical 
Meeting.  San  Diego,  CA,  January'  2002. 

[4]  E.A.  LeMaster.  S.M.  Rock:  “A  Local-Area  GPS  Pseudolite-Based  Navigation 
System  for  Mars  Rovers”.  Journal  of  Autonomous  Robots.  Vol.  14.  No.  2-3.  Mar- 
May  2003,  pp.  209-224. 


Fall  2006 


60 


[5]  J.  Muller.  J.G.  Williams.  S.G.  Tun  shew  P.J.  Shelus:  “Potential  Capabilities  ol 
Lunar  Laser  Ranging  for  Geodesy  and  Relativity".  September  6th.  2005 

[6]  Tom  Murphy  (UCSD)  "‘Next-Generation  Lunar  Laser  Ranging".  Presentation  on 
"APOLLO"  (Apache  Point  Observatory  Lunar  Laser-ranging  Operation) 

[7]  C.  Boucher.  Z.  Altamimi.  P.  Sillard.  M.  Feissel- Vernier:  “The  ITRF  200(T. 
International  Earth  Rotation  and  Reference  Systems  Service  (IERS).  Technical 
Note  No.  31 

[8]  Franco  Bemelli  Zazzera.  Francesco  Topputo.  Mauro  Massari:  “Assessment  of 
Mission  Design  Including  Utilization  of  Lib  ration  Points  and  Weak  Stability 
Boundaries".  Study  developed  under  ESA  Contract  N°.18147/04/NL/MV 


Washington  Academy  of  Sciences 


61 


ROBOTIC  AID  TO  MOON  BASE 

P.  Magnani(*),  B.  Midollini(*),  B.  Papalia(°) 
(*)Galileo  Avionica  S.p.A.  (°)ENEA 


This  article  focuses  on  the  role  played  by  Automation  and 
Robotics  (A&R)  as  part  of  the  logistics  system,  in  support  of  human 
beings  in  the  different  phases  of  their  presence  on  the  Moon:  from 
exploration,  to  base  settlement  and  running,  to  resources  localization  and 
exploitation. 


It  was  late  13th  century  when  Marco 
Polo  reached  China  after  a  15,000 
miles  journey  on  sea  and  land  which 
took  him  three  and  a  half  years. 

Two  centuries  later,  Cristoforo 
Colombo  reached  America  after  having 
sailed  for  79  days  and  3,000  miles:  even  if  shorter  than  the  journey  of  his 
predecessor,  this  event  represented  a  turning  point 
for  our  history,  marking  the  end  of  the  medieval 
age  and  the  start  of  the  modern  era. 

These  are  only  two  outstanding  examples,  but 
human  exploration  was  born  much  earlier  than 
then,  right  since  the  very  first  appearance  of 
humanity  on  the  Earth.  Since  then,  humans  have 
extended  their  exploration,  taking  advantage 
meanwhile  of  the  means  that  scientific  and 
technological  research  were  bringing. 

Thanks  to  these  means  we  have  been  able  to  cover  in  few  days  the  over 
350,000  km  distance  from  our  satellite,  to  orbit  around  it  and,  eventually, 
to  step  on  the  Moon’s  surface. 


Fall  2006 


62 


Courtesy:  NASA 


Now  we  are  going  to  face  this  new 
challenge:  set  an  initial  outpost  on  the  Moon 
in  view  of  creating  a  proper,  self-reliant, 
permanently  inhabited  base,  where  research 
and  experiments  will  be  carried  out,  and 
which  will  be  used  as  a  starting  point  for 
future  missions  of  the  solar  system  human 
exploration. 


Courtesy:  NASA 


The  astronaut  is  not  alone 
in  his  adventure:  the 
relatively  simple  rover 
used  to  explore  limited 
areas  around  the  lander  in 
1970  will  be  now  replaced 
by  systems  employing  the 
technologies  developed  in 
the  last  decades,  which 
will  provide  valid 
instruments  able  to  perform  tasks  otherwise  not  feasible  or  which  will  ease 
human’s  work  on  the  Moon. 


Courtesy:  NASA 


Washington  Academy  of  Sciences 


63 


Among  such  systems  are  those  based  on  robotics  and  automation 
techniques.  These  systems  will  be  effectively  used  in  a  number  of 
applications,  e  g.: 


•  Site  exploration 

•  Site  preparation 

•  Modules  recovery,  placement 
and  integration 

•  Resources  exploitation 

•  Monitoring  and  maintenance 


A  first  example 
is  provided  by 
Moon 

exploration: 
candidate  sites 
for  the  base 
installation 
need  to  be 

surveyed  prior  Courtesy:  ESA 

to  final  selection  in  order  to  check  for  surface  and  subsurface 
characteristics.  In  this  case  A&R  technology  can  provide  light  surveyor 
rovers  with  appropriate  degrees  of  autonomy  equipped  with  stereo  camera 
and  a  positioning  system  (such  as  a  lunar  GPS  or  lunar  Galileo,  or  local 
radar/optical  reference).  This  will  allow  one  to  build  a  topographical 
model  of  the  whole  site  to  be  used  as  reference  for  all  robotic  operations. 
The  rovers  will  also  be  equipped  with  drilling  and  sample  manipulation 
capabilities  to  determine  the  physical  properties  of  soil  and  subsoil. 


Courtesy:  Oak  Ridge  National  Laboratory 


Fall  2006 


64 


A&R  technology  will  be  used  also  for  site  preparation:  stones  and  debris 
will  be  removed  by  means  of  robotic  arms  with  clamps  or  shovels  end 
effectors  and  the  terrain  will  be  properly  leveled.  Digging  devices  will  be 
used  to  prepare  the  terrain  for  those  sections  of  the  base  (like  stock  rooms, 
tanks  or  communication  lines)  which  need  to  be  located  under  the  regolith. 


Once  the  infrastructure  site  has  been  localized  and  prepared,  the  various 
units  and  modules  which  have  been 
sent  to  the  Moon  (and  which  are 
physically  distributed  on  a  certain 
area)  must  be  recovered  from  the 
landing  place  and  transported  to  the 
right  location. 


To  this  end,  a  large  vehicle  will  be 
used,  with  specific  capabilities  for 
handling  and  transportation  of  Courtesy:  Texas  University 

payloads  where  particular  care  must  be  taken  (for  example,  sealing 
capability  to  protect  it  from  dust  or  leveling  capability). 


Finally,  the  modules  have  to  be  assembled:  also 
in  this  case  a  considerable  amount  of  work  can 
be  done  by  tele-controlled  robotic  arms 
installed  on  a  vehicle,  or  running  on  a  rail 
system  around  the 
module. 


Courtesy:  nasa  The  robotic  system 
will  correctly  position  the  various  parts  (in 
terms  of  relative  distance  and  relative 
orientation)  to  allow  inflate  procedure, 
deployment  and  installation;  and  it  will  be 
capable  of  assembling  pipes  or  communication  lines  between  the  different 
modules. 


The  use  of  lunar  resources  is  a  key  point  to  reduce  cost  and 
dependability  from  Earth.  Once  identified  by  sampling  and  analysis,  the 
material  (like  rocks  or  ice)  could  be  collected  by  autonomous  mining 
vehicles  and  then  refined  in-situ  (when  timely  processing  is  required 
because  of  the  presence  of  volatile  material)  or  at  dedicated  processing 
plants,  and  stored  for  later  use.  Loose  material  as  well  as  material  cut  in 


Washington  Academy  of  Sciences 


_ 


65 


particular  shapes  can  be  collected  and  processed,  the  shape  depending 
once  again  on  the  type  of  material  handled. 

But  the  cooperation  between  astronauts  and 
robots  can  also  be  usefully  extended  to  the 
internal  environment.  This  is  a  structured 
environment  which,  however,  may  vary 
either  in  a  correct  way  (i.e.  for  the 
intentional  intervention  of  the  astronauts)  or 
because  of  a  problem  or  malfunction. 

Courtesy:  NASA  ,  , 

Robotic  systems,  basing  on  an  a  priori 
knowledge  of  the  plant  layout,  could  automatically  detect  and  evaluate 
sudden  variations  and  give  a  warning  to  the  crew. 

A&R  systems  can  effectively  perform  tasks  of  surveillance  and  detailed 
inspection,  together  with  transportation  of  parts,  ordinary  maintenance, 
execution  of  tasks  typical  for  the  facility  (for  example,  run  of  experiment 
facility  or  of  production  areas)  and  contingency  operations. 


To  summarize,  the  A&R  equipment  needed  for  the  Moon  Base  can  be 
grouped  as  follows: 


Courtesy:  University  of  Bologna 


Fall  2006 


66 


•  wheeled  or  walking  medium  size  vehicles,  both  for  outer  and  inner 
operations,  allowing  astronaut  transportation  and  dexterous 
interaction  with  the  environment  by  means  of  manipulator  arms 
equipped  with  specialized  end  effectors; 

•  drilling  systems  reaching  a  depth  of  meters  to  tens  of  meters  with 
sampling  capability  and  the  possibility  to  collect  and  temporarily 
store  large  volumes  of  material; 

•  heavy  robots  with  high  thrust  and  stability  against  terrain  and 
special  simple  tools  and  arms  for  interaction  with  the  soil; 

•  large  roving  vehicles  and  cranes  capable  of  handling  big  payloads 
with  simple  interfaces  and  allowing  stable  transportation  also  on  an 
unprepared  terrain; 

•  small  robots  for  inspection,  repair  and  maintenance  with  crawling 
capability  and  smart  articulation 

For  all  these  equipment,  the  main  control  mode  should  be  tele-operation 
with  astronaut  supervision,  while  simple  or  repetitive  tasks  can  be 
performed  autonomously.  In  both  cases,  the  majority  of  tasks  will  be 
planned  and  tested  in  advance. 


The  basic  building  blocks  are  common  to  all  these  robotic  equipment,  and 
this  permits  one  to  undertake  a  modular  approach  for  the  design. 

Also,  Italy  can  rely  on  the  experience  and  capability  gained  not 
only  in  robotics,  but  in  general  in  integrated  space  systems  and  missions 
definition. 

This  allows  us  to  reduce  the  design  efforts  and  to  effectively 
contribute  to  the  implementation  of  the  robotic  support  to  Moon  Base. 


Washington  Academy  of  Sciences 


Fall  2006 


This  page  intentionally  left  blank 


Washington  Academy  of  Sciences 


69 


JANE  AND  JOHN  BORN  IN  LUNAR  JAMESTOWN,  2020 

AN  OUTLINE  OF  A  RESEARCH  PROJECT  ON  SEXUALITY, 
FERTILITY,  PREGNANCY,  AND  BIRTH  ON  THE  MOON 

Roberto  Varrasi,  MD 
Alberto  Revelli,  MD 

Department  of  Obstetrical  and  Gynaecological  Sciences 
University  of  Toruno.  Italy 


Abstract 

In  order  to  have  a  safe  birth  on  the  Moon  or  in  other  low  gravity  or  no 
gravity  environments,  we  need  to  know  much  more  about  how  gravity, 
or  the  absence  of  it  affects  the  human  body  and  its  functioning  related  o 
fertility,  pregnancy,  and  birth.  This  paper  briefly  outlines  the  research 
needed  to  supply  this  know  ledge. 


Introduction 

THE  birth  of  a  baby  represents  the  continuity  of  humanity,  and 
therefore  is  always  a  blessing;  this  is  even  more  valid  if  the  baby  first  sees 
the  light  in  Jamestown,  the  first  human  colony  on  the  Moon.  The  first 
baby  born  in  Jamestown  will  certainly  be  mentioned  in  the  history  books 
as  the  first  real  Universe  explorer.  This  baby  will  be  more  important  than 
Cristoforo  Colombo  or  Neil  Armstrong:  they  were  bom  on  the  Earth, 
while  she  or  he  will  be  the  first  human  being  born  outside  the  “Mother 
Earth,”  and  therefore  will  automatically  become  the  living  symbol  of  the 
humanity  that  wants  to  explore  and  pacifically  colonize  other  Worlds. 

In  order  to  have  a  safe  birth,  at  least  from  the  point  of  view  of 
health,  the  road  is  very  difficult:  a  wide  research  program  is  necessary  to 
guarantee  to  the  baby  and  to  the  mother  the  best  conditions  for  this  very 
important  test.  Studies  of  the  human  body’s  behavior  outside  the 
protection  of  Mother  Earth  are  still  at  the  first  steps;  to  date,  it  is  not  well 
known  what  all  the  consequences  may  be  of  the  prolonged  absence  of 
gravity  or  low  gravity  environment  and  artificial  atmosphere. 


Fall  2006 


70 


This  is,  in  particular,  applicable  to  problems  related  to  sexuality 
and  reproduction. 

Scientific  Program 

Background  and  Rationale 

The  physiology  of  human  reproduction  is  of  major  interest  when 
hypothesizing  a  human  colony  living  in  the  Moonbase  in  Jamestown  in  the 
near  future.  At  present  knowledge  about  human  reproduction  outside  the 
Earth’s  atmosphere  is,  to  the  best  of  our  knowledge,  totally  lacking,  and 
thorough  experimental  work  through  a  broad  research  program  is  needed 
to  estimate  the  reproductive  potential  of  human  beings  on  the  Moon. 

The  scientific  program  will  include  not  only  tests  and  instrumental 
analyses,  but  these  activities  will  be  constantly  supported  by  scientific 
interviews  with  the  persons  involved  in  the  experimental  program. 

The  following  scientific  and  experimental  steps  will  give  us  the 
necessary  information  to  obtain  safe  born  of  a  baby  on  the  Moon. 

Step  1:  Sexuality 

The  first  step  is  aimed  at  identifying  the  behaviours  associated  with 
sexuality.  Two  areas  of  research  are  identified. 

The  Physiology  of  Erection  and  Ejaculation :  Healthy  volunteers 
living  in  the  absence  of  gravity  or  in  a  low  gravity  environment  are  given 
stimulation  and  results  are  measured  with  proper  instrumentation.  Semen 
win  be  collected  and  transferred  to  analysis.  No  severe  technological  issue 
is  identified. 

Intercourse:  The  dynamics  of  sexual  intercourse  in  absence  of  gravity 
or  in  a  low  gravity  environment  are  studied. 

Step  2:  Fertility 

The  second  step  is  to  identify  the  behaviors  associated  with  the 
physiology  of  female  and  male  sexual  apparatus.  With  respect  to  issues 
that  have  high  ethical  contents,  animal  models  will  be  utilized  as  a  first 
step.  The  following  areas  of  research  are  identified. 

Menstruation  and  Owlation:  The  menstrual  rhythm  of  healthy  female 
volunteers  living  in  the  absence  of  gravity  or  in  a  low  gravity  environment 
is  recorded.  During  the  menstrual  cycle  they  are  submitted  to  repeated 
blood  sampling  aimed  to  estimate  the  occurrence  of  ovulation  and  to 


Washington  Academy  of  Sciences 


71 


check  for  the  eventual  increase  of  stress  hormones  potentially  interfering 
with  ovulation. 

Semen  Examination:  Semen  samples  are  objectively  studied  in  situ 
by  a  computer-assisted  analyser.  A  necessary  technological  issue  is  to 
develop  an  apparatus  to  perform  this  test  in  the  absence  of  gravity 

Semen  Ccipcicitation :  Collected  semen  samples  are  processed 
through  an  automatic  semen  capacitating  apparatus;  the  final  preparation 
is  analysed  by  a  computer-assisted  semen  analyser.  The  above  defined 
technological  issue  applies. 

In  Vitro  Fertilization  ( IVF ):  This  issue,  which  implies  ethical 
problems,  is  one  of  the  subjects  to  be  treated  also  by  means  of  animal 
models.  Animal  models  will  be  based  on  associated  studies  of  mice  and 
lab-raised  monkeys. 

The  purpose  of  this  experiment  is  to  obtain  fertilization  of  ova  in 
orbit  or  in  a  low-g  environment. 

Several  analytical  approaches  could  be  used. 

A  first  hypothesis  is  to  analyse  the  fertilization  potential  of  fresh 
oocytes  obtained  on  the  ground  and  incubated  in  orbit  with  fresh  or 
thawed  semen. 

A  second  possibility  could  be  the  utilization  of  frozen  oocytes 
from  fertile  subjects  that  are  thawed  in  the  absence  of  gravity  and 
fertilized  in  vitro  with  spermatozoa  taken  from  frozen-thawed  or  freshly 
produced  semen  samples  of  fertile  healthy  subjects.  In  this  second  case  the 
intracytoplasmic  sperm  injection  (ICSI)  is  applied  to  get  fertilization.  The 
injected  oocytes  are  kept  in  IVF  incubators  and  fertilization  is  assessed  by 
microscopy  some  hours  later.  Mandatory  technological  improvement 
associated  with  this  hypothesis  is  to  develop  an  apparatus  able  to 
automatically  perform  intracytoplasmic  sperm  injection  (ICSI). 

Embryo  Development:  This  issue  is  a  direct  follow-up  of  the 
previous  problem;  therefore,  in  this  case  also  animal  models  will  apply,  in 
accordance  with  the  above  defined  approach. 

The  purpose  of  this  experiment  is  to  observe  the  development,  in 
the  absence  of  gravity  or  in  a  low  gravity  environment,  of  fertilised  ova  in 
the  first  six  days  of  life. 

Fertilized  oocytes  are  kept  in  IVF  incubators  up  to  day  6  of 
development.  Their  growth  is  daily  observed  by  an  invertoscope  until  they 


Fall  2006 


72 


reach  the  blastocyst  stage  (day  6).  After  the  return  to  ground  and  the 
recovery  of  these  embryos,  ultrastructural  and  chromosomal  analysis  will 
be  performed. 

Proper  flight  standards  of  the  normally  utilized  laboratory 
instrumentation  shall  be  developed. 

Step  3 :  Pregnancy 

This  issue,  even  if  does  not  imply  ethical  problems,  is  also  one  of 
the  subjects  to  be  treated  by  means  of  animal  models. 

This  third  step  is  designed  to  analyze  the  behaviors  associated  with 
the  pregnancy.  Two  areas  of  research  are  identified. 

It  has  to  be  observed  that,  considering  the  duration  of  the 
pregnancy  either  of  humans  or  of  the  candidate  animal  models,  several 
possibilities  could  be  considered;  i.e.,  having  the  test  subjects  in  the 
absence  of  gravity  or  in  a  low  gravity  environment  only  for  a  part  of  the 
pregnancy,  eg.  the  first  or  the  central  part  of  the  pregnancy. 

Also  in  this  case,  proper  flight  standards  for  the  normally  utilized 
laboratory  instrumentation  shall  be  developed. 

Utero-placental  Blood  Fluxes:  The  blood  fluxes  in  the  uterine  and 
umbilical  circulation  are  studied  by  ultrasound  Doppler  fluximetry  in  the 
absence  of  gravity  at  different  stages  of  a  normal  pregnancy  (I,  II  and  III 
trimester).  The  foetal  growth  is  assessed  throughout  the  pregnancy  by 
repeated  ultrasound-based  biometry. 

Foetal  Heart  Beal:  The  foetal  heart  beat  of  healthy  foetuses  in  the 
third  trimester  of  pregnancy  are  registered  by  cardiotocography  in  the 
absence  of  gravity. 

Step  4:  Delivery 

The  last  step  is  obviously  the  goal  of  the  entire  scientific  program 
and  is  associated  with  the  delivery  of  the  baby,  and  with  the  problems 
associated  with  neonatal  care.  Animal  models  (i.e.  delivery  of  babies  of 
animals)  will  obviously  anticipate  delivery  of  the  human  baby. 

Spontaneous  Vasinal  Delivery:  A  spontaneous  vaginal  delivery  is 
observed  in  the  absence  of  gravity  or  in  the  presence  of  reduced  gravity. 

Neonatal  Care:  A  newborn  delivered  in  the  absence  of  gravity  or 
in  the  presence  of  reduced  gravity  is  studied  with  particular  attention  to 
the  respiratory  and  cardiovascular  functions. 


Washington  Academy  of  Sciences 


73 


Programmatic  Considerations 

All  the  above  activities  require  several  years  to  be  performed;  this 
applies  in  particular  to  those  experimental  activities  that  need 
development  of  automatic  equipment  to  be  utilized  on  orbit. 

During  the  program  it  is  clear  that  development  of  animal  models 
must  be  performed  before  the  development  of  corresponding  experiments 
on  humans,  in  order  to  benefit  from  the  experience  with  animals. 


Fall  2006 


This  page  intentionally  left  blank 


Washington  Academy  of  Sciences 


* 


A  MOON  BASE 

KNOWLEDGE  AND  IMAGINATION  PORTAL 


75 


Gabriele  Peraldo  Bertinet, 
Rodolfo  Guzzi, 
Bruno  Ratti, 

Anna  Rebecchi 
Geoknowledge  Foundation’ 


“Che  sulla  lima...  non  intendo  gia  come  tuttalvolta  che  non  vi  si 
generino  cose  simili  alle  nostre  " 

(Galileo  “Dialogo  dei  Massiini  Sistemi") 

Abstract 

Knowledge  represents  the  ultimate  rational  of  the  Moon 
Base  Enterprise  and  more  generally  of  space  exploration.  In 
the  present  Knowledge  Society,  knowledge  is  central  for  its 
capability  to  create  value  through  cognitive  multiplication 
based  on  sharing  and  reuse.  In  this  paper  the  creation  of  “A 
Moon  Base  Knowledge  Portal”  infrastructure  to  this  aim  is 
proposed.  The  Portal  will  be  structured  as  a 
multidisciplinary  Portal  of  Portals  based  on  formal 
ontology.  This  Portal  can  permit  achievement  of  important 
objectives  essential  for  the  success  of  the  Moon  Base 
Program,  in  particular:  to  support  an  Inspiration  Program  to 
involve  new  and  future  generations  of  students;  to  obtain 
consensus  of  public  opinion  and  support  of  all  stakeholders 
(government,  science,  universities,  taxpayers)  who  must 
share  aims  and  objectives  of  the  Program. 


Fall  2006 


Introduction 


The  Earth-Moon  System  probably  was  born  as  a  result  of  a 
collision  between  the  young  Earth  and  other  smaller  planetary  bodies  that 
were  also  growing.  One  of  these  hit  Earth  late  in  Earth's  growth  process, 
blowing  out  rocky  debris.  A  fraction  of  that  debris  went  into  orbit  around 
the  Earth  and  aggregated  into  the  Moon  about  4.5  billion  years  ago  (W.  K. 
Hartmann  and  D.  R.  Davis  1975).  The  giant  impact  hypothesis  had  the 
advantage  of  invoking  a  stochastic  catastrophic  event  that  might  happen 
only  to  one  or  two  planets  out  of  nine.  Furthermore,  as  also  was  evidenced 
by  the  rocks  collected  during  the  Apollo  Moon  landings,  the  Moon  has 
exactly  the  same  oxygen  isotope  composition  as  the  Earth,  showing  that 
the  Moon  formed  from  material  found  in  Earth's  mantle.  Mars  rocks  and 
meteorites  from  other  parts  of  the  Solar  System  have  different  oxygen 
isotope  compositions.  This  giant  impact  may  have  also  produced  the 
Earth’s  axial  tilt  and  initial  rotation.  For  these  reasons  the  Moon  has  been 
essential  for  the  formation  of  the  terrestrial  environment  and,  being  part  of 
the  history  of  the  Earth,  has  co-evolved  with  the  birth  and  evolution  of 
mankind. 

Its  presence  in  the  sky  was  never  seen  as  an  anomalous  presence, 
but  humans  always  aspired  to  the  Moon  because  they  understood  its 
influence  on  several  natural  events  that  were  linked  to  life.  Its  presence 
became  part  of  our  biological  and  cultural  memory.  In  the  ancient  cult  the 
Moon  was  a  Goddess,  and  poets,  painters,  and  musicians  have  dedicated 
their  art  to  our  Moon  before  scientists  discovered  that  the  Moon  influences 
several  phenomena  on  the  Earth  from  sea  tides  to  tree  growth  up  to  the 
recently  guessed  influence  on  human  life  and  death. 

For  these  reasons  our  Moon  is  in  the  immaginario  collettivo 
(collective  imagination)  and  its  appeal  crosses  the  whole  range  of  human 
imagination  and  emotions.  Thus  the  Moon  is  not  only  the  subject  of 
knowledge  but  it  is  part  of  our  imagination  and,  today,  the  Earth-Moon 
system  belongs  more  to  the  imagination  than  to  the  science. 

The  aim  of  this  paper  is  to  treat  the  Moon  not  only  as  subject  of 
knowledge,  but  also  part  of  our  imagination,  fantasy,  and  dreams.  The  best 
representation  in  which  science  and  fantasy  are  combined  is  a  pertinent 
portal  whose  structure  and  function  will  be  outlined  in  the  next 
paragraphs. 


Washington  Academy  of  Sciences 


77 


Ontology  Quest  and  Web  Portal  Structure 

The  basis  of  knowledge  is  the  ontology  quest.  This  marks  the 
cultural  aspect  of  people  and  their  heritage.  The  basic  assumptions  for  one 
culture  may  be  not  the  same  for  another.  People’s  queries  differ  from  one 
culture  to  another,  even  though  the  imagination  and  emotions  may  play  a 
certain  role,  and  the  ontology  quest  follows  these  aspects.  Educational 
tools  (Guzzi  et  al  2005)  may  be  developed  by  formal  ontology 
(Cocchiarella  1991)  and  that  are  able  to  reach  different  users: 

■  To  support  an  Inspiration  Program  with  extensive  use  of 
advanced  e-learning  techniques,  to  involve  new  and  future 
generation  of  students,  at  different  degrees,  in  science, 
engineering,  technology  disciplines  relevant  to  Space 
Exploration; 

■  To  obtain  consensus  of  public  opinion  by  media; 

■  To  support  all  stakeholders  (government,  science,  universities, 
taxpayers)  which  must  share  aims  and  objectives  of  the 
Program,  through  an  effective  communication  on  space 
exploration’s  expected  results. 

The  best  solution  to  these  questions,  which  also  causes  web  surfers 
to  dream,  is  to  create  a  web  portal  whose  aim  is  to  produce  knowledge  and 
induce  fantasy.  Such  portals  could  be  the  “Moon  Base  Knowledge  & 
Imagination  Portal.” 

The  access  through  the  Portal  to  video  and  simulation  games  can 
engage,  in  particular,  young  people  very  familiar  with  virtual  reality, 
multimedia,  and  web  techniques,  in  learning  principles  of  celestial 
mechanics,  space  flight,  and  exploration  activities,  achieving  extraordinary 
educational  results  and  benefits.  In  addition  mysterious  rooms  will  be 
created  to  support  the  imagination  and  induce  the  fantasy  of  the  users. 

The  Moon  Base  Knowledge  and  Imagination  Portal  will  be 
implemented  as  a  “Portal  of  Portals”  developing  the  encounter  between 
nations  participating  in  the  Moon  Enterprise.  It  will  be  the  result  of  an 
International  Cooperative  Program  involving: 

■  International  and  national  space  agencies; 

■  Scientific  institutes  and  research  centers; 

■  Universities; 

■  Private  companies  operating  in  space  and  related  sectors; 
and 


Fall  2006 


■  Cultural  associations  and  foundations. 

As  a  result  of  this  cooperative  effort  a  Partnership  will  be 
established  between  participant  entities  inclusive  of  a  management 
structure  responsible  for  direction,  standards  definition  admission  policies, 
intellectual  property,  etc. 


Building  up  the  Web  Portal 

The  Web  Portal  will  be  structured  in  four  major  areas:  scientific 
knowledge,  system  and  engineering  activities,  enabling  technologies  and 
applications,  and  institutional  legal  and  economic  framework. 

Scientific  Knowledge 

This  is  relevant  mainly  to  the  results  of  colonization  and  of  the 
Condominium  of  Observatories  on  the  Moon.  Among  the  topics  to  be 
included: 

■  Cosmology; 

■  Planetology:  with  particular  emphasis  to  study  of  the  Moon-  Earth 

System; 

■  Climatology:  climate  change,  interactions  between  solar  activity 

and  climate  on  the  Earth 

■  Life  science;  and 

■  Energy  generation  and  storage. 

Information  will  be  organized  in  different  levels  of  complexity  for 
different  targets  of  people  accessing  this  Portal,  having  in  mind  the 
objective  of  maximum  dispersion  and  sharing  of  knowledge. 

System  and  Engineering  Activities 

Implementations  and  activities  to  build  up  and  maintain  a  Space 
Infrastructure  on  the  Moon  will  be  outlined  in  this  area,  in  order  to  keep 
people  informed  and  involved  in  the  enterprise.  Among  the  topics 
addressed: 


■  Lunar  Missions  profiles; 

■  Space  transportation  systems; 

■  Living  modules; 

■  Life  support  systems; 

■  Environmental  control  Systems; 

■  Observation  Payloads  and  Infrastructures; 

■  Communication  and  Location  Infrastructure;  and 


Washington  Academy  of  Sciences 


79 


■  Lunar  Resources  Exploitation. 


Enabling  Technologies  and  Applications 

Development  of  the  enabling  technologies  and  applications 
supporting  the  Moon  Program  will  be  indicated  and  summarized  while 
remaining  compatible  with  intellectual  property  rights  and  “sensitivity”  of 
information.  Emphasis  will  be  given  to  technologies  which  may  have  a 
direct  fallout  on  our  day  by  day  lives  as: 

■  Telemedicine; 

■  Advanced  Materials  and  structures 

■  Robotics; 

■  Advanced  nuclear  power  generation; 

■  Wide  Band  Communications; 

■  Nanotechnologies; 

■  Observation  sensors; 

■  Waste  material  recycling;  and 

■  Environmental  Control. 


Institutional  Legal  and  Economical  Framework 

The  establishment  of  an  International  Cooperation  between 
governments  and  Partnerships  with  private  industry  financial  involvement 
(PPP)  will  be  a  way  to  implement  the  Project.  Commercial  exploitation  of 
resources  and  new  products  deriving  from  the  colonization  of  the  Moon 
may  require  the  reconsideration  of  the  1967  Space  Treaty.  Issues  of 
property  rights,  freedom  of  navigation,  and  technology  transfer  will 
become  relevant.  All  these  aspects  and  their  evolution  will  be  dealt  in  this 
area  of  the  Portal. 

Furthermore,  within  the  Web  Portal  mysterious  rooms  will  be 
present  along  the  whole  portal  stimulating  the  fantasy  and  the  imagination, 
in  the  same  way  as  happened  for  the  early  theories  of  the  origin  of  Earth 
and  Moon,  because  fantasy  and  imagination  are  also  crucial  for  future 
discovers  of  mankind. 

The  Geoknowledge  Paradigm 

Within  the  Moon  Base  Knowledge  &  Imagination  Portal  a  GIS 
(Geographic  Information  Systems)  Portal  to  represent  the  “geographic 


Fall  2006 


80 


knowledge”  of  the  Moon  will  be  also  implemented  to  discover  the 
geographic  dimension  of  our  life  in  the  new  continent  of  the  Moon. 

The  Moon  is  considered  as  the  Earth’s  seventh  continent,  the 
colonization  of  which  would  lead  to  increased  acumen,  so  maintaining  the 
geo-knowledge  terminology  for  the  Moon  goes  with  retaining  ontology, 
semantics,  and  methodologies  used  for  the  knowledge  of  the  Earth.  The 
geo-knowledge  of  the  Moon  will  be  referred  to  Earth  sciences  (such  as 
geophysics,  geology,  topography)  and  processes  on  the  Moon  (such  as 
mobility,  production  of  energy,  buildup  of  infrastructures,  exploitation  of 
resources).  For  these  reasons  a  3D  Moon  GIS  has  to  be  implemented  to 
support  missions  on  the  Moon. 


Example  Web  Portal 

Every  information  system  has  its  own  ontology  ascribing  meaning 
to  the  symbols  used  according  to  a  particular  vision  of  the  world  (Guarino 
1995,  1998).  The  Information  System  consists  of  three  different 
components:  application  programs,  information  resources  like  databases 
and/or  knowledge  bases  and  user  interfaces.  The  ontology  impacts  on 
information  systems  by  two  orthogonal  dimensions:  temporal  and 
structural.  In  the  first  case  the  semantics  expressed  by  ontology  is 
transformed  and  translated  into  an  information  system  component.  In  fact, 
users  reuse  the  knowledge  instead  of  the  software  by  using  a  common 
vocabulary  across  heterogeneous  software  platforms  (even  though  current 
ontology  is  limited  yet).  In  the  second  case,  even  though  the  quantity  of 
ontological  knowledge  available  may  be  poor,  the  quality  can  improve  the 
analysis  process. 


Graphic  Interface 

In  a  portal  the  major  role  is  played  by  the  graphics  interface.  We 
don’t  enter  into  the  methodology  of  the  web  portal  building,  but  describe 
here  the  ontology  of  the  portal  and  the  operations  to  be  carried  out.  Users 
act  on  classes  and  descriptors  of  the  system  getting  directories  containing 
images,  data,  movies  and  sound  files  related  to  the  topics  of  the  domain 
knowledge  selected.  During  this  phase,  the  system  acts  selecting  the  best 
items  or  statements  describing  the  domain  knowledge.  The  task  is  carried 
out  by  an  Evaluator,  in  which  the  domain  knowledge  is  explicitly 
transferred  by  means  of  simple  rules  defined  in  system.  This  unit 


Washington  Academy  of  Sciences 


81 


maintains  and  uses  the  Ontology  Integration  System  to  Information 
System.  Figure  1  shows  the  layout  of  the  system.  Figures  2  and  3  show 
both  the  iconographic  aspect  of  the  Portal  and  its  appeal. 


Figure  1:  Structure  of  the  system  with  the  evaluator  between  the 
formalized  knowledge  and  users’  knowledge 


Fall  2006 


82 


Washington  Academy  of  Sciences 


83 


Figure  3:  MOON  BASE  Knowledge  &  imagination  portal.  Contents 
artistic  view 


Movies  and  Role  Games 

Using  formal  ontology  an  educational  tool  about  Moon  missions 
has  been  developed.  It  is  shown  in  figure  4.  The  techniques  have  been 
described  by  Guzzi  et  al  (2005)  for  planetary  missions.  We  referred  to 
some  detailed  scientific  documents  on  space  mission  design  (Wertz, 
Larson  (1999)  and  Doody  (2001))  to  identify  the  main  parameters 
affecting  a  generic  space  mission.  Real  Moon  missions  documented  by 
NASA  (JPL-NASA,  NASA  missions,  JPL  reports,  NASA  database)  have 
been  used  to  tune  mission  parameters  and  main  items  with  real  mission 
quantities.  We  have  fragmented  a  hypothetical  mission  into  several  pieces 
and  within  these  we  have  identified  several  possible  different  cases.  A 
proper  interface  to  QuickTime  has  been  built  up  for  AVI  and  MPEG 
movies,  which  are  in  the  right  part  of  the  picture.  In  the  upper  left  the 
chosen  instrument  for  the  scientific  mission,  in  the  middle  the  instruments 
to  perform  the  space  mission  (from  the  navigation  sensors  to  power,  etc), 
and  in  the  lower  part  the  suitable  amount  of  fuel. 


Fall  2006 


84 


Figure  4:  Graphic  user  interface  of  Space  Mission  as  described  in  the  text 


The  Launch  Button  allows  the  mission  to  run  and  to  be  visualized. 
Vice  versa  when  a  set  up  error  occurs,  a  warning  related  to  the  mission 
failure  is  shown.  The  warning  also  contains  a  link  to  the  space  mission 
online  manual  where  there  is  information  allowing  the  usei  to  understand 
his  error  and  correct  it. 


Conclusions 

In  this  paper  we  have  outlined  a  Portal  of  Portals  for  Moon 
missions  and  discoveries.  Despite  the  technical  aspect  of  the  mission,  we 
have  remembered  the  role  played  by  the  Moon  on  imagination,  our 
memory  and  our  fantasy.  For  this  reason  we  have  introduced  in  our  Portal 
the  concept  of  imagination  and  we  have  drawn  the  pertinent  iconographies 
in  order  to  give  more  appeal  to  the  information  system.  At  the  same  time 
we  have  introduced  the  formal  ontology  and  its  first  application  already 


Washington  Academy  of  Sciences 


85 


done  and  addressed  to  the  Moon  Mission.  Furthermore  since,  nowadays, 
we  have  matured  our  primordial  mental  geography  in  a  way  to  build  the 
first  cartography  of  the  Moon  as  we  made  with  the  first  continents  some 
centuries  ago;  considering  the  Moon  as  the  seventh  continent,  we  also 
introduced  the  first  Moon  GIS  to  give  the  first  geographic  route  for  future 
travelers. 


References: 

Cocchiarella.  N.  B.  1991.  “Formal  Ontology’'  In  H.  Burkhardt  and  B.  Smith  (eds.). 

Handbook  of  Metaphysics  and  Ontology.  Philosophia  Verlag.  Munich:  640-  647 
Doodv  (2001)  “Basics  of  Space  Flight’*  JPL  D-20120.  Dave  Doody.  February  2001.  from 

Imp :  //www .  j  pi .  nasa .  gov/basics/ 

Guarino  N..  “Formal  Ontology.  Conceptual  Analysis  and  Knowledge  Representation” 
International  Journal  of  Human  and  Computer  Studies ,  special  issue  on  The 
Role  of  Formal  Ontology  in  the  Information  Technology  edited  by  N.  Guarino 
and  R.  Poli.  vol  43  no.  5/6.  1995 

Guarino  N..  “Formal  Ontology  and  Information  Systems”.  In  N.  Guarino  (ed.).  Formal 
Ontology  in  Information  Systems  Proc.  of  the  1st  International  Conference. 
Trento.  Italy,  6-8  June  1998.  IOS  Press  (amended  version)  you  may  refer  also  to 
the  online  database 

http:  '  'm’ww. ladseb. yd. cm. n  intor  ontoloz\  'Papers  ( )nioioz\ ’Papers. htnrl #Onlol ogv 

Guzzi  R..  S.  Scarpanti,  G.Ballista.  &  W.  DiNicolantonio  2005  Educational  Technology  & 
Society  8.  (1)  80-90 

Hartmann.  W.  K.  and  D.  R.  Davis  1975  Icarus .  24.  505 
JPL  NASA  home  page  at  http://ww w. jpl.nasa. gov/ 

JPL  NASA  mission  reports  at  http://www.jpl.nasa.gov/status/ 

NASA  planetary  database  at  http : //photo i oumal . jpl . nasa.  gov/ 

JPL  NASA  current  space  missions  at  http://www.jpl.nasa.gov/missions/ 

James  Wertz.  Wiley  Larson.  “Space  Mission  Analysis  and  Design.”  Space  Technology 
Library,  Larson  and  Microcosm,  inc.  Third  Edition.  1999. 


* 

The  GeoKnowledge  Foundation:  Considering  that  space  exploration  has  introduced  a  system  approach  in 
remote-sensing  and  modelling  the  planets  in  our  Solar  System,  thus  improving  the  knowledge  of  the 
planetEarth,  and  that  developments  in  information  and  communication  technology  have  facilitated  a  new  model 
of  knowledge,  based  on  sharing  and  reuse,  the  Foundation’s  aim  is  to  systematically  develop  geo-knowledge 
according  to  the  paradigms  of  the  Knowledge  Society,  activating  a  collaboration  network  of  the  principal  agents 
of  the  geographic  information  and  introducing  a  cognitive  multiplication  factor  into  the  sustainable  usage  of  the 
territory.  The  Foundation  will: 

■  create  a  cultural  meeting  point  for  institutions  and  individuals  who  are  interested  in  geographic 
knowledge,  which  will  generate  scientific,  social,  technological  and  economic  guidelines; 


Fall  2006 


86 


foster  the  creation  of  epistemological  geographic  communities,  where  shared  knowledge  will  be 
based  on  a  common  ontology  of  the  territorial  context  which  unifies  visions,  viewpoints  and 
meanings: 

stimulate  an  interdisciplinary  discussion  on  models  of  geographic  knowledge,  at  national  and 
international  level: 

promote  the  awareness  that  GIS  (Geographic  Information  Systems)  are  essential  tools  for  the 
intelligence  of  natural  and  anthropic  processes  which  exist  in  our  territory'  and  are  a  common 
language  for  increasing  geographic  knowledge. 


Geoknowledge  Foundation  •  Via  Tiburtina.  755  •  Tel.  06  40696201  •  Fax  06  40696333  • 
info(5>geoknowledgefoundation.org 


Washington  Academy  of  Sciences 


87 


NEWS  OF  MEMBERS  AND  AFFILIATED  SOCIETIES 


TO  JERRY  GIBBON,  former  president  of  the  Academy,  our  sincere 
condolences  on  the  death  of  his  son,  in  October. 

SETHANNE  HOWARD,  Associate  Editor  of  this  Journal,  has  announced  the 
forthcoming  publication  of  her  book.  The  Hidden  Giants ,  discussing  the 
4000  year  history  of  women  in  science. 

SAJ  DURRANI  has  been  elected  President  of  the  D  C.  Council  of 
Engineering  and  Architectural  Societies.  The  Council,  formed  in  1936,  has 
more  than  35  affiliated  societies. 

The  American  Statistical  Society’s  new  president  is  Jill 
Montaquila,  of  Westat,  Inc.  Michael  Cohen  is  the  President  Elect  and 
continues  as  the  Society’s  representative  to  the  Academy’s  Board. 

THE  IEEE  COMMUNICATIONS  society  will  hold  a  Global  Communications 
Conference  in  Washington  on  November  26-30,  2007.  This  event 
regularly  attracts  several  thousand  attendees.  Jerry  Gibbon  will  serve  as 
General  Chairman  of  GlobeCom2007.  For  more  information  see 

www.ieee-gl  obecom.  ora/2007. 

Please  send  news  of  Members  and  Affiliated  Societies ,  including  events  to 
occur  after  December  2006  to  the  Editor ,  vcoates@mac.com. 


Fall  2006 


DELEGATES  TO  THE  WASHINGTON  ACADEMY  OF  SCIENCES 
REPRESENTING  AFFILIATED  SCIENTIFIC  SOCIETIES 


Acoustical  Society  of  America 

Paul  Arveson 

American/Intemational  Association  of  Dental  Research 

J.  Terrell  Hoffeld 

American  Association  of  Physics  Teachers 

Frank  R.  Haig,  S.J. 

American  Ceramics  Society 

VACANT 

American  Fisheries  Society 

Ramona  Schreiber 

American  Institute  of  Aeronautics  and  Astronautics 

David  W.  Brandt 

American  Institute  of  Mining,  Metallurgy  &  Exploration 

Michael  Greeley 

American  Meteorological  Society 

Kenneth  Carey 

American  Nuclear  Society 

Steven  Arndt 

American  Phytopathological  Society 

Kenneth  L.  Deahl 

American  Society  for  Cybernetics 

Stuart  Umpleby 

American  Society  for  Microbiology 

VACANT 

American  Society  of  Civil  Engineers 

Kimberly  Hughes 

American  Society  of  Mechanical  Engineers 

Daniel  J.  Vavrick 

American  Society  of  Plant  Physiology 

Mark  Holland 

Anthropological  Society  of  Washington 

Marilyn  London 

ASM  International 

Toni  Marechaux 

Association  for  Women  in  Science  (AWIS) 

Emanuela  Appetiti 

Association  for  Computing  Machinery 

Lee  Ohringer 

Association  for  Science,  Technology,  and  Innovation 

F.  Douglas  Witherspoon 

Association  of  Information  Technology  Professionals 

Barbara  Saffanek 

Biological  Society  of  Washington 

VACANT 

Botanical  Society  of  Washington 

Alain  Touwaide 

Chemical  Society  of  Washington 

James  J.  Zwolenik 

District  of  Columbia  Institute  of  Chemists 

James  J.  Zwolenik 

District  of  Columbia  Psychology  Association 

David  Williams 

Eastern  Sociological  Society 

Ronald  W.  Mandersheid 

Electrochemical  Society 

Robert  L.  Ruedisueli 

Entomological  Society  of  Washington 

F.  Christian  Thompson 

Geological  Society  of  Washington 

Bob  Schneider 

Historical  Society  of  Washington,  DC 

VACANT 

History  of  Medicine  Society 

Alain  Touwaide 

Human  Factors  and  Ergonomics  Society 

Douglas  Griffith 

Institute  of  Electrical  and  Electronics  Engineers 

Gerard  Christman 

Institute  of  Electrical  and  Electronics  Engineers 

Murty  Polavarapu 

Institute  of  Food  Technologies 

Isabel  Walls 

Institute  of  Industrial  Engineers 

Russell  Wooten 

Instrument  Society  of  America 

Hank  Hegner 

Marine  Technology  Society 

Judith  T.  Krauthamer 

Mathematical  Association  of  America 

Sharon  K.  Hauge 

Medical  Society  of  the  District  of  Columbia 

Duane  Taylor 

National  Capital  Astronomers 

Jay  H.  Miller 

National  Geographic  Society 

VACANT 

Optical  Society  of  America 

Jim  Cole 

Pest  Science  Society  of  America 

VACANT 

Philosophical  Society  of  Washington 

Vary  T.  Coates 

Society  of  American  Foresters 

G.  Foster 

Society  of  American  Military  Engineers 

VACANT 

Society  of  Experimental  Biology  and  Medicine 

Darren  Roesch 

Society  of  Manufacturing  Engineers 

VACANT 

Soil  and  Water  Conservation  Society 

Bill  Boyer 

Technology  Transfer  Society 

Clifford  Lanham 

Washington  Evolutionary  Systems  Society 

Jerry  L.R.  Chandler 

Washington  History  of  Science  Club 

Albert  G.  Gluckman 

Washington  Chapter  of  the  Institute  for  Operations 

Research  and  Management  Science 

Russell  Wooten 

Washington  Paint  Technology  Group 

VACANT 

Washington  Society  of  Engineers 

Alvin  Reiner 

Washington  Statistical  Society 

Michael  P.  Cohen 

World  Future  Society 

Russell  Wooten 

MEDIA  MAIL 


Washington  Academy  of  Sciences 
Room  637 

1200  New  York  Ave.  NW 
Washington,  DC  20005 
Return  Postage  Guaranteed 


ERNST  HAYR  LIBRARY 
WUSEUtt  COMP  ZOOLOGY 
HARVARD  UNIVERSITY 
26  OXFORD  ST 
CAHBRIBGEr  MA  02138-2902 


\jor^ 

$333. 


MCZ 

library 

Volume  92 
Number  4 

WAR  1  3  2007 

Winter  2007 

Journal  of  the 

HARVARD 

UNIVERSITY 

WASHINGTON 

ACADEMY  OF  SCIENCES 

Contents 


The  Editor  Comments  .  i 

Instructions  to  Authors  . ii 

Athanasios  Diamandopolous,  Artemisia  Revisited . 1 

Robert  M.  Cutler,  The  Paradox  of  Intentional  Emergent  Coherence  . 9 

Roulette  William  Smith,  Nature  versus  Nurture  in  Evolution,  and  Emergences  of  Designs  in 
Genetics,  the  Immune  System,  and  the  Brain  . 29 

Thomas  Meylan,  Environmental  Impacts  on  Human  Moods  and  Emotions:  Implications  for 
Workplace  and  Workflow  Design . 51 

Vary  Coates,  BOOK  REVIEW:  The  Best  Science  Writing  2006,  ed.  by  Atul  Gawande  . 73 

News  of  Members,  Fellows,  and  Affiliates . 75 

Affiliated  Institutions  . 78 

Membership  Directory . 79 

Affiliated  Societies . Inside  back  cover 


ISSN  0043-0439 


Issued  Quarterly  at  Washington  DC 


l^asiJjtngton  Scabemp  of  Sciences! 

Founded  in  1898 


Board  of  Managers 

Elected  Officers 

The  Journal  of  the  Washington  Academy  of 
Sciences 

President 

William  Boyer 

President  Elect 

Alain  Towaide 

Treasurer 

Harvey  Freeman 

Secretary 

James  Cole 

Vice  President,  Administration 

Rex  Klopfenstein 

Vice  President,  Membership 

Thomas  Meylan 

Vice  President,  Junior  Academy 

Paul  L.  Hazan 

Vice  President,  Affiliated  Societies 
Mark  Holland 

Members  at  Large 

Sethanne  Howard 

Donna  Dean 

Frank  Haig,  S.J. 

Jodi  Wesemann 

Vary  Coates 

The  Journal  is  the  official  organ  of  the  Academy. 
It  publishes  articles  on  science  policy,  the  history  of 
science,  critical  reviews,  original  science  research, 
proceedings  of  scholarly  meetings  of  its  Affiliated 
Societies,  and  other  items  of  interest  to  its  members. 
It  is  published  quarterly.  The  last  issue  of  the  year 
contains  a  directory  of  the  current  membership  of 
the  Academy. 

Subscription  Rates 

Members,  fellows,  and  life  members  in  good 
standing  receive  the  Journal  free  of  charge. 
Subscriptions  are  available  on  a  calendar  year  basis, 
payable  in  advance.  Payment  must  be  made  in  U.S. 
currency  at  the  following  rates. 

US  and  Canada .  $25.00 

Other  Countries  .  30.00 

Single  Copies  (when  available)  .  10.00 

Claims  for  Missing  Issues 

Claims  must  be  received  within  65  days  of  mailing. 
Claims  will  not  be  allowed  if  non-delivery  was  the 
result  of  failure  to  notify  the  Academy  of  a  change 

Peg  Kay  of  address. 

Past  President:  F.  Douglas  Witherspoon 

Notification  of  Change  of  Address 


AFFILIATED  SOCIETY  DELEGATES: 
Shown  on  back  cover 

Address  changes  should  be  sent  promptly  to  the 
Academy  Office.  Notification  should  contain  both 
old  and  new  addresses  and  zip  codes. 

Editor  of  the  Journal 

Vary  T.  Coates 

Associate  Editors: 

Alain  Touwaide 

Sethanne  Howard 

Elizabeth  Corona 

POSTMASTER: 

Send  address  changes  to  WAS,  Rm.631, 

1200  New  York  Ave.  NW 

Washington,  DC.  20005 

Journal  of  the  Washington  Academy  of  Sciences 
(ISSN  0043-0439) 

Academy  Office 

Washington  Academy  of  Sciences 
Room  63 1 

1200  New  York  Ave.  NW 
Washington,  DC  20005 

Phone:  202/326-8975 
email:  was@washacadsci.org 

Published  by  the  Washington  Academy  of  Sciences 
202/326-8975 

website:  www.washacadsci.org 

THE  EDITORS  COMMENT: 


MCZ1 

LIBRARY 

MAR  1  3  2007 


HARVARD 

A  DIRECTORY  OF  THE  FELLOWS  AND  MEMBERS  of  WA$^^?tF?SlTY 
as  is  traditional,  in  the  Winter  Issue  of  the  Journal.  The  dues, 
contributions,  and  volunteer  labor  of  these  members  support  all  of  the 
activities  of  WAS:  for  example,  the  publication  of  the  Journal,  the 
planning  and  organization  of  the  biannual  Capital  Science  Conferences 
and  other  specialized  conferences  throughout  the  year,  and  the  Junior 
Academy  STARS  program  that  provides  judges  and  awards  for  many  local 
school  science  fairs. 

We  take  this  opportunity  to  make  a  special  appeal  to  the  Members 
and  Fellows — for  help  in  reaching  out  to  the  many  scientific  societies  in 
the  Washington  Metropolitan  Area  in  order  better  to  carry  out  our  goal  of 
showcasing  and  encouraging  the  work  of  local  scientists,  engineers,  and 
science  teachers.  We  seek  to  do  this  by  publishing  their  papers  in  this 
peer-reviewed  Journal,  and  by  organizing  conferences  and  seminars  on 
topics  of  special  interest  to  them. 

Most,  if  not  all,  WAS  members  belong  to  local  chapters  of 
scientific  societies,  nearly  60  of  which  are  formally  affiliated  with  WAS 
(see  the  list  on  the  inside  back  cover  of  the  Journal).  The  Affiliates’  other 
members,  however,  are  not  always  familiar  with  the  Academy  and  with  its 
activities,  or  with  the  Journal.  We  ask  that  when  you  go  to  meetings  of 
these  societies,  you  take  one  or  more  issues  of  the  Journal  to  display,  and 
invite  the  attendees  to  subscribe,  and  to  contribute  papers.  The  WAS 
office  can,  on  request,  supply  some  additional  copies  of  the  Journal  for 
this  purpose. 

We  thank  our  members  and  subscribers  for  their  participation  and 
support  throughout  2006,  and  look  forward  to  working  for  you  and  with 
you  in  2007-2008.  Please  contact  the  Editors  to  submit  papers  or  book 
reviews,  to  send  news  items  about  your  own  or  others’  science  related 
activities,  and  to  suggest  ways  to  improve  the  Journal. 

Editor:  Vary  Coates,  v coates@ m ac . com 

Assoc.  Editor:  Alain  Touwaide,  atouwaide@hotmail  com 

Assoc.  Editor:  Sethanne  Floward,  sethanneh@msn.com 

Assoc.  Editor:  Elizabeth  Corona,  el i zab eth corona@ mnai  1 . com 


Winter  2006 


II 


INSTRUCTIONS  FOR  AUTHORS 


THE  JOURNAL  of  the  Washington  Academy  of  Sciences  is  a 
peer-reviewed  journal.  Exceptions  are  made  for  papers  requested 
by  the  editors  or  positively  approved  for  presentation  or 
publication  by  one  of  our  affiliated  scientific  societies. 

We  welcome  disciplinary  and  interdisciplinary  scientific  research 
reports  and  papers  on  technology  development  and  innovation, 
science  policy,  technology  assessment,  and  history  of  science  and 
technology.  Book  reviews  are  also  welcome. 

Contributors  of  papers  are  requested  to  follow  these  guidelines 
carefully. 

•  Papers  should  be  submitted  as  e-mail  attachments  to  the  chief 
editor,  vcoatesc/  mac.com.  along  with  full  contact  information  for 
the  primary  or  corresponding  author. 

•  Papers  should  be  presented  in  Word:  do  not  send  PDF  files. 

•  Papers  should  be  6000  words  or  fewer.  If  more  than  6  graphics  are 
included  the  number  of  words  allowed  will  be  reduced 
accordingly. 

•  Graphics  must  be  in  black  and  white  only.  They  must  be  easily 
resized  and  relocated.  It  is  best  to  put  graphics,  including  tables,  at 
the  end  of  the  paper  or  in  a  separate  document,  with  their  preferred 
location  in  the  text  clearly  indicated. 

•  References  should  be  in  the  form  of  endnotes,  and  may  be  in  any 
style  considered  standard  in  the  discipline(s)  represented  by  the 
paper. 


Winter  2006 


1 


ARTEMISIA  REVISITED 


Diamandopoulos  Athanasios* 

Saint  Andrews  Regional  Hospital 
Patras.  Greece 


Abstract 

Recently,  several  groups  have  promoted  the  use  of  artemisinin,  a 
derivative  of  the  plant  artemisia.  as  a  potent  antimalarial.  It  is  cultivated 
widely  in  China  and  there  is  hope  that  it  could  replace  quinine  to  which 
plasmodium  falciparum  has  developed  resistance.  In  this  article,  we 
trace  the  use  of  artemisinin  for  treating  malaria  as  far  back  as  the  1st 
century  AD.  in  the  Eastern  Mediterranean  Basin.  Its  employment  for 
this  purpose  continued  in  a  semi-official  mode  until  the  19th  century.  It 
was  finally  replaced  by  quinine,  albeit  now  we  witness  a  comeback.  In 
conclusion,  artemisinin  is  but  an  old  drug  of  Western  medical  tradition 
in  new  exotic  clothes,  still  potent  and  beneficial. 


The  use  of  Artemisinin,  a  derivative  of  the  plant  Artemisia,  as 
an  anti-malaria  drug,  has  recently  gotten  extended  publicity  both  in  the 
daily  [1,2]  and  the  scientific  press  [3,4],  Specifically,  there  was  strong 
criticism  of  the  World  Health  Organization  (WHO)  on  the  grounds  that  it 
failed  to  foresee  in  time  the  potential  of  the  substance  and  use  it  in  Africa, 
where  the  disease  is  endemic  and  frequently  fatal  [5,6,7],  In  summary,  it 
was  pointed  out  that  the  Organization  “ignored  a  new  Chinese  drug 
proved  efficient  to  treat  malaria  since  the  Vietnam  War  [8],  favoring  the 
old  drug,  quinine.” 

Although  from  a  strictly  medical  point  of  view  the  statement  seems 
sound,  from  a  historical  perspective  we  suggest  that  all  three  components 
of  the  statement  should  be  altered.  The  drug  is  not  new,  it  is  not 
exclusively  Chinese,  and  its  efficiency  against  malaria  and  other  parasitic 
diseases  has  been  well  known  since  antiquity.1  Malaria  has  been  an 
endemic  disease  in  Greece  and  the  Mediterranean  Basin  since  the 
Neolithic  period,  with  exacerbations  in  the  Classic  and  Roman  eras. 


The  author  is  President  of  the  Panhellenic  Society  for  the  History  and  Archaeology  of 
Medicine. 


Winter  2006 


2 


Hippocrates  in  his  works  De  aere  aquis  et  locis  and  De  mortis 
popularity s  made  several  references  to  tertian  and  quartian  fevers 
accompanied  with  symptoms  resembling  the  malarial  ones.  Similarly, 
Plato  referred  to  this  kind  of  fevers  in  Timeus.  An  extended  overview  of 
the  incidence  of  malaria  in  Greece  can  be  found  in  the  work  of  Mirko 
Grmek,  “Les  maladies  a  l’aube  de  la  civilisation  occidental  ”  [9] 

No  specific  treatment  was  mentioned  in  the  Classic  period, 
although  there  are  allusions  related  to  the  plant  artemisia  in  the 
mythological  cycle  of  the  god  Dionysus.  [10]  An  early  medical  reference 
to  it  was  made  by  the  celebrated  Greek  medical  writer  Dioscorides  in  his 
work  De  Materia  Medica,  in  the  1st  cent.  AD,  wherein  its  properties  as  a 
warming,  drying  and  purgative  drug  are  described.  [11]  The  plant  was  also 
called  oxetesia,  ephesia,  anactorios,  sozusa,  lea,  lycophrys,  sanguis 
hominis,  chrysanthemon,  herba  regia,  rapium,  tetrabageta,  ponem,  zuoste. 
[12]  It  was  synonymous  and/or  likened  to  absinth.  [13]  It  was 
recommended  for  many  diseases  such  as  parasites  of  the  bowel,  kidney 
stones  and  intended  miscarriage  of  dead  embryos.  More  relevant  to  the 
topic  of  this  article  is  that  Dioscorides  also  proposed  it  as  a  cure  for  tertian 
or  quartian  fevers,  [14,15]  terms  that  are  still  used  to  describe  malarial 
fever. 


If  there  is  any  doubt  about  the  inclusion  of  malaria  in  the  array  of 
diseases  manifested  with  this  kind  of  fever,  Galen  (lst/2nd  cent.  AD) 
clarifies  the  issue.  He  proposed  the  use  of  artemisia  against  tertian  and 
quartian  fevers  and  describing  other  findings  of  the  disease  he  added:  “it  is 
accompanied  by  [.  .  .]  infection  and  enlargement  of  the  spleen,”  [16] 
findings  typical  for  malaria.  [17] 

We  can  hardly  expect  the  use  of  the  actual  word  malaria  in  texts 
written  20  centuries  ago.  The  notion  was  repeated  by  many  medical 
writers  of  Late  Antiquity  and  Byzantium  [18,19,20,21,22]  who 
continuously  suggested  the  use  of  artemisia  in  tertian  and  quartian  fevers. 
It  was  also  recommended  specifically  against  the  hardening  of  the  spleen. 
[23]  The  practice  of  using  artemisia  as  a  medicament  passed  via 
Dioscorides’  De  Materia  Medica  in  the  Latin  West  and  then  in  the 
Medieval  Herbaria.  With  the  invention  of  printing  in  the  15th  century, 
ancient  medical  knowledge  spread  quickly  across  Europe  and 
consequently  did  so  the  use  of  artemisia.  [24] 


Washington  Academy  of  Sciences 


3 


With  the  appearance  of  the  Enlightenment  and  the  domination  of 
scientific  over  traditional  medicine,  the  use  of  artemisia  as  an  antimalarial 
agent  started  to  fade.  It  was  preserved  in  some  odd  quarters,  like  the 
“kitchen  -  medicine”  of  the  Henriot  sisters  in  the  Swiss  town  of  Covet, 
where  they  prepared  at  their  tiny  still  on  a  kitchen  stove  a  concoction 
containing,  between  other  ingredients,  absinth,  otherwise  known  as 
artemisia,  in  the  time  honoured  way  of  the  “wise  women  ”  It  was 
consumed  as  a  general  panacea  for  many  diseases  or  just  as  a  health 
booster.  A  French  royalist,  the  physician  Pierre  Ordinaire  (1741-1821), 
staying  in  self-exile  in  the  village,  was  acquainted  with  the  drink  and 
began  peddling  it.  The  concoction  was  called  Absinthe  as  the  Latin  name 
of  one  of  its  ingredients  (wormwood,  meaning  in  German  “preserver  of 
the  mind”)  was  Artemisia  Absinthum .  Later  (in  1797),  the  recipe  was 
bought  by  a  Major  Henri  Dubied  and  through  his  enterprising  heirs  who 
opened  a  larger  distillery  at  Pontarlier,  France,  was  introduced  to  the 
French  illustrious  society,  initiating  the  absinth  frenzy.  [25]  This  was  a 
repetition  of  the  ancient’s  idea  that  the  substance  was  simultaneously  a 
potent  stimulant  and  an  agent  promoting  soberness  to  its  consumer,  while 
he  was  drinking  heavily.  [26] 

The  consumption  of  liquors  or  other  mixtures  containing  absinth 
was  consequently  banned  from  Europe  because  of  its  side  effects,  except 
for  a  few  countries  like  Denmark  and  Czechoslovakia.  Through  the 
Internet  it  started  again  to  be  advertised  world  -  wide  as  a  symbol  of 
culture  and  invigoration.  [27]  Traditionally,  it  belongs  to  the  bitter  drinks 
group  like  Vermouth  and  Amaretto,  the  latter  containing  bitter  almonds 
instead  of  absinth.  Galen  had  already  suggested  the  replacement  of  absinth 
by  bitter  almonds,  if  there  was  a  lack  of  the  former  [28], 

Recently,  the  Chinese  promoted  it  again  as  an  ancient  plant  from 
their  traditional  pharmacopoeia,  reputably  incorporated  in  the  Sheng 
Nung’s  writings  circa  101  BC,  as  introduced  by  the  Divine  Ploughmen 
around  2800  BC.  [29]  They  cultivated  it  widely,  extracting  its  active 
substance  artemisinin,  which  is  a  sesuiterpene  with  five  oxygen  atoms, 
two  of  them  in  a  peroxide  bridge  system  over  a  seven-member  ring  with 
two  others  in  a  lactone  ring  structure  [30]  and  branding  it  under  the  name 
of  artemisinin. 

In  May  2001,  the  big  drug  company  Novartis  made  a  public- 
private  collaboration  agreement  with  the  World  Health  Organization 


Winter  2006 


4 


(WHO)  in  the  fight  against  malaria.  The  essence  of  the  agreement  is  that 
Novartis  commits  to  making  Coartem®  (a  drug  containing  a  derivative  of 
artemisia)  available  on  a  "not-for-profit"  basis  for  distribution  to  public 
sector  agencies  of  malaria-endemic  developing  countries.  Through  grants 
provided  by  “The  Global  Fund  to  fight  AIDS,  Tuberculosis  and  Malaria”,  j 
Novartis  has  equally  undertaken  to  supply,  under  the  aegis  of  WHO, 
Coartem®  to  public  sector  agencies  [3 1  ^Consequently,  we  do  not  meet 
the  problem  of  a  new  drug  facing  difficulties  in  ousting  an  old,  established 
one,  but  on  the  contrary,  a  very  old  drug,  trying  hard  to  reclaim  its 
position  from  the  usurper  quinine.  It  is  characteristic  that  currently  there 
are  roughly  seventeen  thousand  entries  on  the  Internet  discussing 
artemisia,  one  thousand  four  hundred  proclaiming  its  antimalarial 
properties  and  only  one  connecting  it  with  Dioscorides,  but  failing  to 
notice  the  indication  in  De  Materia  Medica  for  treating  tertian  and 
quartian  fevers. 

It  is  the  usual  “I  said  it  first”  syndrome  we  observe  in  the  History 
of  Medicine.  [32]  Everyone  rediscovering  an  abandoned  treatment 
advertises  it  loudly,  ignoring  or  forgetting  to  state  that  many  before  him 
had  used  it.  But  the  really  great  scientist  always  announces  his  debt  to 
previous  writers.  As  for  example  Isaac  Newton  (1642-1727)  who  wrote: 

“If  I  have  seen  further  it  is  by  standing  on  the  shoulders  of  giants.” 
However  he  forgot  to  add  that  the  same  expression  has  been  used  by 
Robert  Burton  (1577-1640)  in  his  introduction  to  “Democritus  to  the 
Reader,”  who  did  not  mention  that  it  belonged  to  Diego  de  Estella’s 
(1524-1578)  “In  Sacrosanctum  Iesu  Christi  Evangelium  secundum  Lucam 
enarratio,”  who  failed  to  report  that  it  had  been  used  by  Bernard  de 
Chartres  (d.  1126)  [33],  In  spite  of  the  above  implication  of  practising 
plagiarism  it  would  be  wrong  to  assume  that  modem  researchers  in 
general,  and  the  advocates  of  artemisinin  in  particular,  do  not  do  anything 
more  than  copy  the  ancients.  Because:  “even  if  the  ancients  had 
discovered  everything,  one  thing  will  be  always  new,  the  application  of 
the  discoveries  already  made  and  their  interpretation.”  [34] 


1  Etymologically.  The  botanical  species  name  of  wormwood,  absinthium ,  is  indeed  the 
classical  Latin  name  for  that  plant  and  derives  from  Greek  apsinthion  [a\|/iv0iov]  (in  the 
New  Testament  apsinthos  [a\j/iv0og]):  the  word  still  lives  in  some  Romance  tongues: 
Italian  assenzio .  Spanish  ajenjo ,  Galician  axenxo  and  Portuguese  absinto.  It  lias  also 
spread  to  some  unrelated  languages,  like  Basque  axinse  and  Hebrew  absint  [ttf’DDKn]. 


Washington  Academy  of  Sciences 


5 


The  etymology  of  Greek  apsinthion  is  not  clearly  explained:  a  theory  derives  it  from  a- 
(negation)  +  psinthos  |vj/iv0oq].  an  obscure  adjective  meaning  "enjoyable.”  cf.  also 
Sanskrit  ashiva  "unpleasant,  pernicious  ".  The  meaning  of  the  compound,  "unpleasant,"" 
would  seem  fit  for  a  bitter  herb,  but  may  well  be  the  product  of  folk  etymology.  A  better 
guess  is  that  tire  name  actually  stems  from  some  Middle  Eastern  language:  in  Middle 
Persian,  the  name  aspemd  is  recorded  for  a  bitter  plant  (perhaps  Syrian  rue.  Peganum 
harmala ):  modem  Farsi  has  afsentin  [  "wormwood"'  and  espemd  [>hH]  ""Syrian 

me".  This  plant  is  not  related  to  the  herb  commonly  called  rue.  (http://www.uni- 
eraz.atATatzer/end/Artc  vul.html#absinthe) 


REFERENCES 

1.  BBC  Homepage.  20/5/2004.  http :/Avw w . bbc . co.uk/dna/h2 g2/ A 7 84046 

2.  DJ  McNeil.  Switch  to  herbal  malaria  drug  is  on.  Herald  Tribune .  (10/5/2004).  p.  1. 

3.  G  Yamey.  Health  agencies  and  in  -  fighting  on  malaria.  BMJ  (2004),  328: 1095  (8/5) 

4.  MB  Denis.  TM  Davis.  S  Hewitt,  et  al..  Efficacy  and  safety  of  dihydroartemisinin  - 
piperaquine  ( Artekin)  in  Cambodian  children  and  adults  with  uncomplicated  falciparum 
malaria.  Clin  Infect.  Dis.  35  (2002).  pp  1469  -  1476. 

5.  N  White.  F  Noston.  A  Bjorkman  et  al.  WHO.  the  Global  Fund,  and  medical 
malpractice  in  malaria  treatment.  Lancet.  363  (2004),  p.  1 160. 

6.  A  Attaran.  KI  Bames.  C  Curtis  et  al.  WHO.  The  global  Fund,  and  medical  malpractice 
in  malaria  research.  Lancet.  363  (2004).  P.  237. 

7.  K  Bames.  J  Mwenechanya,  M  Tempo,  et  al.  Efficacy  of  rectal  artesunate  compared 
with  parental  quinine  in  initial  treatment  of  moderately  severe  malaria  in  African  children 
and  adults:  a  randomized  study.  Lancet.  363  (2004)  p.  1598. 

8.  T  Hien,  C  Dolescek.  P  Phuong  et  al.  Dihydroartemisinin  -  piperaquine  against 
multidrug  -  resistant  Plasmodium  falciparum  malaria  in  Vietnam:  randomized  clinical 
trial.  Lancet.  363  (2004).  Issue  9402.  p.  18  -  22. 

9.  M  Gnnek.  Les  maladies  a  l'aube  de  la  civilisation  occidentale.  Payot  (edt)  (1983). 
Paris. 

10.  Bibliotheca,  ed.  R  Henry.  Photius.  Bibliotheque.  8  vols.  Paris:  Les  Belles  Lettres. 
1:1959:  codex.  190,  p,150a.*  1.  27. 

11.  De  materia  medica,  ed.  M.  Wellmann.  Pedanii  Dioscuridis  Anazarbei  de  materia 
medica  libri  quinque.  3  vols.  Berlin:  Weidmann.  3:1914.  book  3.  ch.  113.  s.  1. 1.  1. 


Winter  2006 


6 


12.  R  Gunther.  The  Greek  Herbal  of  Dioscorides.  illustrated  by  a  Byzantine  A.D.  512, 
Englished  by  Jolm  Goody er  A.D.  1655.  Edited  and  first  printed  A.D.  1933.  Oxford 
University  Press  (1934).  Oxford,  p.  357. 

13.  De  materia  medica.  ibid,  book  4.  eh.  60. 1.1. 

14.  De  materia  medica.  ibid,  book  4.  ch.  60.  s.  1. 1.  6. 

15.  Euporista  vel  De  simplicibus  medicinis.  ed.  M.  Wellmann.  Pedanii  Dioscuridis 
Anazarbei  de  materia  medica  libri  quinque.  vol.  3.Berlin,  1914  (repr.  1958):  book  2,  ch. 
19.  s.  1.1.  5 

16.  De  tv pis  liber,  ed.  C.G.  Kohn.  Claudii  Galeni  opera  omnia,  vol.  7.  Leipzig: 

Knobloch.  1824  (repr.  Hildesheim:  Olms.  1965):  463-474  book.  7,  p.  469.  L  18. 

17.  Harrison's  Principles  of  Internal  Medicine.  6th  edition  (1970).  McGraw  -  Hill  Book 
Company.  New  York  etc.  pp  1030  -  1034. 

18.  Iatricorum  liber  v.  ed.  A.  Olivieri,  Aetii  Amideni  libri  medicinales  v-viii  [Corpus 
medicorum  Graecorum.  vol.  8.2.  Berlin:  Akademie-Verlag.  1950]:  6-119.  ch.  80. 1.  7 

19.  Eclogae  medicamentorum.  ed.  J.  Raeder.  Oribasii  collectionum  medicarum  reliquae. 
vol.  4  [Corpus  medicorum  Graecorum.  vol.  6.2.2.  Leipzig:  Teubner.  1933]:  185-307.  b. 
45.  s.  6.1.  1. 

20.  De  febribus.  ed.  T.  Puschmann.  Alexander  von  Tralles,  vol.  1.  Vienna:  Braumoller. 
1878  (repr.  Amsterdam:  Hakkert.  1963):  291-439.  (Cod:  20.743:  Med.)  1.  373.4 

21.  Epitomae  medicae  libri  septem.  ed.  J.L.  Heiberg.  Paulus  Aegineta.  2  vols.  [Corpus 
medicorum  Graecorum,  vols.  9.1  &  9.2],  Leipzig:  book.  2.  ch.  19.  p.  1  1.  6. 

22.  De  virtutibus  herbarum  (e  cod.  Paris,  gr.  2502  and  Vindob.  med.  gr.  23.  ed.  H.-V. 
Friedrich.  Thessalos  von  Tralles  [Meisenheim  am  Gian:  Hain.  1968]:  43-44.  56.  59.  b  1. 
ch.  8,  s.  t  l.L 

23.  Synopsis  ad  Eustathium  filium,  ed.  J.  Raeder.  Oribasii  synopsis  ad  Eustathium  et  libri 
ad  Eunapium  [Corpus  medicorum  Graecorum.  vol.  6.3].  Leipzig:  Teubner.  1926  (repr. 
Amsterdam:  Hakkert.  1964). 

24.  G  Penso.  Les  Plantes  Medicinales.  Roger  Dacosta  (Edt)  (1986).  Paris,  p.  101  and 
139. 

25.  Man  and  Scythe  Inn  absinthe  page:  http ://w w w . Manandscvthc.co.uk/absinte.htm 

26.  Libri  ad  Eunapium,  ibid.  1.  12.  5.  1. 

27.  Welcome  to  Bar  Absinthe.  Find,  buy  and  appreciate  great  absinthe  here. 

http :  /Avww .  bar-absinthe  .co  m/ 


Washington  Academy  of  Sciences 


7 


28.  De  succedaneis  liber,  ed.  C.G.  Kuhn.  Claudii  Galeni  opera  omnia,  vol.  19.  Leipzig: 
Knobloch.  1830  (repr.  Hildesheim:  Olms.  1965):  721-747. 

29.  M  R.  Lee.  Plants  against  malaria.  Part  2.  Artemisia  Annua.  Qingshaosu  or  the  SW 
Wormwood.  J.  R.  Coll.  Physicians  Edinb  2002.  32:300 

30.  hi tp :/Av w vv.ifpma . ora/Heal tli/ma la ria/health  coartem  ma  1 . aspx 

31.0.  Famin.  H  Ginsburg.  Differential  effects  of  4  -  aminoquinine  -  containing 
antimalarial  drugs  on  hemoglobin  digestion  in  plasmodium  falciparum-infected 
erythrocytes.  Biochemical  Pharmacology  (2002).  63  (3):  393 

32.  A  Diamandopoulos.  The  use  of  Ancient  and  Medieval  Greek  literature  for  avoiding 
the  «I  said  it  first»  research  syndrome.  Proceedings  of  the  2nd  World  Congress  '‘Ancient 
Greece  and  the  Modem  World  '.  Olympia.  12-17  July  2002.  University  of  Patras  Press. 
Patras.  2003.  p.  40.  (in  Greek). 

33.  P  Prioreschi.  The  idea  of  progress  in  Antiquity  and  in  the  Middle  Ages.  Vesalius 
(2002).  VIII.  1.34-45. 


34.  Seneca.  Ad  Lucilium  epistulae  morales.  64.  vii.  ix.  in:  Prioreschi  P..  ibid.  p.  36. 


Winter  2006 


This  page  intentionally  left  blank 


Washington  Academy  of  Sciences 


9 


THE  PARADOX  OF  INTENTIONAL  EMERGENT 
COHERENCE: 

ORGANIZATION  AND  DECISION  IN  A  COMPLEX  WORLD 

Robert  M.  Cutler1 

Institute  of  European  and  Russian  Studies.  Carleton  University 


Abstract 

The  work  presented  here  draws  upon  previous  research  into  the 
development  of  a  particular  type  of  international  organization,  the 
international  parliamentary  institution  (IPI).  of  which  the  European 
Parliament  is  the  best-known  example.  It  generalizes  the  concepts 
framing  that  research  to  organizations  at  large,  including  but  not  limited 
to  political  nation-states.  By  explicating  how  that  framework  is 
grounded  in  the  theory  of  complex  systems,  the  present  article  shows 
how  it  is  applicable  to  social  systems  in  general.  The  tasks  facing  IPIs 
seeking  to  survive  and  grow,  on  the  one  hand,  and  on  the  other  hand  the 
tasks  facing  states  responding  to  the  international  political  environment, 
are  in  fact  similar  in  cybernetic  terms.  Since  only  the  resources 
available  to  them  differ,  the  organizational  and  human  information¬ 
processing  phenomena  that  form  and  constrain  state  foreign  policy 
decision-making  are  comparable  to  those  that  are  expressed  in  the 
epigenetic  development  of  an  IPI.  Since  the  general  framework 
presented  here  provides  a  way  to  take  into  account  the  differences  in 
resources  to  which  states  and  IPIs  have  access,  it  is  applicable  by 
extension  to  any  political  or  social  system  or  organization  that 
continually  confronts  an  external  environment  that  it  itself  helps  to 
shape  through  interpretation  as  well  as  action.  This  general  framework 
may  therefore  be  properly  applied  not  just  to  analyzing  the  behavior  of 
formal  social  organizations  constmcted  by  human  collectivities  for 
defined  political  purposes,  but  also  to  the  analysis  of  decisions,  growth, 
and  development  of  individual  human  beings  in  life  itself. 


“Complexity  Science”  or  “The  Complex  Sciences”? 


Complex-science  or  “complex-scientific”  studies,  like  structuralist 
studies,  are  an  approach  to  the  creation  of  knowledge.  Consequently,  it  is 
more  appropriate  to  speak  of  “the  complex  sciences”  than  of  “complexity 
science.”  The  former  locution  makes  it  clear  that  the  complex  sciences  are 


Winter  2006 


10 


not  a  delimited  set  of  fields  of  knowledge  to  be  explored  from  either  a 
complexity  standpoint  or  a  non-complexity  standpoint,  but  rather  the 
manifestation  of  one  perspective  on  the  world  and  knowledge-creation 
about  the  world.  In  order  to  underline  this  distinction,  for  the  purpose  of 
the  present  article  the  compound  attributive  adjective  “complexity- 
science”  (as  in  “complexity-science  approaches”)  is  replaced  by 
“complex-scientific”  (thus  “complex-scientific  approaches”). 

It  is  necessary  to  explicate  the  distinctive  characteristics  of  the 
complex  sciences  in  order  to  ascertain  what  the  limits  to  knowledge  about 
them  are.  The  best  vehicle  for  that  explication  is  by  analogy  to  the 
exegesis  of  Levi-Strauss’s  structuralism  by  Piaget,  who  shows  how 
different  applied  structuralisms  within  various  fields  of  knowledge  are 
conditioned  by  the  sociologies  of  knowledge  constructing  these 
disciplinary  fields  of  study,  which  in  turn  vary  across  time  and  space,  even 
within  the  same  field  of  knowledge.  The  three  central  notions  of  Piaget’s 
exegesis  of  Levi  -Strauss  are  totality ,  self-regulation ,  and  transformation . 
Defining  these  three  categories  as  the  components  of  a  structure,  Piaget 
distinguishes  how  they  manifest  and  differentiate  comparative 
structuralisms  across  fields  of  knowledge  from  mathematics  to 
anthropology,  passing  through  the  natural  sciences,  life  sciences,  and 
social  sciences.11 

Winch’s  application  of  a  Wittgensteinian  approach  to  social 
science  sets  Piaget’s  notion  of  structuralism  into  relation  with  the 
epistemology  of  complex  systems.  In  particular.  Winch  explains  how  any 
social  science  may  construct  its  epistemology  by  establishing  the 
categories  of  structure,  norms,  and  behavior,  and  using  any  two  of  these  to 
study  the  third.111  Inspection  of  Piaget’s  exegesis  of  Levi-Strauss  reveals 
“totality”  as  the  principal  characteristic  of  a  given  structure ,  “self¬ 
regulation”  as  principally  characterizing  norms  (since  it  is  according  to 
norms  that  such  self-regulation  occurs),  and  “transformation”  as  a 
characterization  principally  of  behcnior  (involving  change  over  time  and 
therefore  differential).  In  other  words,  structures  define  what  is  possible 
while  norms  operate  within  structural  constraints  so  as  to  generate  that 
which  actually  manifests  in  the  world.  So  it  is  that  norms  operate  upon 
structure  to  produce  behavior;  or,  put  another  way,  structure  is  mediated 
through  norms  into  behavior.  Stated  with  a  still  greater  degree  of 
generality:  Totality  is  akin  to  a  domain,  self-regulation  akin  to  a  function, 
and  transformation  akin  to  a  range:  the  laws  of  self-regulation  act  upon  the 
totality  and  result  in  transformation.1'' 


Washington  Academy  of  Sciences 


11 


Complex-scientific  approaches  to  knowledge-creation  are  the 
constructive  response  that  goes  beyond  the  critical  destructiveness  of 
“post-structuralism. ”v  Building  blocks  of  the  complex  sciences  include 
three  fundamental  categories  that  are  extensions  of  the  three  categories 
forming  the  basis  for  the  structuralist  epistemology  that  Piaget  extracts 
from  Levi-Strauss.  In  particular:  (l)the  complex-scientific  extension  of 
the  structuralist  category  of  “totality”  is  coherence ,  (2)  the  complex- 
scientific  extension  of  the  structuralist  category  of  “self-regulation”  is 
autopoiesis ,V1  and  (3)  the  complex-scientific  extension  of  the  structuralist 
category  of  “transformation”  is  emergence.  However,  Winch’s 
epistemology  as  applied  to  the  structuralist  social  sciences  does  not 
provide  the  basis  for  extension  to  the  epistemology  of  the  complex 
sciences;  specifically,  in  the  complex  sciences  it  is  not  the  case  that  any 
two  fundamental  categories  (among  coherence,  autopoiesis,  and 
emergence)  can  be  used  to  study  the  third.  Rather,  autopoiesis  mediates 
coherence  and  emergence. 

There  are  three  interrelated  approaches  to  the  modern  study  of 
complex  systems,  each  focusing  on  one  of  the  components  of  a  social 
science  as  enumerated  above  according  to  Winch:  (1)  how  interactions 
give  rise  to  patterns  of  behavior ,  a  largely  North  American  approach 
typified  by  an  emphasis  on  “complex  adaptive  systems”; 
(2)  understanding  the  different  ways  in  which  complex  systems  may  be 
normative/y  described ,  a  mostly  European  approach  characteristic  of  the 
natural  sciences  and  typified  by  Prigogine  and  the  approach  to 
thermodynamics;  and  (3)  the  process  of  structural  formation  of  complex 
systems  through  pattern  formation  and  evolution,  a  cybernetics-based  and 
system-theory-oriented  approach  adopted  in  both  Europe  and  North 
America.™  The  present  article  is  situated  in  the  tradition  of  the  third  of 
these  approaches,  which,  in  Winch’s  terms,  combines  the  study  of 
behavior  and  of  norms  in  order  to  explain  structure. 


Functionalism  and  Organizational  Development 

Within  this  general  approach,  the  present  article  establishes  a 
framework  for  evaluating  the  growth  and  decline  of  organizations  and 
other  social  systems,  determining  what  leads  some  of  them  to  respond 
adequately  to  demands  imposed  upon  them  by  their  environment,  and 
others  not.  The  framework  synthesizes  two  apparently  mutually  exclusive 
taxonomies:  one  concerning  how  organizations  maintain  homeostasis  in 


Winter  2006 


12 


order  to  survive,  the  other  about  how  organizations  develop  and  adapt  in 
order  to  grow.'111  Table  1  summarizes  the  first  taxonomy,  which  concerns 
how  organizations  survive;  it  emphasizes  the  creation  of  organizational 
structures  so  as  to  accomplish  functional  tasks;  it  comprises  the  two 
principal  categories,  internal  functions  and  external  functions.  The  order  in 
which  any  organization  accomplishes  the  internal  functions  in  fact  defines 
an  evolutionary  sequence:  (1)  informational  activities,  (2)  normative 
activities,  (3)  rule-creating  activities,  (4)  rule-supervisory  activities,  and 
(5)  operational  activities.  Only  organizations  that  successfully  perform 
lower-numbered  internal  functions  have  the  opportunity  to  move  on  to  the 
higher-numbered.  “Operational  activities”  are  activities  undertaken  with 
reference  to  the  physical  and  institutional  environment.  They  represent  the 
spillover  from  the  full  development  of  internal  functions  to  the 
deployment  of  external  functions. 1X 


Internal  Functions 

External  Functions 

1)  Informational  activities 

2)  Normative  activities 

3)  Rule-creating  activities 

4)  Rule-supervisory  activities 

5)  Operational  activities 

1 )  Interactions  with  other  organizations 

2)  Adaptation 

3)  Normative  integration 

4)  Cultural  issues 

Table  1.  Internal  and  External  Functions  of  an  Organization. 

Table  2  summarizes  the  second  taxonomy,  which  adopts  an 
“epigenetic”  approach,  concentrating  not  on  established  functions  (as  does 
the  first  taxonomy),  but  rather  on  the  new  functions  that  must  develop  for 
effective  growth;  it  addresses  the  growth  of  institutions  and  communities 
through  an  almost  biological  metaphor.  This  second  taxonomy  establishes 
four  categories,  each  of  which  identifies  each  category  with  an  analytical 
task  in  the  study  of  organizational  survival  and  development.  Setting  these 
two  taxonomies  in  relation  to  one  another  problematizes  the  relevant 
theoretical  questions,  establishes  the  necessary  constraints  on  the  research 
design,  and  suggests  how  to  code  information  concerning  the  development 
of  these  organizations  x 


Washington  Academy  of  Sciences 


13 


ERKLARUNG 

(“Explanation,”  its  “variables,”  and  t 

heir  categories) 

VERSTEHEN 

(“Understanding”) 

,f  Variables  ” 

(in  “Explan¬ 
ation”) 

Independent: 

Evolution¬ 
ary  level 

Intervening: 

Locus  of  power 

Dependent: 

Performance 

[Interpretation:] 

Sequence  of 
integration 

Categories  of 
variables 

Stage  of: 

-  Initiation 

-  Takeoff 

-  Spillover 

Degree  of: 

-  Elitism8 

-  Internalization3 

-  Responsiveness  to 
demands  and 
feedback 

Robustness  of:b 

-  Information 
and  commu¬ 
nication 

-  Motivation 

-  Spheres  of 
competence 

Nature  of:c 

-  Merging  units 

-  Emerging  unit 
-Functional 

statements  vs. 

“real  sequences” 

Table  2.  Elements  for  an  Epigenetic  Analysis  of  General  Organizational  Evolution. 

“  The  present  short  article  does  not  address  these  categories;  however,  operationalizations  may  be  found  in 
Hayward  R.  Alker,  “On  Political  Capabilities  in  a  Schedule  Sense:  Measuring  Power,  Integration,  and 
Development,”  pp  307-73  in  Mathematical  Approaches  to  Politics,  ed.  by  H  R.  Alker,  K.W.  Deutsch.  and 
A.H.  Stoetzel  (Amsterdam:  Elsevier,  1973). 

b  See  the  reformation  of  Etzionf  s  categories  as  explicated  in  the  text. 

c  This  column  of  the  Table  is  more  directly  pertinent  to  Etzioni's  original  concern  with  international 

communities,  than  to  state  foreign  policy  making  per  se.  therefore,  it  is  included  for  completeness  but  not 
discussed  in  the  present  article. 


The  first  taxonomy  emphasizes  the  creation  of  organizational 
structures  to  fulfill  and  accomplish  prescribed  “internal”  and  “external” 
functions  while  the  second,  adopting  an  epigenetic  approach  and 
concentrating  on  new  behaviors  called  forth  by  the  environment, 
comprises  four  principal  categories:  stages  of  development,  locus  of 
power,  performance,  and  sequences  of  integration.  The  innovation  and 
incorporation  of  procedures  for  accomplishing  “internal  functions” 
represent  a  response  to  developmental  challenges  in  the  life  of  the 
institution.  Organizational  success  in  adapting  to  these  tasks  therefore 
represents  a  passage  from  one  phase  to  another.  Organizations  must  as  a 
rule  first  evolve  internal  functions  permitting  them  to  exist  stably  in 
relation  with  their  constituent  parts.  Only  then,  according  to  this  idealized 
functional  sequence,  may  they  engage  pro-actively  with  the  external 
environment.  Therefore  the  development  of  internal  and  external 
functions,  posited  by  the  functional  taxonomy,  may  be  heuristically  treated 
as  a  teleology  of  potentials  for  the  evolution  of  any  given  organization. 

The  functional  framework  is  thus  first-order  cybernetic,  pertaining 
to  the  cybernetics  of  observed  systems;  the  epigenetic  is  second-order 
cybernetic,  pertaining  to  the  cybernetics  of  observing  systems.M  Their 


Winter  2006 


14 


synthesis,  denoted  the  “paradox  of  intentional  emergent  coherence,”  lays 
stress  on  autopoiesis  and  learning.  In  this  presentation,  the  two  taxonomic 
frameworks  are  intermediated  by  necessary  constructs  that  may  properly 
be  called  one-and-a-half-order-cybernetic.  These  constructs  are  akin  to  the 
“middle  voice”  of  verbs  in  classical  Greek,  where  the  subject  acts  on  or  for 
itself,  halfway  between  (passive)  first-order-cybernetic  and  the  (active) 
second-order-cybernetic  frameworks.  Indeed,  insofar  as  constructivism  in 
the  social  sciences  dissolves  the  distinction  between  norms  and  structures, 
treating  them  equally  as  merely  different  ways  of  regarding  institutions,  it 
represents  a  bridge  to  autopoiesis  as  the  mediating  term  between 
coherence  and  emergence.  This  is  the  process  that  fundamentally 
characterizes  the  Paradox  of  Intentional  Emergent  Coherence.™ 

David  Easton’s  application  of  systems  theory  to  the  study  of 
politics  distinguished  among  the  elite,  regime,  and  community  sectors  of 
the  political  system.™1  What  the  elite  is,  is  self-evident.™  The  “regime 
sector”  comprises  those  institutions  of  the  political  system  through  which 
governance  is  executed.  The  community  sector  is  basically  everything 
else.  David  Apter  explicitly  reintroduced  the  notion  of  qualitative 
communication  among  these  sectors.  He  drew  particular  attention  to  the 
flow  of  “information”  from  the  community  to  the  regime  and  from  the 
regime  to  the  elite,  and  of  “coercion”  in  the  reverse  direction.  His  mature 
theoretical  work  is  inspired  more  by  Deutsch’s  than  by  Almond’s 
adaptation  of  Easton’s  systems  approach.™  Karl  Deutsch  replaced 
Easton’s  “less  precise  concepts  of  demands  and  supports”  with  “the 
concept  of  message  units  or  informative  bits.”™1  More  important,  he 
introduced  the  crucial  distinction  between  primary  and  secondary 
feedback  in  the  attempt  to  grasp  what  “consciousness”  and  “learning” 
might  mean  where  political  systems  were  concerned.  His  overall  goal  was 
to  explicate  such  philosophic  categories  as  “choice,”  “will,”  and 
“autonomy”  in  information  processing  terms.  Following  Deutsch’s 
definition  of  “information”  as  a  “patterned  relationship  between  events,” 
Apter  retuned  to  the  emphasis  placed  by  Max  Weber’s  American  student 
Talcott  Parsons  on  action ,  “a  more  narrow  term  that  includes  choice  and 
will,”  as  opposed  to  behavior ,  which  “may  include  the  mechanistic 
response  characteristic  of  lower  animals.”™11 

Work  by  Karl  Deutsch  helps  to  render  Etzioni’s  epigenetic 
framework  operationally  second-order  cybernetic.  According  to  Deutsch, 
foreign  policy  learning  may  be  cognitively  manifested  either  through  the 
transformation  of  goals  held  at  the  outset  into  goals  not  previously 
conceived,  or  through  the  choice  of  pre-existing  alternative  goals  over 


Washington  Academy  of  Sciences 


15 


other  goals  originally  held.  His  project  in  his  classic  Nerves  of 
Government  was,  among  other  things,  to  translate  the  categories  of 
classical  European  philosophy  (before  the  latter  was  depsychologized  by 
Bertrand  Russell  and  his  continuators)  into  cybernetic  language™11 
However,  the  limitations  of  the  conceptual  apparatus  available  to  social 
science  at  the  time  he  wrote  leave  him  often  a  prisoner  of  a  functionalist 
and  homeostatic  framework.  In  the  later  sections  of  the  book  where  he 
reintroduces  the  category  of  will  and  other  volitional  concepts,  he  points 
explicitly  towards  the  key  second-order  cybernetic  principle  of 
autopoiesis. 

Indeed,  Deutsch’s  translation  of  “faith”  and  “grace”  into  cybernetic 
language  are  remarkable  attempts  to  overcome  the  limitations  of  first- 
order  cybernetics.  Deutsch’s  cybernetic  treatment  of  categories  of  classical 
European  philosophy  goes  some  distance,  though  not  all  the  way,  towards 
infiltrating  an  autopoietic  (second-order  cybernetic)  aspect  into  the  first- 
order  cybernetic  framework  that  he  inherited  from  the  structural- 
functionalist  application  of  general  systems  theory  within  political  science. 
In  this  way  Deutsch  partly  infiltrated  an  autopoietic  aspect  into  the  first- 
order  cybernetic  framework  that  he  inherited  from  general  systems  theory. 

Epigenesis  and  Organizational  Autopoiesis 

Etzioni’s  epigenetic  “performance”  categories  were 

(1)  communication,  (2)  information,  and  (3)  control.  However,  these 
categories  are  still  somewhat  limited  by  the  dynamics  of  first-order 
cybernetics.  It  is  therefore  necessary  somewhat  to  reconceptualize  them. 
One  additional  well-known  mainstream  political-science  work,  the 
seminal  work  on  the  foreign-policy  decision-making  approach  from  the 
early  1960s,  completes  the  integration  of  Etzioni  and  Deutsch  into  a  fully 
second-order  cybernetic  outlook.  X1X  It  not  only  helps  to  correct  Etzioni’s 
taxonomic  triad  of  performance  variables,  but  also  supplies  a  framework 
in  which  certain  categories  of  Deutsch  specify  and  operationalize  that 
performance  triad  of  Etzioni.  That  work  enumerated  three  “clusters  of 
variables”  (or  sets  of  phenomena),  each  of  which  addresses  some  facet  of 
how  people  making  decisions  in  organizations  operate. 

These  clusters  are  (1)  communication  and  information, 

(2)  motivation,  and  (3)  spheres  of  competence.  Together  they  motivate  the 
re-specifications  of  Etzioni’s  performance  categories.  The  relation  to 
Etzioni’s  triad  of  performance  variables  is  as  follows.  The  communication 
and  information  cluster  is  about  communication  in  organizations,  and  also 


Winter  2006 


16 


about  the  organization  of  communications;  it  subsumes  two  of  Etzioni’s 
original  categories.  The  motivation  cluster  is  about  goals  in  organizations 
and  is  entirely  absent  in  Etzioni.  Deutsch  uses  it  along  with  other 
volitional  concepts.  The  cybernetic  term  for  autonomy  of  motive  is 
autopoiesis ,  and  this  concept  is  the  lever  with  which  to  open  the  overall 
problematique  to  second-order  cybernetic  considerations.  The  spheres  of 
competence  cluster  is  about  authority  in  organizations;  it  is  a  less 
mechanistic,  more  second-order  cybernetic  expression  of  Etzioni’ s 
category  of  “  control.” 

The  three  “stages  of  development”  inherent  in  the  epigenetic 
framework  (initiation,  takeoff,  and  spillover)  may  be  considered  as  not 
“stages”  per  se  fully  describing  a  continuum  of  development,  but  rather 
phases  in  the  transition  of  an  organization  from  one  category  to  another  in 
this  typology.  With  this  expanded  theoretical  content,  it  produces  a 
hierarchy  for  classifying  organizational  development  (see  Table  3)  that 
serves  also  as  a  template  for  mapping  foreign  policy  decisions.xx 

Still  closer  inspection  of  the  synthesis  of  the  functional  and 
epigenetic  taxonomies  discovers  the  influences  upon  an  organization’s 
achievement  or  failure  to  move  from  one  rung  of  the  developmental  ladder 
to  another.  These  influences  are  expressed  in  particular  by  the  setting  into 
mutual  relation  of  the  “locus  of  power”  and  “performance”  categories  of 
the  epigenetic  framework  on  the  one  hand  with,  on  the  other  hand,  the 
organization’s  “external  functions”  as  specified  in  the  functionalist 
framework.  External  functions  have  two  aspects:  normative  and 
behavioral.  The  key  to  examining  performance  is  an  assessment  of  the 
“behavioral  aspects.”  The  functionalist  framework  defines  this  as  the 
organization’s  adaptation  and  its  interactions  with  other  organizations. 
(“Organizations”  may  be  construed  in  the  sociological  sense  to  include 
non-bureaucratic  structures  as  well  as  structures  in  the  environment.) 
Adaptation  occurs  (or  fails  to  occur)  with  respect  to  the  organization’s 
institutional  environment,  and  in  the  second  instance,  with  respect  to  the 
demands  (both  internal  and  external)  on  the  nascent  organization. 


Washington  Academy  of  Sciences 


17 


(Implem  en  tat  ion) 

Phase  3 

Competence 

Operational  activities 

Spillover 

( Preparation  for 
implementation) 

Phase  2 

Motivation 

R  u  1  e-  s  upend  s  ory 
activities 

Takeoff  (2nd  moment) 

Rule-creating  activities 

Takeoff  (1st  moment) 

( Processing  of 
information ) 

Phase  1 

. 

.  . 

Communication 

Normative  activities 

Initiation  (2nd  moment) 

and  information 

Informational  activities 

Initiation  (1st  moment) 

(Collection  of 
information) 

Phase  0 

(Phase  of  for¬ 
eign  policy  deci¬ 
sion  making ) 

Functional  requisite  of 
organizational  devel¬ 
opment  (see  Table  1) 

Evolutionary 
phase  of 
organization 

Epigenetic  moment  in 
organizational  devel¬ 
opment  (see  Table  2) 

Table  3.  Concordance  between  Functional  Requisites  and  Epigenetic  Moments  in 
Organizational  Development,  and  also  by  extension  to  Phase  of  State 
Foreign  Policy’  Decision  Making.  (Read  this  Table  from  bottom  to  top.) 


There  is  a  one-to-one  correspondence  between  the  analytical 
subcategories  of  these  interactions  with  other  organizations,  under  the 
functionalist  framework,  and  the  criteria  of  performance  under  the 
epigenetic  framework.  To  be  exact,  there  are  under  the  functionalist 
framework  three  elements  of  “interactions  with  other  organizations” 
(threat  systems,  hierarchies,  and  goal  definition  and  realization),  and  under 
the  epigenetic  framework  there  are  likewise  three  performance  criteria. 
Table  4  sets  these  into  a  one-to-one  correspondence,  establishes  the 
synthesis  and,  through  the  Deutsch-Etzioni  transformation,  it 
operationalizes  the  first-order-cybernetic  categories  necessarily  in  a 
second-order-cybernetic  manner 


Winter  2006 


18 


EPIGENETIC  “LOCUS  OF  POWER  ” 

[  Response  to  demands/feedback  =  Performance  ] 

Performance 
variables  operational¬ 
ized  as  “capacities  ” 

Hierarchies 

Goal  definition  and 
realization 

Threat  systems 

Primacy  capacity 

Secondary  capacity' 

1)  Fundamental 
restructuring 

2)  Inner  rearrange¬ 
ment 

1 )  Steering  capacity 

2)  Depth  of  memory 

1 )  Power 

2)  Intake  channels 

Functional  behavior 
characteristic  of  the 
performance  variable 

Communication  and 
information 

Motivation 

Spheres  of 
competence 

[  Interactions  with  other  organizations  ] 

FUNCTIONAL  “EXTERNAL  BEHAVIOR” 

Table  4.  How  Deutschian  “Capacities”  Operationalize  Performance  V  ariables  from  the 

Functional  and  Epigenetic  Taxonomies,  Transforming  the  First-Order  Cybernetic 
into  the  Second-Order-Cybernetic. 


The  introduction  these  second-order  cybernetic  correctives  to 
Etzioni’s  original  epigenetic  taxonomy  of  performance  variables 
transforms  the  functionalist  framework,  which  treats  organizations 
homeostatically,  into  an  epigenetic  cycle  of  organic  development. 
Deutsch’s  remarks  on  obstacles  to  learning  in  cybernetic  systems  are 
directly  to  this  point.  Specifically,  he  discusses  “losses,”  any  of  which  can 
prevent  effective  learning.  Since  the  inverse  of  such  a  loss  is  a  capacity,  he 
in  fact  enumerates  six  capacities  that  promote  learning.  It  turns  out  that  a 
different  pair  of  these  six  capacities  is  related  to  each  of  the  functionally 
defined  external  behaviors  (threat  systems,  hierarchies,  and  goal  definition 
and  realization).  Thus  the  three  pairs  of  capacities  are  mutually  exclusive; 
and  collectively,  they  exhaust  the  set  of  six,  as  follows: 


Washington  Academy  of  Sciences 


19 


1.  Hierarchies  are  inhibited  primarily  by  ‘loss  of  capacity  for 

fundamental  restructuring”  and  secondarily  by  “loss  of  depth  of 
memory  ”  Cognitive  hierarchies  are  not  internalized  and 

organizational  hierarchies  are  not  imprinted  in  the  absence  of 
information  and  communication.  These  requisite  capacities 
therefore  depend  most  closely  upon  the 

information  communication  performance  variable,  which  is  in 
turn  most  characteristic  of  the  initiation  phase  of  epigenetic 
development. 

2.  Goal  definition  and  realization  are  inhibited  primarily  by  “loss 
of  steering  capacity”  and  secondarily  by  “loss  of  capacity  for 
inner  rearrangement.”  Goals  are  not  defined  or  realized  in  the 
absence  of  motivation.  The  requisite  capacities  therefore 
depend  most  closely  upon  the  motivation  performance  variable, 
which  is  in  turn  most  characteristic  of  the  takeoff  phase  of 
epigenetic  development. 

3.  Threat  systems  are  inhibited  by  “loss  of  power”  and 
secondarily  by  “loss  of  intake  channels.”  Threat  systems  are  at 
best  ineffective,  and  at  worst  nonexistent,  in  the  absence  of 
competence.  The  requisite  capacities  depend  most  closely  upon 
the  competence  performance  variable,  which  is  in  turn  most 
characteristic  of  the  spillover  phase  of  epigenetic  development. 


Theoretical  structures  of  second-order  cybernetics,  outlined  above, 
establish  links  between  members  of  this  conceptual  triplet  and  the  one 
immediately  preceding.  However,  it  would  be  inconsiderate  to  adduce 
supernumerary  abstractions  to  such  a  demonstration,  particularly  in  the 
presence  of  editorial  limits  on  the  length  of  this  article.  Therefore  Table  5 
summarizes  those  connections  in  apothegms  unifying  them  with  the 
fundamental  analytical  issues  in  the  complex  sciences,  enumerated  at  the 
outset  of  the  chapter. 


Winter  2006 


20 


APPOSITE  APOTHEGM 

(Second-order  cybernetic)  perform¬ 
ance  criterion  and  corresponding  phase 
of  epigenetic  evolution 

Complex' -scientific  category  corresponding 
to  tire  functionally  defined  behavior 

(Consolidated  explication) 

Epigenetic  phase 

Performance 

criterion 

Manifesting  complex- 
scientific  behavior 

j  Functional 
j  behavior 

During  the  phase  of  ini¬ 
tiation ,  information  and 
communications  emerge 
as  [and  manifest  as  cog¬ 
nitive  and  organizational] 
hierarchies. 

Initiation 

Information  and 
communications 

Emergence 

!  Hierarchies 

During  the  phase  of  take¬ 
off \  motivation  changes 
and  stabilizes  [and  mani¬ 
fests  as  and  sustains]  goal 
definition  and  realization. 

Takeoff 

Motivation 

Autopoiesis  (for 
sustainability) 

!  Goal  definition 
and  realization 

During  the  phase  of  spill¬ 
over,  competence  self- 
organizes  [and  manifests 
and  coheres  as)  threat 
systems. 

Spillover 

Competence 

(Self-organized) 

Coherence 

!  Threat  systems 
;  (for  response 

I  to  threat) 

Table  5.  Concordance  between  Categories  for  Assessing  Institutional  Development  of  International 
Organizations  and  Foreign -Policy  Decision  Making  on  the  One  Hand,  and,  on  the  Other 
Hand,  Fundamental  Issues  in  the  Complex  Sciences  Enumerated  at  the  Outset  of  This 
Article, 


The  Paradox  of  Intentional  Emergent  Coherence 

The  paradox  of  intentional  emergent  coherence  is  a  condensation 
and  a  transformation  of  this  first  order  cybernetic  system  into  a  second 
order  cybernetic  system.  It  is  explicitly  second-order-cybernetic  and 
founded  in  the  theory  of  complex  systems.  Consequently  it  operationalizes 
autopoiesis  in  particular  better  than  any  recitation  of  functional 
mechanisms.  For  an  organization,  autopoiesis  is  the  crucial  attribute 
marking  the  successful  performance  of  functional  tasks  associated  with  the 
developmental  stage  of  take-off.  It  signifies  the  capacity  proactively  to 
undertake  relations  with  other  organizations,  as  opposed  to  remaining  only 
a  coordinating  center  for  actions  of  its  own  component  organizational 
elements.  It  is  the  foundation  of  autonomous  motive. 


Washington  Academy  of  Sciences 


21 


Figure  1.  A  Functionalist  General-systems  Inventory  of  Influences  on  State  Foreign  Policy  Behavior. 


Figure  2.  The  Paradox  of  Intentional  Emergent  Coherence:  A  Cyclical  Developmental  Framework 

for  Organizations  and  Social  Systems,  Unifying  the  Functional  and  Epigenetic  Approaches. 


Figure  1  portrays  a  standard  functionalist  general-systems 
flowchart  of  influences  on  the  foreign-policy  behavior  of  a  state. 
Inspection  will  show  that  all  “variables”  (boxes  with  Roman  numerals) 
and  “causes”  (labeled  arrows)  in  Figure  1  appear  also  in  Figure  2,  albeit 
some  in  abbreviated  form.  [I],  [II],  and  [III]  represent  the  demands  upon 
the  political  system  and  supports  to  it  and  their  reciprocal  intermediation 
by  (A-l),  (A-2),  and  (B-l).  Those  three  flows,  together  with  (B-2)  are  seen 
in  Figure  2  to  represent  the  flow  from  [I]  to  [III],  interpreted  as  transition 
from  the  phase  of  Emergence  to  that  of  Coherence,  in  turn  represented  in 
Table  3  as  progression  from  Phase  0  to  Phase  1.  The  movement  from  [III] 
to  [V]  in  Figure  1  is  the  beginning  of  processing  of  information  on  the  new 


i 


Winter  2006 


22 


situation  after  it  has  been  collected;  “Conversion”  is  the  structural- 
functional  category  identifying  this  process,  and  it  is  there  intermediated 
by  the  flows  (C),  (D),  and  (E).  In  Figure  2,  this  represents  the  transition 
from  Coherence  to  Crisis,  as  the  new  challenge  begins  coming  to  a  head; 
and  in  Table  3,  it  is  the  progression  from  Phase  1  to  Phase  2,  where  the 
crucial  issue  of  motivation  comes  to  the  fore,  especially  along  with 
associated  second-order-cybernetic  concerns  with  autopoiesis  and  goal 
definition. 

The  critical  elaboration  of  a  response  (“Decision  and 
Implementation”)  is  portrayed  schematically  in  Figure  1  by  the  move  from 
[V]  to  [VI]  via  (F),  interpreted  in  Figure  2  as  the  resolution  of  the  Crisis  by 
Performance  (either  good  or  bad,  and  implemented  either  poorly  or 
efficiently),  and  captured  in  Table  3  as  the  movement  from  Phase  2  to 
Phase  3.  Finally,  in  Figure  1  for  the  case  of  state  foreign-policy  decision¬ 
making,  there  is  feedback  to  the  international  and  domestic  political 
environments,  i.e.  from  [VI]  to  [I]  and  [II],  via  the  flows  (G-l)  and  (G-2). 
These  are  also  represented  in  Figure  2;  for  Table  3,  it  is  the  “relapse”  from 
Phase  3  to  Phase  0,  awaiting  a  new  situation  of  challenge  to  arise. 

The  progressive  transformation  of  Figure  1  into  Figure  2 
demonstrates  that  the  functionalist  schema  hides  an  organic  cycle  of 
epigenetic  development.  Indeed,  forsaking  the  functionalist  for  the 
epigenetic  standpoint  in  fact  renders  the  schema  more  parsimonious  with 
no  sacrifice  of  analytical  rigor;  the  analytical  rigor  is  enhanced,  as  two 
conceptual  consolidations  transform  the  functionalist,  first-order- 
cybemetic  schema  in  Figure  1,  so  as  to  reveal  its  epigenetic,  second-order- 
cybemetic  essence  in  Figure  2.  First:  The  epigenetic  approach  discovers 
that  the  endogenous  demands  and  supports  evolve  with  each  metamorphic 
stage  through  which  an  organization  passes;  therefore,  these  demands  and 
supports  may  be  treated  together  as  a  single  expression  of  the 
developmental  stage  that  it  has  reached.  Second:  Conversion,  decision, 
and  implementation  together  constitute  the  response  to  the  epigenetic 
challenge  at  hand;  therefore,  these  may  be  collapsed  to  single  category 
representing  the  organization’s  performance-response  to  outgrow  that 
developmental  stage  answer  the  self-transformative  challenge  to  enter  the 
next.™ 


Conclusion 

The  basis  for  the  two  correspondences  just  enumerated  is  that  the 
innovation  and  incorporation  of  procedures  for  the  various  kinds  of 


Washington  Academy  of  Sciences 


23 


activities  specified  under  “internal  functions,”  represent  a  response  to  a 
developmental  challenge.  Successful  adaptation  therefore  represents  the 
passage  from  one  metamorphic  stage  to  another,  in  the  epigenetic  course 
that  inheres  in  the  organization,  whether  it  is  a  state  or  an  international 
organization  or  any  other  social  system.  The  assertion  that  initiation, 
takeoff,  and  spillover  are  not  point  discontinuities  between  different 
phases  but  rather  are  themselves  transition  phases  in  the  life  of  the 
organization  is  validated  by  their  specification  as  being  composed  of 
developmental  functional  tasks.  The  transformation  of  Figure  1  into  Figure 
2  discovers  the  ladder  of  epigenetic  development  concealed  by  the 
functionalist  cycle  and  embedded  in  it.  This  ladder  indeed  conditions  the 
very  parameters  of  that  functional  cycle. 

The  transformation  of  the  functionalist  into  the  epigenetic 
framework  demonstrates  the  intrinsic  unity  of  the  two  approaches.  It  maps 
unambiguously  the  “internal  functions”  and  the  “stages  of  epigenetic 
development”  onto  a  common  domain.  The  complete  ladder  of 
institutional  development  is  a  hierarchy  representing  a  typology  of  the 
nature  of  the  emerging  units  (a  category  under  the  “sequences  of 
integration”  category  of  the  epigenetic  framework  in  Table  2).  This 
correspondence  integrates  the  principal  category  of  the  epigenetic 
framework  with  the  principal  taxonomy  of  the  functional  framework.  The 
Paradox  of  Intentional  Emergent  Coherence  unfolds  as  a  result.  Thus 
Table  5  sustains  and  illustrates,  as  was  asserted  in  the  beginning,  that  in 
the  complex-scientific  approach,  autopoiesis  intermediates  emergence 
with  coherence;  and  from  that,  the  rest  of  this  theoretical  construct  is 
generated  and  supported. 


REFERENCES 


i.  Senior  Research  Fellow.  Institute  of  European  and  Russian  Studies.  Carleton 
University.  Postal  address:  Station  H.  Box  518.  Montreal.  Quebec  H3G  2L5.  Canada. 
Email  address  <rmcf§.alum.mit.edu>;  website  <http://www.robertcutler.org>.  Member 
of  the  Washington  Evolutionary  Systems  Society.  A  draft  of  this  article  was  presented  to 
the  Washington  Academy  Conference  CapSci2006.  The  author's  first  oral  presentation 
of  these  ideas  was  to  two  interdisciplinary  conferences  sponsored  by  the  New  England 
Complex  Systems  Institute  in  Boston  in  October  1998  and  March  1999. 


Winter  2006 


24 


ii.  Jean  Piaget.  Le  structuralisme  (Paris:  Press  universitaires  de  France.  1968);  for  a 
summary,  see  Jean  Piaget.  Epistemologie  des  sciences  de  I  'homme  (Paris:  Gallimard. 
1970).  pp.  278-86. 

iii.  Peter  Winch.  The  Idea  of  a  Social  Science  and  Its  Relation  to  Philosophy  (London: 
Routledge  and  Kegan  Paul.  1958);  2nd  ed.  (London:  Routledge.  1990).  For  a 
commentary  comparing  the  arguments  in  the  two  editions,  see  Philip  Pettit  "Winch's 
Double-edged  Idea  of  a  Social  Science."  History  of  the  Human  Sciences  13.  no.  1 
(February  2000):  63-77. 

iv.  Totality  is  thus  as  like  a  mathematical  Object:  self-regulation,  an  Operation;  and 
transformation,  a  Relation.  See  Arthur  F.  Bentley.  "Sociology  and  Mathematics"  [first 
published  in  1931].  pp.  53-100.  in  Bentley.  Inquiry  into  Inquiries:  Essays  in  Social 
Theorw  ed.  with  Introd.  bv  Sidney  Ratner  (Boston.  Mass.:  Beacon  Press.  1954).  at  56- 
59. 

v.  Paul  Cilliers.  Complexity’  and  Postmodernism:  Understanding  Complex  Systems 
(London:  Routledge.  1998). 

vi.  Autopoiesis  is  the  capacity’  of  complex  systems,  and  especially  complex  adaptive 
systems,  to  set  their  own  goals  through  progressive  interaction  with  their  environment 
and  through  learning  in  response  to  this.  John  Holland.  Hidden  Order:  How  Adaptation 
Builds  Complexity  (New  York:  Perseus  Books.  1996);  Niklas  Lulnnan  Soziale  Svsteme: 
Grundriss  einer  allgemeine  Theorie  (Frankfurt:  Suhrkamp.  1984).  translated  as  Niklas 
Luhmann.  Social  Systems  (Writing  Science ).  trans.  John  Bednarz  and  Dirk  Baecker 
(Stanford.  Calif.:  Stanford  University’  Press.  1995). 

vii.  Based  on  [Yaneer  Bar-Yam].  "NECSI  Guide:  About  Complex  Systems,” 
<http:/Avyvyv. necsi.org/guide/study .html>.  accessed  10  October  2006. 

viii.  Respectively:  The  United  Nations  System:  International  Bibliography,  ed.  by  Klaus 
Hiifner  and  Jens  Naumann  (Munich:  Verlag  Dokumentation.  1976-present);  and  Amitai 
Etzioni.  "The  Epigenesis  of  Political  Communities  at  the  International  Level,”  American 
Journal  of  Sociology.  68.  no.  4  (December  1963):  407-21.  reprinted  at  pp.  346-58  in 
International  Politics  and  Foreign  Policy,  ed.  by  James  N.  Rosenau.  rev.  ed..  (New  York: 
Free  Press.  1969). 

The  Hufner-Nauman  taxonomy  is  based  in  the  "structural-functionalist"  school 
of  political  analysis,  which  emphasizes  "capabilities"  of  a  different  nature  than  discussed 
here,  and  of  which  the  locus  c/assicus  is  Gabriel  A.  Almond  and  G.  Bingham  Poyvell.  Jr., 
Comparative  Politics:  A  De\’elopmental  Approach  (Boston:  Little.  Broyvn.  1966). 
drayving  heavily  but  implicitly  on  the  magnum  opus  of  Almond's  teacher  Talcott  Parsons. 
The  Social  System  (Neyv  York:  Free  Press.  1951).  which  was  unimaginably  influential  in 
its  time.  Parsons  in  turn  was  the  chief  American  exegete  of  the  great  German  sociologist 
Max  Weber,  yvhose  attention  to  historical  detail  he  however  sacrificed  for  the  gain  of 
abstract  conceptualization  at  the  grandest  level  of  theory  .  The  resulting  deficiencies  (and 


Washington  Academy  of  Sciences 


25 


they  continue  still  today  to  affect  organizational  and  foreign-policy  analysis  by  many 
mainstream  North  American  political  scientists)  are  trenchantly  laid  out  by  the  pioneering 
American  sociologist  William  Foote  Whyte.  “Parsonian  Theory  Applied  to 
Organizations,’'  pp.  250-267  in  The  Social  Theories  of  Talcott  Parsons:  A  Critical  View, 
ed.  by  Max  Black  (Englewood  Cliffs,  N.J.:  Prentice-Hall.  1961). 

The  issues  that  Etzioni  outlines  are  not  far  from  a  the  recent  attempt  in  Security 
Communities .  ed.  by  Emanuel  Adler  and  Michael  Barnett  (Cambridge:  Cambridge 
University  Press.  1998),  to  revive  the  approach  by  Karl  W.  Deutsch  et  al. ,  Political 
Community  and  the  North  Atlantic  Area  (Princeton.  N.J.:  Princeton  University  Press. 
1957).  to  the  study  of  security  communities.  However.  Etzioni' s  framework  better 
houses  recent  advances  in  social  network  analysis,  including  the  important  qualitative 
differences  now  rigorously  demonstrated  between  triads  and  the  dyadic  relationships 
emphasized  by  methodological-individualist  approaches:  see.  e.g ..  Ronald  L.  Breiger. 
Explorations  in  Structural  Analysis:  Dual  and  Multiple  Networks  of  Social  Structure 
(New  York:  Garland  Press.  1991).  and  Stanley  Wassennan  and  Katherine  Faust.  Social 
Network  Analysis  (Cambridge:  Cambridge  University  Press.  1994):  compare  Bam 
Buzan  and  Ole  Waever.  Regions  and  Powers:  The  Structure  of  International  Security. 
(Cambridge:  Cambridge  University  Press.  2003).  At  the  same  time,  social  network 
analysis  conserves  the  systems-theory  approach  underlying  Deutsch' s  perspective  and 
easily  accommodates  the  cybernetic  aspects  of  complex  systems,  as  explained  among 
others  by  Paul  A.  Stokes.  “Socio-Cybemetics  and  the  Project  of  Scientificization  of 
Sociology."  pp.  3 1 1-334  in  Self-Steering  and  Cognition  in  Complex  Systems:  Towards  a 
New  Cybernetics .  ed.  by  Francis  Heylighen.  Eric  Rosseel.  and  Frank  Demeyene  (New 
York:  Gordon  and  Breach.  1990). 

ix.  For  a  more  elaborate  argument  of  some  of  these  points,  see  Cutler,  ‘The  Emergence 
of  International  Parliamentary  Institutions"  (fn  2). 

x.  For  an  example  of  how  this  synthesis  generates  a  philosophically  grounded  and 
empirically  applicable  coding  methodology  for  organizational  development,  see  Robert 
Cutler  and  Alexander  von  Lingen.  “The  European  Parliament  and  European  Security  and 
Defence  Policy,"  European  Security  12.  no.  2  (June  2003):  1-20.  also  at 

<http  ://www .  robertcutler.org/ar0  3  es .  htm> . 

xi.  As  Felix  Gever  notes  in  “The  Challenge  of  Sociocybemetics,”  Kvbernetes ,  24.  no.  4 
(1995):  5-32.  another  main  difference  as  set  out  by  Heinz  von  Forster  was  that  “second- 
cybernetics  explicitly  includes  the  observer(s)  in  the  systems  to  be  studied  [and]  generally 
deals  with  living  systems." 

xii.  Compare  Alicia  Juarrero.  Dynamics  in  Action:  Intentional  Behavior  as  a  Complex 
System  (Cambridge:  MIT  Press.  1999).  pp.  109-25  passim. 

xiii.  David  Easton.  The  Political  System  (New  York:  Knopf.  1953).  and  several 
subsequent  monographs  on  the  same  time. 


26 


xiv.  Actually,  elite  studies  have  long  been  an  identified  topic  within  the  comparative 
politics  subdiscipline  of  political  science,  complete  with  its  own  problematization  and 
definition  of  different  elites.  What  an  elite  is.  is  nevertheless  fairly  clear  in  an  ordinary- 
language  way.  w  hereas  that  is  not  necessarily  the  case  for  the  concept  of '  regime.” 

xv.  David  E.  Apter.  Choice  and  the  Politics  of  Allocation  (New  Haven.  Conn.:  Yale 
University  Press.  1971). 

xvi.  Law  rence  C.  Mayer.  Comparative  Political  Inquin'  (Homew  ood.  Ill.:  Dorsev. 

1972) .  p.  136. 

xvii.  Ibid.,  p.  127. 

xviii.  Kari  W.  Deutsch.  The  Nerves  of  Government:  Models  of  Political  Communication 
and  Control  (New  York:  Free  Press  of  Glencoe.  1963).  pp.  96.  210.  222.  The 
transformation  of  goals  held  at  the  outset  into  goals  not  previously  conceived  is  rare  and 
cannot  be  programmed.  Haas  reserv  es  the  term  “learning*'  for  goal  transformation,  as 
distinct  from  “adaptation."  See  Ernst  B.  Haas.  “Collective  Learning:  Some  Theoretical 
Speculations."  in  Learning  in  U.S.  and  Soviet  Foreign  Policy \  pp.  62-99.  esp.  pp.  72-97 . 
Compare:  James  N.  Rosenau.  “Foreign  Policy  as  Adaptive  Behavior:  Some  Preliminary 
Notes  for  a  Theoretical  Model.*'  Comparative  Politics  2.  no.  3  (April  1970):  365-387. 
Rosenau.  The  Study  of  Political  Adaptation:  Essays  on  the  Analysis  of  World  Politics 
(London:  Frances  Pinter.  1981);  Steve  Smith.  Foreign  Policy’ Adaptation  (Famborough. 
Gower.  1981):  Smith.  “Rosenau* s  Adaptive  Behaviour  Approach."  Review  of 
International  Studies  7.  no.  2  (1981)  pp.  107-26. 

xix.  Richard  W.  Snyder.  H.W.  Bruck.  and  Burton  Sapin.  “Decision-making  as  an 
Approach  to  the  Study  of  International  Politics."  pp.  106-170  in  Foreign  Policy  Decision 
Making:  An  Approach  to  the  Study  of  International  Politics,  ed.  by  Snyder.  Bruck.  and 
Sapin  (New  York:  Free  Press  of  Glencoe.  1962). 

xx.  The  italicized  entries  in  the  alternate  cells  in  the  left-hand  column  of  Table  3  are 
taken  from  Robert  Axelrod.  “Schema  Theory:  An  Information  Processing  Model  of 
Perception  and  Cognition.”  American  Political  Science  Review  67.  no.  4  (December 

1973) :  1248-1266:  compare  Alessandro  Bruschi.  “Informazione  e  processi  decisionali 
nel  sistema  politica."  pp.  165-213  in  Ministero  degli  affari  esteri.  Istituto  diplomatico. 
Relazioni  internazaionali:  metodi  e  tecniche  di  analisi  (Florence:  Centro  Studi  e  ricerche 
di  politica  comparata.  1973);  also  the  periodization  of  decision-making  in  domestic 
affairs  by  Peter  H.  Solomon.  Soviet  Criminologists  and  Criminal  Policy:  Specialists  in 
Policy-Making  (New  York:  Columbia  University  Press.  1978).  p.  1 14.  Fig.  1. 

xxi.  The  successful  implementation  of  a  good  decision  strengthens  these  supports  and 
develops  new  capabilities.  A  poor  decision  w  eakens  supports,  and  a  poor  implementation 
(even  of  a  good  decision)  stunts  the  development  of  new  capabilities.  Successes  and 
failures  are  not  isolated  events  but  experiences  that  condition  the  whole  of  future 


Washington  Academy  of  Sciences 


27 


evolution.  For  a  discussion  of  this  notion  in  relation  to  the  development  of  domestic 
political  systems,  see  Leonard  Binder  et  a /.,  Crises  and  Sequences  in  Political 
Development  (Princeton.  N.  J. :  Princeton  University  Press.  1971);  and  Crises  of  Political 
Development  in  Europe  and  the  United  States,  ed.  by  Raymond  Grew  (Princeton.  N.J.; 
Princeton  University  Press.  1978). 


Winter  2006 


This  page  intentionally  left  blank 


Washington  Academy  of  Sciences 


NATURE  VERSUS  NURTURE  IN  EVOLUTION, 
AND  EMERGENCES  OF  DESIGNS  IN 
GENETICS,  THE  IMMUNE  SYSTEM  AND  THE  BRAIN 


29 


Roulette  William  Smith 

Institute  for  Postgraduate  Interdisciplinary  Studies 
Palo  Alto,  CA 


Abstract 

Even  though  much  remains  to  be  explicated,  it  generally  is  agreed  that 
DNA  provides  a  parsimonious  basis  for  evolution  associated  with 
speciation.  differentiation,  development,  immunity,  cognition  and 
behavior,  aging,  and  dying  and  death.  Until  published  reports  of  two 
human  genome  projects  (HGPs)  in  mid-Feb ruarv  2001.  it  had  not  been 
appreciated  that  proteomic  components  of  the  human  genome  account 
for  approximately  2%  of  the  human  genome  (-30.000  genes)  and  at 
most  25%  of  the  genome  comprise  the  proteome  (i.e..  the  protein¬ 
encoding  portion  of  the  genome)  and  its  regulatory  elements.  Based  on 
studies  of  “slow  viruses”  and  their  roles  in  “dementia"  in  both  the  brain 
and  the  immune  sy  stem.  Smith  hypothesized  that  DNA  must  be  the 
repository  of  long-term  memories  in  living  systems  (LTM)  -  and 
especially  in  the  brain  -  with  broad  implications  for  evolution  (1979). 
Contemporaneously.  Tonegawa  demonstrated  rearrangements  in  DNA 
account  for  immunoglobulin  specificity  (1978;  Sakano  et  al. .  1979). 
Taken  together,  these  findings  support  three  interdependent 
evolutionary7  schemas  in  humans  and  other  higher  animals  with  bony 
crania  (Smith.  2006a;  Smith.  2006b).  One  evolutionary  scheme  is 
associated  with  Daryvinian  proteomic  (i.e.,  genetic  and  epigenetic) 
evolution.  A  second  interdependent  evolutionary  pathway  is  associated 
with  7/7  utero  transmission  of  immunoglobulins  and  other  evolutionary 
information  (Vemy  and  Kelly,  1981/1983).  and  breastfeeding  - 
especially  in  humans.  The  third  pathway  is  associated  with  imitation  in 
behavior  associated  with  mirror  neurons  (Arbib  et  al.,  2000;  Rizzolatti 
and  Craighero.  2004;  Iacoboni  et  al.,  2005)  and  other  mechanisms 
involving  the  transmission  of  information  to  and  within  the  brain  -  the 
latter  generally  comprising  DNA  changes  in  non-proteomic  portions  of 
the  genome.  This  report  explores  emergences  in  spontaneous,  “natural" 
and  aberrant  designs  and  heuristics  among  the  three  interdependent 
evolutionary7  subsystems  and  their  associated  transmission  mechanisms. 
Emergent  designs  associated  with  proteomic  regions  of  the  genome 
largely  represent  designs  in  “nature,"  whereas  emergent  designs 
associated  with  the  remaining  interdependent  evolutionary7  schemas 
often  reflect  consequences  of  designs  associated  with  “nurturance." 


Winter  2006 


30 


Background  and  Introduction 

On  Design  -  ‘ Caveats  Emptor '  and  Limiting  the  Scope  of  this  Report 


The  central  theme  of  this  symposium  -  “emergences  in  designs”  - 
conjures  a  variety  of  meanings  and  opportunities.  This  report  focuses  on 
four  (of  eight)  definitions  of  design  taken  from  the  Merri am -Webster 
Online  Dictionary  (www.m-w.com).  These  definitions  are  “...  4)  a 
preliminary  sketch  or  outline  showing  the  main  features  of  something  to 
be  executed  :  <the  design  of  the  stadium>;  5  a)  an  underlying  scheme  that 
governs  functioning,  developing,  or  unfolding  :  PATTERN,  MOTIF  <the 
general  design  of  the  epic>;  5  b)  a  plan  or  protocol  for  carrying  out  or 
accomplishing  something  (as  a  scientific  experiment);  also  :  the  process  of 
preparing  this;  6)  the  arrangement  of  elements  or  details  in  a  product  or 
work  of  art;  and,  7)  a  decorative  pattern  ...”  Because  of  space  limitations, 
this  report  focuses  solely  on  those  definitions  of  design  which  are  nouns. 
A  goal  is  to  disambiguate  nature  and  nurture  in  evolution,  and  elucidate 
emergences  in  designs  associated  with  three  “emergent”  interdependent 
evolutionary  schemas  to  be  described  momentarily. 


On  Evolution  An  Introduction  to  a  Post-Darwinian  Model  of  Evolution 

Charles  Darwin’s  theories  of  evolution  and  speciation  (1859), 
while  widely  accepted  by  life  scientists  and  most  scholars,  have  been 
challenged  by  other  scholars  who  praise  Jean-Baptiste  Lamarck 
(1819/1984)  on  adaptation,  Pierre  Teilhard  de  Chardin  (1959)  on 
phenomenology  and  ‘inward’  reflection,  and  others  ( e.g .,  Huxley,  1941). 
Reported  findings  in  mid-February  2001  from  two  human  genome  projects 
(HGPs),  when  coupled  with  experimental  findings  and  speculations  about 
DNA  dynamics  (McClintock,  1950;  Tonegawa  et  al. ,  1978;  Sakano  et  al ., 
1979;  Smith,  1979;  Zou  and  Buck,  2006)  now  support  a  novel  post- 
Darwinian  tripartite  model  of  evolution. 

The  tripartite  model  derives  from  the  HGPs’  revelation  that, 
according  to  its  most  generous  interpretation,  at  most  25%  of  the  human 
genome  accounts  for  the  proteome,  with  as  much  as  75%  or  more  of  non- 
proteomic  regions  of  the  human  genome  remaining  to  be  explicated  (see 
February  2001  issues  of  Nature  [Volume  409,  15  February  2001]  and 
Science  [Volume  291  (5507),  16  February  2001]).  Non-proteomic  regions 
of  the  genome  sometimes  are  referred  to  as  “junk”  DNA. 


Washington  Academy  of  Sciences 


31 


Although  notions  of  proteomic  and  non-proteomic  regions  of  a 
genome  were  unknown  in  1979,  Smith  (1979)  anticipated  that  changes  in 
DNA  (i.e.,  LTM)  would  take  place  in  the  non-contiguous  and  non- 
proteomic  regions  in  the  brain  and  immune  system,  that  endogenous 
retrovirus-like  elements  may  play  roles  in  depositing  information  in  non- 
proteomic  regions  of  the  genome,  and  that  Francis  Crick’s  “central 
dogma”  was  grossly  deficient  because  of  overwhelming  evidence  of  an 
inverse  information  pathway  (cf.  Crick,  1958;  Crick  et  a /.,  1961;  Crick, 
1970;  Smith,  1979;  Smith  2006a;  Smith,  2006b).  The  subsequent  design 
and  invention  of  proXomc-ehcixomc-iomc-phoionic  molecular  calculating 
(preliophic  moleculating)  devices  and  processes  demonstrated  the 
potential  for  inverse  molecular  information  pathways  (Smith  and  Shadel, 
2003  patents  pending).  It  then  was  proposed  that  trinucleotide  repeat 
(TNR)  diseases  provide  further  support  for  inverse  molecular  information 
pathways  and  junctions  between  proteomic  and  non-proteomic  regions  of 
genomes  (Smith,  2003;  cf.  Cleary  and  Pearson,  2003). 

Clues  to  the  need  for  a  tripartite  model  of  evolution  derive  from 
DNA  rearrangements  associated  with  immunoglobulin  production  and 
DNA  changes  in  non-proteomic  regions  of  the  brain  -  neither  of  which 
are  transmitted  to  the  germ-line.  Previously,  interpretations  of  DNA 
rearrangements  associated  with  immunoglobulin  specificity  were 
interpreted  as  evidence  for  rejecting  the  “one  gene  -  one  protein”  dogma. 
Scholars  had  overlooked  that  those  gene  rearrangements  were  not  being 
passed  along  to  germ-line  tissue.  A  need  for  a  second  (non-Darwinian) 
interdependent  evolutionary  schema  was  evident  upon  reflection  on 
passive  immunity  associated  with  in  utero  transmission  of 
immunoglobulins  and  breastfeeding,  the  effects  of  addictive  drugs  on 
fetuses  during  pregnancy  and  the  newborn,  and  possible  transfers  of 
soulful  information  in  utero  (Verny  and  Kelly,  1981/1983).  A  need  for  a 
third  evolutionary  pathway  became  evident  because  changes  in  DNA  in 
the  brain  are  unlikely  to  be  accompanied  by  cell  division  (i.e.,  a  significant 
evolutionary  event  was  to  retain  the  ability  to  change  DNA  without  cell 
division,  especially  for  cells  constrained  by  a  bony  cranium),  because  of 
the  complexity  of  neural  networks  fostering  the  need  for  efficient 
intracellular  communication  through  axons  and  dendrites,  because  DNA 
changes  in  brains,  may  represent  a  priori  events  with  changes  in  axons- 
dendrites  representing  a  posteriori  consequences  of  those  DNA  changes, 
and  because  DNA  changes  in  the  brain  are  not  transmitted  to  the  germ¬ 
line.  In  short,  a  tripartite  system  of  evolution  became  essential  because  of 
separate  transmission  mechanisms  associated  with  genetic  reproduction 


Winter  2006 


32 


associated  with  the  germ-line,  passive  immune  transfers  in  utero ,  and 
DNA  changes  in  the  brain.  Moreover,  transmission  mechanisms  in  the 
brain  may  invoke  mirror  neuron  systems  (Arbib  el  cil,  2000;  Rizzolatti 
and  Craighero,  2004;  Iacoboni  el  a/.,  2005;  Blakeslee,  2006)  for  acquiring 
information  by  imitation  and  mimicry  (cf.  Ekman,  1973;  Ekman  and 
Friesen,  1975;  Ekman,  2003).  Evidence  of  “psychoviruses”  (Smith,  1987; 
Smith,  1988)  and  life-long  consequences  of  traumatic  events  also  support 
infectious  cognitive  snippets  leading  to  DNA  changes  (Smith,  2006a; 
Smith,  2006b).  Thus,  transmissions  of  evolutionary  information  associated 
with  changes  in  DNA  (in  humans)  may  involve:  a)  sexual  reproduction;  b) 
in  utero  transfers  and  breastfeeding;  and  c)  imitation,  mimicry,  and  trauma 
and  psychoviruses. 

Based  on  considerable  theoretical,  experimental  and  clinical 
evidence,  DNA  changes  in  the  brain  probably  involve  changes  from 
adenine*thymine-rich  regions  to  guanine*cytosine-richer  regions  in 
genomes  in  selected  neurons.  Rates  of  changes  should  differ,  say,  for  the 
forebrain  (i.e.,  associated  with  cognition)  and  cerebellum  ( i.e associated 
with  acquired  sensory-motor  responses).  One  crude  measure  of  nurturance 
is  the  ratio  of  guanine* cytosine  base-pairs  ::  adenine*thymine  base-pairs 
in  selected  regions  (Smith,  2003b;  hereafter  designated  G*C  ::  A*T).  We 
also  introduce  the  term  sytitropy  to  refer  to  mathematical,  chemical  and 
physical  representations  of  increased  ordering  and  organization  in 
information. 

Intriguing  consequences  of  the  tripartite  model  of  evolution  are  its 
predictions  of  significant  roles  for  nurturance,  and  the  organization  and 
ordering  of  information.  Both  G*C  ::  A*T  ratios  and  syntropy  have 
important  implications  for  emergences  in  designs.  Moreover,  whereas 
G*C  ::  A*T  ratios  represent  crude  quantitative  measures  of  nurtures,  for 
DNA  in  the  brain,  (G*C  ::  A*T)changing ltm  /  (G*C  ::  A*T)baseiine  are  crude 
measures  of  ‘syntropy’  -  with  (G*C  ::  A*T)baseiine  being  determined  for 
tissue  not  undergoing  changes  in  DNA  ( e.g .,  DNA  from  a  hair  follicle). 
Clearly,  if  (G*C  ::  A*T)baSeiine  is  chosen  as  a  crude  measure  of  nature,  then 
(G*C  ::  A*T)changing  LTM  /  (G*C  ::  A*T)baseiine  also  is  an  expression  of  the 
vast  importance  of  nurture  relative  to  nature.  These  crude  measures  of 
nature,  nurture  and  syntropy  also  represent  crude  measures  of  designs. 
[Increases  in  G*C  ::  A*T  are  indicators  of  increased  avidity  in  chemical 
bonding] 


Washington  Academy  of  Sciences 


33 


Nature  versus  Nurture  in  Emergences  in  Designs 

Because  of  ongoing  speculations  regarding  possible  intelligent 
designs,’  this  report  now  has  an  opportunity  to  both  debunk  intelligent 
design  arguments  while  clarifying  how  designs  may  emerge  (or  arise 
spontaneously  or  aberrantly)  in  living  and  non-living  systems.  In  the 
context  of  the  tripartite  model  of  evolution,  discussions  of  emergences  in 
designs  also  provide  unique  opportunities  to  highlight  sciences  of  the 
'real’  and  ‘artificial,’  as  well  as  evolution  in  designs.  Finally,  a  focus  on 
abnormal,  aberrant,  disordered  and  dysfunctional  designs  (cf.  Smith,  1979) 
can  provide  rich  opportunities  to  parlay  “Murphy’s  Law”  into  a  theory  of 
“debugging”  design,  disorder,  and  disease  (Smith,  1979).  Equally 
important,  an  emphasis  on  design  provides  opportunities  for  the  arts  to 
inform  the  sciences  and  engineering  ( e.g .,  see  “help  for  the  unknowingly 
needy  and  worried  well”  at  the  end  of  the  Glossary),  and  vice  versa. 

Debunking  ‘ Intelligent  ’  Design 

If  there  are  any  take-home  lessons  from  the  tripartite  model  of 
evolution,  they  are  that  the  three  interdependent  evolutionary  systems  are 
co-evolving  from  fundamentally  distinct  initial  “designs”  -  even  though 
those  systems  are  parsimonious  insofar  as  DNA  is  the  thread  underlying 
all  evolution.  The  Darwinian  model  of  evolution  works  well  for  random 
mutations,  selections  and  survival  of,  say,  the  most  ‘fit.’  It  even  can 
account  for  symbiosis  and  Archaebacteria  being  precursors  of 
mitochondria,  or  hydra  representing  an  assemblage  of  cells. 

Darwinian  evolution  is  less  successful  in  accounting  for 
consciousness,  nurturance,  spirituality,  elder  wisdom  or  other  elements  of 
the  transpersonal.  Each  of  these  evolutionary  developments  represents 
“designs”  according  to  definitions  cited  earlier,  though  there  are  no 
underlying  elements  of  intelligence  -  nor  are  there  any  reasons  to  invoke 
intelligence.  Indeed,  some  “designs”  may  be  extraordinarily  “beautiful,” 
“exquisite,”  and  “elegant”  -  as  in  the  structure  and  function  of 
countercurrent  mechanisms  in  kidneys  and  oceans.  Other  designs  may 
represent  “kludges”  (e.g.,  in  the  liver  and  brain)  or  even  the  unintelligent 
(e.g.,  in  the  interaction  between  human  female  ovaries  and  the  associated 
fimbrae).  Moreover,  the  evolutionary  emergence  of  the  umbilicus  and 
bony  cranium  may  have  been  central  to  the  tripartite  model,  whereas  in 


Winter  2006 


34 


other  animals  (or  plants)  entirely  different  evolutionary  schemas  may  be 
necessary. 

Perhaps  most  important,  the  evolution  of  gods  and  godliness  may 
recapitulate  the  evolution  of  evolution,  which,  in  turn,  may  recapitulate  the 
dispersion  of  matter  and  information  secondary  to  a  presumptive  'big 
bang.’  Stated  differently,  if  there  is  intelligence,  it  certainly  changes  and 
differs  throughout  evolution! 


Emergences  in  Designs  Associated  with  Nature 

As  noted,  the  proteome  comprises  less  than  2%  of  the  human 
genome,  or,  being  generous,  at  most  25%  if  allowances  are  made  for 
unknown  or  uncertain  regulatory  processes.  That  said,  evolutionary  and 
developmental  biologists  (/.£.,  the  “evo-devo”  movement)  have 
characterized  a  variety  of  "designs”  in  genes,  structures  and  functions 
which  are  parsimonious  across  species  and  over  time.  Indeed,  attendees  at 
the  Washington  Evolutionary  Systems  Society  Symposium  (March  25-26, 
2006)  were  treated  to  a  marvelous  plenary  presentation  by  Francis  Collins 
(Director  of  the  National  Human  Genome  Research  Institute)  in  which  he 
described  the  use  of  haplotype  mapping  (HapMap)  and  HGP  technologies 
to  identify  a  gene  site  implicated  in  progeria,  a  premature  aging  disease 
caused  by  a  “ de  novo”  gene  mutation  (Collins,  2006). 

Table  1  (see  the  end  of  the  article)  represents  an  attempt  to 
schematize  and  systematize  some  of  those  and  other  "designs”  associated 
with  nature,  though  without  unduly  focusing  on  specific  genes,  structures 
or  functions.  That  said,  it  is  instructive  to  contrast  humans  to  chimpanzees. 
One  finds  greater  than  95%  homology  between  their  proteomes,  yet  all 
will  agree  to  their  remarkable  differences.  Homologies  between  human 
and  chimpanzee  proteomes  and  genomes  also  underscore  their 
extraordinary  "designs.”  Both  species  retain  similar  and  analogous: 
physical,  biological,  and  physiological  characteristics;  sensory  and 
perceptual  features;  asymmetries  and  symmetries;  codes  and 
redundancies;  etc.  Differences  between  humans  and  chimpanzees  probably 
are  represented  in  the  structure,  codes,  and  values  of  information 
represented  in  non-proteomic  regions  of  their  genomes.  This  could 
represent  as  much  as  98%  of  the  human  and/or  chimp  genomes,  with  the 
95+%  homology  comprising  approximately  2%  of  their  genomes. 

Table  1  also  provides  a  paradigm  for  conceptualizing  emergences 
in  designs,  and  especially  in  living  systems.  "'Man-made’  /  artificial” 


Washington  Academy  of  Sciences 


35 


constructs  reveal  opportunities  for  inventions  and  professional  responses, 
whereas  “abnormal,  aberrant,  disordered,  and  dysfunctional”  possibilities 
point  to  diseases,  outliers,  and  other  anomalous  situations.  The  term 
“aberrant”  specifically  is  chosen  to  minimize  stereotyping  and 
unnecessary  (or  inappropriate)  value  judgments.  Its  earliest  use  is  in  our 
research  involves  individuals  whose  commonsense  differs  fundamentally 
from  others’  commonsense,  though  with  those  individuals’  commonsense 
generally  serving  their  needs  (Smith,  1987;  Smith,  1988). 


Emergences  in  Designs  Associated  with  Nurture  / Nurturance 

Possibly  the  most  significant  contribution  of  the  tripartite  model  of 
evolution  is  that  it  moves  discussions  of  evolution  beyond  phenotypes  and 
speciation.  According  to  the  DNA  change  hypotheses  in  the  immune 
system  and  the  brain,  evolution  exists  within  individuals  and  their  species 
-  and  especially  in  humans.  Passive  immunity  was  cited  because  of  its 
elegance  when  dealing  with  novel  pathogens  for  highly  mobile 
populations.  Indeed,  perhaps  the  greatest  tragedy  of  the  20th  century  has 
been  the  failure  to  recognize  that  the  AIDS  pandemic  fundamentally 
involves  distinctions  between  relatively  common  versus  relatively 
uncommon  pathogens  (Smith,  2004).  A  potential  avian  influenza 
pandemic  could  underscore  this  point  of  view,  though  with  far  more 
significant  consequences. 

As  noted,  nurturance  is  central  to  the  second  and  third  evolutionary 
pathways.  Table  2  (see  the  end  of  the  article)  represents  an  attempt  to 
systematize  and  schematize  emergent  designs  associated  with  nurture  / 
nurturance.  Not  surprisingly.  Tables  1  and  2  overlap  in  many  ways,  even 
though  examples  cited  in  Table  2  are  deliberately  limited.  If  one  considers 
the  contrast  between  humans  and  chimpanzees  cited  earlier,  it  immediately 
is  apparent  that  nurturance  contributes  greatly  to  the  explication  of  the 
“man-made  /  artificial”  and  the  “abnormal,  aberrant,  disordered  and 
dysfunctional.”  Nurturance  also  contributes  greatly  to  “life-span,” 
“methodological,”  “philosophical,”  and  the  “m eta-evolutionary  and 
metaphoric”  categories. 


What 's  Ahead  . . . 

The  tripartite  model  of  evolution  may  foreshadow  several  long- 
range  possibilities.  Just  as  Darwin’s  theory  of  evolution  could  give  rise  to 


Winter  2006 


36 


the  tripartite  model  of  evolution  in  some  animal  species,  future 
evolutionary  schema  may  include  heretofore  unforeseen  additional 
pathways  producing  further  possibilities  for  emergent  designs.  Man- 
machine  and  man-chemical  interfaces  must  rank  among  high-probability 
future  evolutionary  pathways.  Hang-gliding,  paragliding  and  rock- 
climbing  activities  all  point  to  the  potential  for  an  intersection  between 
nature  and  nurture  to  alter  phenotypic  genetic  traits  and  other  patterns  of 
nurtured  transmissions  over  time.  These  examples  are  cited  because  they 
reveal  man’s  potential  to  acquire  skills  generally  thought  to  be 
inaccessible  to  humans  ( e.g .,  flying).  Other  examples  include  the  use  of 
chemicals  to  enhance  performance  (e.g.,  in  sports)  or  interactions  with 
distant  contacts  using  computers  (e.g.,  using  the  Internet  or  when 
communicating  with  man  or  other  living  systems  during  space  travel).  Not 
to  be  overlooked  would  be  novel  possibilities  for  the  design  of  drugs  and 
other  pharmaceuticals.  Drugs,  chemicals  and  other  substances  targeting 
non-proteomic  regions  of  the  genome  -  and  especially  in  the  brain  - 
could  lead  to  alternative  treats  for  schizophrenia,  dissociative  identity 
disorders  (e.g..,  multiple  personalities),  “brain-washing,”  post-traumatic 
stress  syndromes,  and  other  diseases  /  syndromes  (cf.  Smith,  2003a) 

In  citing  these  examples,  one  should  not  interpret  them  as 
predictions.  Our  goal  is  to  stimulate  others’  imaginations  regarding  future 
evolution,  the  evolution  of  evolution,  and,  emergent  designs  associated 
with  nature  and  nurture  in  evolution.  A  further  goal  is  to  encourage 
dialogues  regarding  moral,  ethical,  and  other  philosophical  issues  (cf 
Gaarder,  1994). 

One  last  comment  in  regard  to  evolution  and  its  implications  for 
nature,  nurture,  and  emergent  designs.  Just  as  Darwin  found  evidence  for 
his  theory  of  evolution  of  species  after  exploring  the  Galapagos  Islands,  an 
‘island  theory’  also  has  value  in  the  exploration  of  nurture  (cf.  Smith, 
1994).  Much  can  be  learned  from  studies  of  nurturance  in  small  groups 
and  on  islands  -  whether  physical  (e.g.,  in  New  Zealand,  Sicily,  Sardinia, 
Hainan,  Iceland),  social  (e.g.,  cults,  prisons,  militaries  and  armies, 
ghettos),  political  (e.g.,  “red”  and  “blue”  states  in  the  USA)  and/or 
economic  (e.g.,  associated  with  caste  and  class).  Even  though  Hardy- 
Weinberg  equations  may  not  apply  in  regard  to  nurture,  analogous 
heuristics  reveal  the  value  of  studies  of  island  populations. 


Washington  Academy  of  Sciences 


37 


Conclusions 

A  novel  post-Darwinian  model  of  evolution  is  discussed  in  this 
brief  essay.  The  model  represents  a  modest  attempt  to  integrate  evolution 
across  species  with  a  broader  view  of  evolution  within  individuals,  within 
their  species,  and  across  species.  The  proposed  tripartite  model  of 
evolution  has  five  additional  salutary  features.  First,  it  provides  clues  to 
quantitative  assessments  of  nurture  and  the  relative  contributions  of  nature 
and  nurture  in  animals  with  brains  in  bony  crania  -  though  especially  in 
humans.  Second,  the  model  reveals  the  importance  of  understanding 
different  modalities  for  the  transmission  of  evolutionary  information;  to 
wit,  genetic  and  germ-line  transmission  constitute  a  relatively  small  part  of 
evolution.  Third,  the  model  points  to  alternative  futures,  some  of  which 
had  been  anticipated  accurately  ( e.g .,  HIV/AIDS;  see  Smith,  1979;  Smith, 
1984;  Smith,  1994;  Smith,  2001;  Smith,  2003a;  Smith,  2004),  and 
alternative  evolutionary  schemas.  Fourth,  the  model  reveals  that 
Lamarckian  and  Darwinian  notions  of  evolution  are  parsimonious,  with 
Lamarck’s  notion  of  adaptation  (Lamarck,  1819)  comporting  well  with  our 
notion  of  nurturance.  Finally,  it  may  be  reasonable  to  anticipate  drugs  and 
other  chemicals  (including  psychedelic  preparations)  having  direct  effects 
on  non-proteomic  regions  of  genomes  and  attendant  biochemical  pathways 
(e.g.,  associated  with  adenine  <->  adenosine  biochemical  pathways). 

The  possible  distinction  regarding  the  relative  contributions  of 
nature  and  nurture  in  evolutionary  settings  provides  a  unique  opportunity 
to  discuss  emergent  designs  associated  with  nature  and  nurture.  Although 
Tables  1  and  2  point  to  some  emergent  design  considerations  related  to 
nature  and  nurture  respectively,  these  Tables  are  by  no  means  complete. 
For  example,  psychological,  political,  and  economic  aspects  of  emergent 
designs  associated  with  nature  and  nurture  are  not  discussed.  The  latter 
would  include  discussions  of  early  detection  of  emergent  designs,  memory 
for  emergent  designs  (e.g.,  ‘Oscar  moments’  involving  memory  for  scenes 
in  movies),  transmission  of  emergent  designs  (e.g.,  on  radio,  television, 
and  the  Internet),  aberrant  processes  (e.g.,  “transmissible  negativism”  and 
aberrant  commonsense;  out-of-body,  near-death  and  past-life  experiences; 
etc.),  and/or  cultural  aspects  of  emergent  designs  (cf.  Smith,  2006c).  Nor 
do  Tables  1  or  2  address  emergent  design  issues  specifically  related  to  the 
separate  interdependent  evolutionary  pathways  -  and  especially  in  utero 
and  other  placental  pathways  (Gross,  2006;  Kriegs  et  ciL,  2006).  These 
will  be  the  basis  for  future  studies.  Finally,  opportunities  for  social  designs 
and  engineering  are  not  discussed  (e.g.,  novel  designs  and  theories  of 


Winter  2006 


38 


measurement  and  testing  which  take  into  account  nurturance  and  peer 
group  processes). 


References 

Aibib.  M.  A..  Billard.  A..  Iacoboni.  M..  and  E.  Oztop.  (2000).  Synthetic  Brain  Imaging: 

Grasping.  Mirror  Neurons  and  Imitation.  Neural  Networks  13.  975-997. 

Blakeslee.  S.  (2006).  Cells  That  Read  Minds.  New’  York  Times  January  10.  2006. 

Cleary.  J.  D..  and  Pearson.  C.  E.  (2003).  The  contribution  of  c/s-elements  to  disease- 
associated  repeat  instability  :  Clinical  and  experimental  evidence.  Cytogenetics  and 
Genome  Research  100:25-55. 

Collins.  F.  (2006).  Plenary  lecture  at  the  “2nd  Biennial  Pan-Affiliates  Conference  of  the 
Washington  Academy  of  Sciences  (Capital  Science  2006).“  [Washington.  DC  - 
March  25], 

Crick.  F.  H.  C.  (1958).  On  Protein  Synthesis.  Symposium  of  the  Society' for  Experimental. 

Biology’.  New  York:  Academic  Press.  XII.  138-163. 

Crick.  F.  (1970).  Central  dogma  of  molecular  biology.  Nature  227:561-563. 

Crick.  F.  H.  C..  Barnett.  L..  Brenner.  S..  and  Watts-Tobin.  R.  J.  (1961)  General  nature  of 
the  genetic  code  for  proteins.  Nature  192.  1227-1232. 

Darwin.  C.  (1859).  On  the  Origin  of  Species  by  Means  of  Natural  Selection,  or  the 
Presentation  of Fa\’oured  Races  in  the  Struggle  for  Life.  London:  John  Murray. 
Abermarle  Street. 

Ekman.  P.  (1973).  Darwin  and  Facial  Expression:  A  Century '  of  Research  in  Re\’iew\ 
New  York:  Academic  Press. 

Ekman.  P.  (2003).  Emotions  Revealed:  Understanding  Faces  and  Feelings.  New  York: 
Henry  Holt  and  Company. 

Ekman.  P..  &  Friesen.  W.  V.  (1975).  Unmasking  the  Face:  A  Guide  to  Recognizing 
Emotions  from  Facial  Clues.  Upper  Saddle  River.  NJ:  Prentice-Hall. 

Gaarder.  J.  (translator  Moller.  P.)  (1994).  Sophie's  World  -  A  Novel  About  the  History 
of  Philosophy.  New  York:  Berkley  Books  [See  Chapter  on  Plato  pp.  78-93.  and 
especially  the  discussions  on  “The  World  of  Ideas"  and  ‘‘True  Knowledge"]. 

Gross.  L.  (2006).  Resolving  the  Familv  Tree  of  Placental  Mammals.  PLoS  Biology  4(4). 
ell  1-el  12. 

Horrobin.  D.  F.  (1982).  "Peer  review:  is  the  good  the  enemy  of  the  best?"  Journal  of 
Research  Communication  Studies  3(4):327-34. 

Huxley.  J.  (1941).  The  Uniqueness  of  Man.  London:  Chatto  and  Windus. 

Iacoboni.  M..  Molnar-Szakacs.  I..  Gallese.  V..  Buccino.  G..  Mazziotta.  J.C..  and 

Rizzolatti.  G.  (2005).  Grasping  the  intentions  of  others  with  one's  own  mirror  neuron 
system.  PloS  Biology’  3(3).  E79.  pp.  0529-0535. 

Kriegs.  J.  O..  Churakov.  G..  Kiefmann.  M..  Jourdan.  U..  Brosius.  J..  and  Schmitz.  J. 
(2006).  Retroposed  elements  as  archives  for  the  evolutionary  history  of  placental 
mammals.  PLoS  Biology’  4(4).  0537-0544. 

Lamarck.  J-B.  (1819/1984).  Zoological  Philosophy  :  An  Exposition  With  Regard  to  the 
Natural  History  of  Animals  (reprint).  Chicago:  University  of  Chicago  Press. 


Washington  Academy  of  Sciences 


39 


McClintock.  B.  (1950).  The  Origin  and  behavior  of  mutable  loci  in  maize.  Proceedings  of 
the  National  A  cade  my  of Sciences  ( USA )  36:3 44-3  5 5 . 

Rizzolatti.  G.  and  Craighero.  L.  (2004).  The  mirror-neuron  system.  Annual  Review  of 
Neuroscience  27.  169-192. 

Sakano.  H.,  Huppi.  K..  Heinrich.  G..  and  Tonegawa.  S.  (1979).  Sequences  at  the 
recombination  sites  of  immunoglobulin  light-chain  genes.  Nature  280:288-294. 

Smith,  R.  W.  (1979).  Long-Term  Memories:  Where  Does  the  'Buck'  Stop?  —  Toward  a 
Testable  Theory  of  Debugging  the  Molecular  Basis  of  Long-Term  Memories  in 
Living  Organisms.  Abstracts .  Seventh  Meeting  of  the  International  Society  for 
Neurochemistry  f Jerusalem.  ISRAEL  —  September  2-6].  p.  590. 

Smith,  R.  W.  (1984).  AIDS  and  'Slow  Viruses'.  Annals  of  the  New  York  Academy  of 
Sciences  437:576-607. 

Smith.  R.  W.  (1987).  The  National  Impact  of  Negativistic  Leadership:  A  Need  for 
National  Caveats  Emptor.  In  Abstracts,  1987  Amiual  Meeting  of  the  Western 
Political  Science  Association  [Anaheim,  CA  —  March  26-28].  p.  28. 

Smith.  R.  W.  (1988).  Transmissible  Negativism  and  Its  Possible  Relation  to  Irrational 
Behavior  and  Poor  Common  Sense.  Presented  to  the  XXIV International  Congress  of 
Psychology  [Sidney.  AUSTRALIA  —  August  28  -  September  2\. 

Smith.  R.  W.  (1994).  On  Mechanisms  of  Slowness  and  Progressiveness  in  Slowly 
Progressive  Processes.  Annals  of  the  New  York  Academy  of  Sciences  724:430-434. 

Smith.  R.  Wm.  (2001).  Opinion:  The  Durban  Declaration.  The  Scientist  15(2):39. 

<hllp://w ww '.the-scicntisl.com/vT200 1/ian/opin  010 1 22  html> 

Smith.  R.  Wm.  (2003a).  Autovirulence  -  An  epigenetic  model  of  molecular  mimicry  in 
living  systems:  Broad  implications  for  aberrant  translation  products  and 
autoimmunity.  Abstracts .  XIX  International  Congress  of  Genetics  [Melbourne. 
AUSTRALIA  -  July  6-1 1 1.  Abstract  #2.G.0246.  p.  225. 

Smith.  R.  Wm.  (2003b).  Revisiting  the  Molecular  Biology.  Genetics  and  Genomics  of 
Long-Term  Memory  in  Living  Systems.  Abstracts ,  XIX  International  Congress  of 
Genetics  [Melbourne.  AUSTRALIA  -  July  6-11],  Abstract  #5.C.0802.  p.  133. 

Smith.  R.  Wm.  (2004).  Two  axioms  from  sentential  logic  used  to  disambiguate  cause, 
consequence,  infectious  and  non-infectious  cofactors,  opportunism  and  association  in 
lentivirus  and  other  infections:  Implications  for  revising  the  Henle-Koch  Postulates 
and  for  public  health  policy  on  infectious  diseases,  vaccines  and  vaccinations. 

Abstract  submitted  for  presentation  at  the  XV  International  AIDS  Conference 
[Bangkok.  THAILAND  -  January  4], 

Smith.  R.  Wm.,  and  Shadel.  R.  (2003).  Preliophic  Moleculator  Using  Electric  Fields  And 
Gradients  For  Manipulating  Molecules.  US  Patent  Application  #10/879.627. 

Smith.  R.  Wm.  (2006a).  Evolution  and  Long-Term  Memories  in  Living  Systems:  Using 
molecular  biology  to  resolve  three  great  debates  . . .  Lamarck  versus  Darwin.  Nature 
versus  Nurture,  and  the  Central  Dogma.  Presentation  at  the  Winter  Chaos  2006 
Conference  /  Snowflake  Forum  ( <ww  w.blueberrv- 

brain  om/winterchaos/.snowflake2006  html#roulette>)  [Pittsburgh,  PA  -  Februarv  3- 

5]. 

Smith.  R.  Wm.  (2006b).  Evolution  and  Long-Term  Memories  in  Living  Systems:  Using 
molecular  biology  to  resolve  three  great  debates  . . .  Lamarck  versus  Darwin.  Nature 
versus  Nurture,  and  The  Central  Dogma.  Presentation  to  the  San  Francisco  Tesla 
Society  ( <ww w , sftesla .org/N ewsletters/newslett  2006 .  htm>)  [San  Francisco.  CA  - 
February  12] 


Winter  2006 


40 


Smith.  R.  Wm.  (2006c).  Evolution  and  Long-Term  Memories  in  Humans:  Implications 
for  Theoretical  and  Clinical  Research  on  Consciousness  and  the  Transpersonal.  In 
Consciousness  Research  Abstracts  (Journal  of  Consciousness  Studies).  "Toward  a 
Science  of  Consciousness  2006"  Conference.  #215.  p.  154. 

Teilhard  de  Chardin.  P.  (1959).  The  Phenomenon  of  Man  (translated  from  Le  Phenomene 
Humain  by  Bernard  Wall).  New  York:  Harper  &  Row  Publishers.  Inc. 

Tonegawa.  S..  Maxam.  A.  M..  Tizard.  R..  et  al.  (1978).  Sequence  of  a  mouse  germ-line 
gene  for  a  variable  region  of  an  immunoglobulin  light  chain.  Proceedings  of  the 
National  Academy  of  Sciences  (USA)  75:1485-1489. 

Vemy.  T.  R..  and  Kelly.  J.  (1981/1983).  Das  See/enleben  des  Ungeborenen  -  Wie  Mutter 
und  J  ’ (iter  schon  fur  der  Geburt  Personlichkeit  und  Gluck  ihres  Kindes  fordern 
konnen  [The  Soul  Life  of  the  Unborn  -...].  Berlin:  Rogner  &  Bernhard. 

Zou.  Z..  and  Buck.  L.  B.  (2006).  Combinatorial  Effects  of  Odorant  Mixes  in  Olfactorv 
Cortex.  Science  311 : 1477-148 1 . 

Acknowledgments 

I  am  deeply  indebted  to  Jerry  L.  C.  Chandler,  Neill  Edwards,  Carlos  Torre, 
Anita  Rodriquez,  and  Vijay  Padmanabhan  for  their  personal  and 
professional  assistance.  Their  assistance  was  truly  invaluable! 

Glossary 

Aberrant  /  Aberration  /  Aberrancy  -  a  notion  that  some  traits  /  behaviors 
may  differ  from  normal,  modal  or  median-like  traits  /  behaviors.  The  term 
aberration  is  used  to  distinguish  between  occasional  situations  that  may 
arise  in  life  though  in  contrast  to  blatant  abnormality,  disease,  illness, 
and/or  other  “wrong”  or  implicitly  “negative”  contexts.  For  purposes  of 
this  report,  we  posit  my  variant  on  “Murphy’s  Law;”  to  wit,  “if  things  can 
be  different  in  living  situations,  those  different  situations  will  arise  in  life 
-  possibly  as  aberrations,  aberrant  situations  or  outliers.” 

Genome  -  The  full  complement  of  DNA  in  a  cell  of  a  particular  organism. 
In  humans,  the  genome  comprises  23  pairs  of  chromosomes  along  with  an 
“X”  and/or  “Y”  chromosome,  and  mitochondrial  DNA. 

Murphy's  Law  -  “If  something  can  go  wrong,  it  will  ...” 

Non-proteomic  -  referring  to  that  portion  of  cellular  DNA  which  does  not 
encode  for  proteins. 


Washington  Academy  of  Sciences 


41 


Nurture  -  Nurture,  which  often  is  contrasted  to  nature,  refers  to  the 
process  of  acquiring  and  replicating  learned,  cultural,  or  other  experiential 
information  and  its  possible  transmission  (usually  via  non-genetic,  non- 
proteomic  means)  to  others.  In  this  report,  a  central  thesis  and  premise  is 
that  nurture  most  often  is  reflected  in  non-proteomic  changes  in  DNA 
which  are  not  transmitted  to  the  host’s  germ-line  (/.*?.,  sperm  or  ova), 
whereas  nature  is  reflected  in  genetic  and  proteomic  (i.e.9  non-learned) 
traits  that  generally  are  transmitted  via  the  germ-line.  For  clinical  and 
pragmatic  purposes,  it  is  hypothesized  that  the  ratios  of  C*G  ::  A*T  DNA 
base-pairs  in  select  organs  ( e.g .,  in  selected  regions  in  the  brain  and  the 
immune  system)  represent  crude,  albeit  quantifiable ,  measures  of  nurture 
-  perhaps  the  first  quantitative  measures  of  nurture! 

Parsimony  -  A  notion  that  a  single  or  “best-fit”  “thread”  or  theme  may 
underlie  one  or  more  phenomena.  The  notion  of  parsimony  put  forth  in 
this  report  deliberately  goes  beyond  the  notion  of  Ockham’s  Razor  insofar 
as  those  which  may  be  “best-fit”  may  fundamentally  involve  complexities 
rather  than  mere  simplicities.  The  old  saw  or  adage  of  “keeping  it  simple, 
stupid”  [KISS]  is  rejected,  particularly  as  it  applies  to  the  transpersonal 
and  its  evolution. 

Preliophic  moleculator  -  Devices  and  processes  invented  in  1996  to 
capture  the  bidirectional  flow  of  molecular  information  based  on 
PRotonic-ELectronic-IOnic-PHotonlC  gradients  (Roulette  Wm.  Smith  and 
Robert  Shadel,  international  patents  pending).  The  device  is  called  a 
moleculator  for  MOLECUlar  calcu LATOR. 

Proteome  -  The  portion  of  the  genome  which  encodes  for  proteins. 

Proteomic  -  referring  to  the  proteome;  to  wit,  that  portion  of  the  genome 
which  encodes  for  proteins. 

Syntrophy  -  mathematical,  chemical  and  physical  measures  of  tendencies 
toward  organization  and  order  in  information  (as  contrasted  to  entropy). 
For  DNA  in  brain  and  if  G*C  connotes  Guanine*Cytosine  base  pairs  and 
A*T  connotes  Adenine*Thymine  base-pairs,  then  (G*C  ::  A*T)Changing  ltm 
/  (G*C  ::  A*T)baseiine  are  crude  measures  of  ‘syntropy’  -  with  (G*C  :: 
A*T)baseiine  being  determined  for  tissue  not  undergoing  changes  in  DNA 
(e  g.,  DNA  from  a  hair  follicle).  [NB:  An  obvious  example  of  syntropy 
occurs  when  single  complementary  strands  of  DNA  are  deposited  in  an 
aqueous  solution,  with  their  complementary  base-pairing  occurring  after 


Winter  2006 


42 


relatively  short  order.  For  practical  considerations,  'time’  is  not  a  variable 
in  the  calculation  of  syntropy;  rather,  syntropy  is  a  measure  of  increased 
order/organization.] 

Transmissible  -  the  acquisition  of  information  or  traits  by  genetic  and  or 
other  non-prole omic  genomic  mechanisms. 

Transpersonal  -  stages  and/or  states  of  human  development  through 
which  a  person's  self-awareness  extends  beyond  the  personal.  Classic 
examples  of  the  transpersonal  include  consciousness,  rational,  spirituality, 
mystical,  dreams,  etc.  For  this  discussion,  the  transpersonal  is  more 
broadly  defined  and  includes  soul,  spirit,  knowledge  and  beliefs,  erotetics 
(that  is,  the  logic  of  one’s  [especially  good]  questions),  competence, 
“commonsense,”  appearance,  taste  and  aesthetics,  wisdom  and  elder 
nurturance,  persistence  and  tenacity,  and  antecedent  (and  sometimes 
evolutionary)  conditions  which  may  define  or  shape  one’s  development. 

Help  for  Unknowingly  Needy  and  Worried  Well  -  (A  model  for  social 
design  and  engineering) 

It  is  said  that  ... 

“Mankind  may  be  divided  into  four  classes: 

( 1 )  Those  who  KNOW  and  know  that  they  KNOW  -  of  them 
seek  knowledge; 

(2)  Those  who  KNOW  but  do  not  know  that  they  KNOW  - 
awaken  them; 

(3)  Those  that  do  not  KNOW  and  know  that  they  do  not 
KNOW  -  instruct  them; 

(4)  Those  who  do  not  KNOW  but  think  that  they  KNOW- 
they  are  fools,  dismiss  them.” 

Salomon  Ibn  Gabirol  (also  known  as  Avicebron) 

In  Mibhar  Hu-Peninim  [ Choice  of  Pearls ] 

No.  60  (circa  1050  AD) 

[NB:  An  analysis  of  this  adage  reveals  that  those  with  and  without 
knowledge  may  be  partitioned,  albeit  somewhat  simplistically, 
according  to  their  education,  alertness,  motivation  and  ability 
to  educate.  The  range  of  possibilities  is  even  more  instructive  if 
the  words  “NEED”  and/or  “HELP”  are  substituted  for  the 
capitalized  and  italicized  word  “KNOW.”] 


Washington  Academy  of  Sciences 


43 


Table  1  -  Examples  of  Emergences  of  Designs  in  Nature 


Concepts, 
Formations  & 
Formalities 

Examples 

Man-Made  / 
Artificial 

Abnormal, 
Aberrant, 
Disordered  & 
Dysfunctional 

Geological 

formations 

Crystals;  geodes; 

Quartz  timers  and 
transistors 

Earthquakes; 
tsunamis;  etc. 

Physical, 
biological  and 
physiological 

Counter-current 
mechanisms  in 
oceans  and 
kidneys; 

Viruses  and 
infectiousness; 
Stem  cells; 
Velocity  - 
Electro 
negativity  of 
phosphates, 
sulfates,  etc.; 
genetic  tools 
(e.g.,  restriction 
enzymes, 
nucleases, 
kinases, 
proteases,  etc.); 

Submarines  and 
ballasts; 

Genetic 

engineering; 

Stem  cells; 

Gradients  in 
preliophic  systems; 
In  vitro 
fertilization; 

Cell  sorting; 

Genetic, 
congenital  and 
other  birth  or 
developmental 
disorders; 
trauma; 
bioterrorism; 
HIV/AIDS 

Sensory  and 
perception 

Sounds;  shapes; 
sights;  sizes; 
smells;  etc. 

Sounds;  shapes; 
sights;  sizes; 
smells;  etc. 

Synesthesia; 
dyslexia;  other 
diseases  of 
sensation  & 
perception; 

Biophysical 

Movements  & 
gradients; 
Microtubulin- 
associated 
proteins  (MAPs) 

Electrophoresis; 

Isoelectric 

focusing; 

Ampholytes; 

Aberrantly 

synthesized 

molecules 

Processes 

Cellular 

Electrophoresis; 

slow  viruses; 

molecular 

Preliophics; 

HIV/AIDS; 

information 

processes 

Winter  2006 


44 


Pathways 

Electron  chains 
and  proton 
processing 
Mitochondria 
Chloroplasts; 
Central  dogma; 
Inverse 
molecular 
information 
pathway; 
Biochemical 
pathways; 

Preliophic 

moleculators 

Genetic 
diseases  and 
defects 

Asymmetry, 
symmetry, 
parallelism  and 
segmentation 

Organisms 

Preliophics 

Tumors  & 

cancers 

Mathematics  & 
codes 

Fibonacci 

sequences 

Recursion; 

constants 

Redundancy 

G*C  and  A*T 
base  pairings; 

Redundancy  and 
fault  tolerance; 

Space  probe 
disasters 

Philosophical 

In  vivo; 
hi  virtualis 
(pre/iophics); 

In  vivo; 

A  priori  A 
posteriori; 
Evolutionary; 
Complexity; 

In  vitro; 

In  vivo^in 

vir  tit  a  list-tin  vitro; 

A  priori  A 
posteriori; 
Parsimony; 
Ockham’s  razor; 

Infection  & 
disease; 
Causality 
versus 

consequences; 
Illogical  & 
Aberrant 
logics; 

Metaphoric 

“Ontogeny 

recapitulates 

phylogeny” 

“Engineering 

recapitulates 

reality” 

Computer 
worms  & 
viruses; 
psychoviruses; 
re¬ 
engineering; 

Washington  Academy  of  Sciences 


Table  2  -  Examples  of  Emergences  of  Designs  Involving  Nurture 


Concepts, 

Formations 

& 

Formalities 

Examples 

Man-made  / 
Artificial 

Abnormal, 
Aberrant, 
Disordered  & 
Dysfunctional 

Geological 

formations 

and 

representa- 

Tions 

Statues; 

sculpture; 

arts; 

dream 

space  & 

time 

among 

aboriginal 

peoples 

“down 

under” 

Timers  and 
clocks;  cave 
paintings; 
movies;  edifices 
and  monuments 

Aberrant  measurements 
and  prurient  constructs; 
Inappropriate  habitats 
along  coasts, 
earthquake  faults, 
liquefaction  sites 

Physical, 
biological, 
physiological 
,  social, 
educational 

Nesting; 

parenting; 

fight- 

flight- 

fright 

mechan¬ 

isms; 

menses 

and  meno¬ 
pause; 
gender; 
passive 
immune 
transfers 
&  breast¬ 
feeding; 
imitation 
&  psycho¬ 
virus 
transmis¬ 
sions 

Homes;  schools; 
villages;  hunter- 
gathers;  justice; 
caste  (Laws  of 
Manu);  class; 
slavery;  war; 
religion; 
government; 
printing  press; 
agriculture,  radio, 
television, 
computers; 
internet;  iPODs; 
social  inventions; 
pushing  the  limits 

Prisons;  trauma; 
orphanages; 
bankruptcies; 
medical/surgical 
consequences; 
discrimination;  racism; 
divorce;  pandemics; 
war;  bombs;  autism 
syndromes;  HIV/AIDS; 
bioterrorism 

Winter  2006 


46 


Sensory  and 
perception 

Tastes; 

smell; 

other 

tactile 

responses; 

food; 

rituals; 

beliefs 

versus 

reality; 

Creativity;  fine 
arts;  literature; 
cuisine;  apparel; 
music;  dance; 
recreational  and 
ritual  substances; 
shamanistic 
practices; 
enhancement 
substances  and 
activities;  radio; 

TV 

Psychedelics  and 
substance  abuse; 
lack  of  thorough 
documentation  of 
extinctions; 
destruction  and 
disorders  of  senses  and 
perceptions; 

Biophysical 
and  meta¬ 
physical 

Sleep, 
dreams, 
rest  and 
relaxation, 
contempla 
tion; 
anticipa¬ 
tion 

Clairvoyance 

Sleep  disorders;  out-of- 
body  experiences;  near 
death  experiences;  past 
life  experiences; 
apparitions 

Transpersona 

1  and  other 
processes 

Conscious 

ness; 

commonse 

nse; 

desire; 

spirituality 

;  wit; 

humor; 

wisdom; 

compassio 

n;  gifts  & 

volunteeri 

sm; 

heroism 

Religion; 

philanthropy; 

archetypes 

Trauma;  autism; 
transmissible 
negativism; 
psychoviruses;  lying; 
Temporary’  autism; 
bigotry 

Pathways 

Careers; 

avocations 

and 

interest 

Religion;  spiritual 
& 

transformational 

quests 

Cults;  terrorism; 
exploitation; 

Asymmetry, 

symmetry. 

Central 

dogma 

Cooperation; 

competition; 

Trinucleotide  (TNR) 
diseases;  bankruptcies; 

Washington  Academy  of  Sciences 


47 


parallelism 

and 

segmentation 

versus 

inverse 

molecular 

informatio 

n 

pathways; 

genetic 

code 

versus 

inverse 

code 

governing 
storage  of 
molecular 
informatio 
n  in  DNA 

economic  models 
and  decision 
theory; 

globalization  in 
labor,  economies 
and  war;  business 
cycles; 

Mathematics 
and  codes 

Fibonacci 

sequence 

Recursion; 
heuristics; 
computability; 
Dewey  decimal 
system  &  Library 
of  Congress  filing 
schemas; 

international  book 
numbering 
systems;  DOI 
article  and  journal 
referencing 
codes; 

cryptography 

Computer  viruses  and 
worms 

Redundancy 

Redundan 
cy  in 
language 
and  brain 

structures 

/ 

processes 

Redundant 
designs  in 
aircraft, 
emergency 
vehicles,  & 
hospital  services 
and  procedures 

Apollo,  Challenger  and 
space  probe  failures 

Life-span 

Aging; 

elder 

wisdom; 

Insurance;  Social 
Security; 

Medicare  / 

Accidents,  war; 
diseases 

Winter  2006 


48 


legacy; 
impact; 
impor¬ 
tance; 
knowing 
when  and 
how  to 
stop; 
extinct¬ 
ions 

Medicaid; 
Elderhostels; 
adult  education; 
life  after  death 
and  reincarnation; 
archetypes 

Methodologi 

cal 

Census; 

simulation 

Epidemiology; 

preliophics 

Failure  to  record  and 
report  common  versus 
uncommon  pathogens 

Philosophical 

Subtlety; 
elegance; 
appreciati 
on;  open- 
minded¬ 
ness; 

verisimili¬ 

tude; 

reliability; 
validity; 
efficiency; 
phenomen 
ology; 
logic; 
causality; 
conse¬ 
quences; 
“good” 
question¬ 
asking  & 
“good” 
question 
answering 
(erotetics), 
paradigms 
& 

paradigm 

Peer  review; 
juries; 
autotoxicity; 
autovirulence; 
context- 
specificity; 
knowing  when 
and  how  to  stop 
(involving 
decisions, 
experiments, 
gambling, 
substance  abuse 
and  other 
addictions,  and 
war);  ge  dan  ken 
studies;  Henle- 
Koch  postulates 
for  a  single 
pathogen  causing 
a  single  disease 

“Is  the  good  the  enemy 
of  the  best?”  (Horrobin, 
1982);  Henl e-Koch 
postulates  for  complex 
infections  (e  g.,  EBV 
and  HIV);  Dogma 
including  Lthe  central 
dogma’,  'one  gene  - 
one  protein,  HIV  is  the 
sole  cause  of  AIDS, 
and  infectious 
pathogens  must  include 
nucleic  acids 

Washington  Academy  of  Sciences 


49 


shifts; 
morals  & 
ethic; 

Meta¬ 

evolutionary 

and 

metaphoric 

Structure 

versus 

function; 

cosmolog 

y  and  the 

cosmic 

gaming 

‘Big  Bang’; 
inflation  theory; 
God  and 
godliness; 
syntropy  versus 
entropy;  Does 
structure  precede 
function,  or  does 
function  precede 
structure?  Do 
sciences 

recapitulate  arts? 

War;  decline  and  fall  of 
empires;  Holocausts; 
extinctions  of  the 
endangered 

Winter  2006 


This  page  intentionally  left  blank 


Washington  Academy  of  Sciences 


51 


ENVIRONMENTAL  IMPACTS 
ON  HUMAN  MOODS  AND  EMOTIONS: 
IMPLICATIONS  FOR  WORKPLACE  AND  WORKFLOW 

DESIGN 

Thomas  Meylan,  Ph  D. 

EvolvingSuccess® 

Burtonsville.  MD 

Abstract 

The  human  body  is  designed  to  monitor  a  variety  of  multi-band 
channels  delivering  extremely  large  amounts  of  information  from  the 
environment.  It  is  also  designed  to  monitor  vast  amounts  of 
information  regarding  its  own  internal  states  and  conditions.  This 
information  flow  is  filtered  and  assessed  by  a  large  number  of  control 
loops  that  prepare  the  body  for  life-sustaining  activity.  These 
preparations  also  generate  a  large  number  of  subliminal  emotions  that 
start  to  intrude  themselves  into  a  person’s  inner  dialog  if  the  control 
loops  generating  them  remain  unclosed  or  unsatisfied  for  a  long  enough 
period  of  time.  The  lack  of  congruence  between  natural  environments 
and  the  typical  workplaces  inhabited  by  know  ledge  workers  means  that 
the  information  flow  received  by  these  control  loops  lacks  evidence 
that  the  workplace  can  sustain  life,  and  the  loops  remain  unclosed  and 
unsatisfied.  The  implication  of  these  information  deficient  work 
environments  on  knowledge  worker  mental  health  and  productivity  is 
discussed,  and  general  recommendations  made  for  re-engineering 
w  orkspaces  and  w  orkflow  . 


Introduction 

The  human  organism  is  intimately  linked  to  its  environment 
through  the  intake  of  massive  amounts  of  information  from  the 
environment.  This  is  facilitated  through  a  wide  variety  of  multi-band 
channels  into  the  body.  The  availability  and  openness  of  these  channels, 
combined  with  the  density  of  data  being  delivered  through  them,  set  the 
pedestal  level  for  the  psychological  comfort  of  the  healthy  human 
individual. 

In  addition,  information  flow  within  the  body  is  highly  dependent 
on  the  stimulation  received  through  the  senses,  and  upon  the  chemical  and 
hydration  state  of  the  body  itself.  Information-expectant  control  loops, 
either  chemically  based  or  based  in  the  nervous  system,  monitor  both  the 


Winter  2006 


52 


state  of  the  body  and  the  state  of  the  environment.  Depending  on  the  states 
detected,  these  control  loops  create  motivations  for  the  body  to  act  in 
accord  with  the  three  primary  drives  presented  in  an  earlier  paper,  and 
listed  in  a  section  below.1 


Illustrations  to  Expose  the  Phenomena 

On  average,  the  highest  rates  of  suicide  among  industrialized 
nations  occur  in  Scandinavian  countries.  Research  into  this  statistical 
oddity  revealed  a  phenomenon  now  called  seasonal  affective  disorder 
(SAD).  It  has  been  connected  to  the  relative  lack  of  exposure  to  full 
spectrum  sunlight  during  winter  months  brought  about  by  the  high 
latitudes  of  these  countries. 

Interestingly,  during  the  past  two  years,  news  reports  have 
occasionally  appeared  describing  research  on  sun-tanning  addiction. 
Apparently,  exposing  the  skin  to  solar  levels  of  UV  radiation  triggers  the 
release  of  endorphins.  For  many  people,  this  creates  a  condition  similar  to 
runner’s  high  (which  we’ll  talk  about  in  a  minute). 

Now  that  we  have  two  firmly  established  data  points  on  a 
phenomenon,  we  can  draw  our  uncontestable  straight  line  through  them 
and  (also)  draw  the  following  conclusion:  human  bodies  are  programmed 
to  seek  out  exposure  to  sunlight.  Why?  Mostly,  in  addition  to  triggering 
the  release  of  endorphins,  exposure  to  sunlight  also  triggers  the 
photosynthesis  of  vitamin  D  in  the  skin.  This  is  a  vital  element  in  good 
health. 

Let’s  get  back  to  runners’  high  to  anchor  another  uncontestable 
two-point  conclusion.  Physical  exertion  in  competitive  sports  is  widely 
known  to  create  positive  moods  in  a  large  segment  of  the  physically  active 
population.  Anecdotally,  one  often  hears  of  people  claiming  to  feel  very 
good  after  completing  heavy  physical  chores,  especially  if  they  have 
sedentary  professional  occupations. 

To  define  our  second  point  on  this  line,  we  look  at  the  rise  of  cases 
of  depression  during  the  20th  Century.  While  some  arguments  could  be 
made  that  a  significant  rise  in  depression  is  due  to  increased  reporting  of 
such  cases  to  doctors,  the  rise  of  depression  maps  very  well  to  the 
transition  of  large  segments  of  the  American  population  out  of  rural-based 
farming  lifestyles  into  urban-based,  less  physically  demanding  lifestyles. 


Washington  Academy  of  Sciences 


53 


The  level  of  physical  activity  has  continued  to  drop  off  for  most  people 
through  the  remainder  of  the  20th  Century  to  the  present,  and  the  condition 
of  mental  health  in  the  country  remains  relatively  poor.  (See  Seligman 
19952,  p.  37  for  summaries  of  four  studies  in  the  time-development  of 
depression  in  the  American  population.) 

What  do  we  conclude  here?  The  human  body  is  programmed  to 
monitor  its  own  level  of  physical  activity,  and  apparently  is  also 
programmed  to  reward  the  owner  with  certain  positive  emotional  perqs  if 
it  achieves  certain  levels  of  exertion.  Why  should  this  be?  Perhaps  the 
answer  is  as  simple  as  this:  active  animals  are  better  players  at  the  game  of 
natural  selection.  Successful  animals  have  to  hustle,  at  least  once  in  a 
while.  “Doing”  trumps  “being”  in  the  animal  kingdom.  So,  the  emotional 
programming  of  the  human  body  includes  emotional  rewards  based  on 
physical  exertion,  and  the  control  loops  that  deliver  them. 

In  other  words,  information  about  physical  exertion  is  generated  in 
the  body’s  chemistry.  The  readout  of  that  information  affects  the 
emotional  state  of  the  individual.  Various  levels  of  exertion  apparently 
create  a  spectrum  of  positive  emotional  experiences,  while  sedentary 
lifestyles  generate  increasingly  inert  piles  of  complacency  or  even  mild 
depressions.  The  control  loops  monitoring  life-sustaining  activities 
“know”  when  a  body  isn’t  being  used  properly,  and  “punish”  accordingly 
with  an  increasingly  unsatisfying  emotional  experience  of  life. 

These  anecdotal  musings  provide  us  with  two  illustrations  of 
information  management  in  a  human  body.  One  illustration  provides  us  a 
glimpse  of  information  collection  strategies  to  monitor  external 
conditions.  The  second  one  shows  us  that  large  amounts  of  information 
are  also  generated  within  the  body  itself,  and  is  in  fact  utilized  in  creating 
or  destroying  various  motivational  mechanisms.  Let’s  mark  down  two 
simple,  more  or  less  self-evident  conclusions: 

1.  The  emotional  experiences  of  people  are  highly  dependent  on 
information  collected  from  the  environment  within  which  they 
find  themselves,  and, 

2.  The  emotional  experiences  of  people  are  highly  dependent  on 
the  information  generated  within  their  own  bodies  about  the 
conditions  within  them. 

In  the  rest  of  this  article  we  will  propose  answers  to  the  following 
questions: 

•  Where  does  all  of  this  information  come  from? 


Winter  2006 


54 


•  How  does  this  information  affect  emotional  states? 

•  Why  do  modem  work  environments  affect  so  many  people 
adversely? 

•  How  can  a  manager  re-engineer  workspace  and  workflow  to 
emulate  the  environments  human  bodies  expect  to  function  in? 


Methodology 

There  may  be  some  who  will  be  less  than  satisfied  with  the  lack  of 
rigorous  connection  to  a  base  of  scholarly  or  professional  literature  in  this 
presentation.  However,  the  literature  in  evolutionary  psychology  (“ev 
psych”  for  convenience),  the  field  which  provides  the  basis  of  this 
presentation,  is  aimed  at  a  very  different  set  of  studies.  To  be  both  candid 
AND  fair,  it  has  established  a  single  principle  of  study,  that  being  the 
application  of  natural  selection  to  the  formation  of  successful  animal 
behavior,  and  for  humans,  the  formation  of  both  successful  behavior  and 
thought.  If  we  view  thought  as  a  form  of  behavior  instead  of  as  a  non¬ 
material  phenomenon  associated  exclusively  with  humans,  we  get  a  better 
handle  on  dealing  with  this  admittedly  difficult  topic. 

The  problem  with  much  literature  in  ev  psych  is  that  it  has  jumped 
to  big  problems  before  is  has  refined  the  way  its  primary  principle  is 
applied  to  human  thought  and  behavior  at  a  smaller  level.  It  is  trying  to 
answer  questions  from  other  fields  before  it  has  adequately  defined 
questions  arising  from  its  own  initial  inquiries. 

To  illustrate  this  point,  let’s  briefly  look  at  the  history  of 
astronomy.  Ancient  Greek  astronomy  took  its  foundational  premise  from 
contemporaneous  philosophy,  which  stated  that  in  the  perfection  of  the 
heavens,  all  heavenly  bodies  travel  in  perfect  circles.  As  we  know  from 
the  history  of  astronomy,  it  was  very  difficult  to  reconcile  eye-ball 
observations  of  planetary  motions  with  this  “accepted  truth.”  With  this  is 
mind,  let’s  consider  an  example  from  literature  in  ev  psych.  An  early 
popular  book  in  this  field  made  a  similar  jump  from  philosophy.  In  The 
Moral  Animal  (Wright  19943)  the  author  explicitly  believes  that  humans 
exhibit  moral  behavior,  and  then  attempts  to  use  ev  psych  to  explain  why 
this  is  so.  However,  like  the  premature  acceptance  of  circles  in  Ptolemaic 
astronomy,  the  use  of  the  term  “moral  behavior”  is  premature  in  that  it 
establishes  a  category  which  may  not  necessarily  simplify  the  scientific 
study  of  human  thought  and  behavior,  especially  when  the  term  is  used  to 


Washington  Academy  of  Sciences 


55 


form  value  judgments  instead  of  define  a  quantity  suitable  for  objective 
measurements. 

What  we  instead  are  attempting  in  this  presentation  is  “rough 
science”  (to  steal  the  title  of  a  great  PBS  television  program)  on  field 
studies  of  large  primates  in  their  most  commonly  observed  setting.  These 
are  populations  of  humans  at  work.  Our  team  at  EvolvingSuccess  has  a 
combined  experience  of  over  100  years  in  observing  and  working  with 
people  in  a  very  wide  range  of  knowledge-based  industries.  We  also  have 
over  50  years  of  combined  experience  managing  people  in  professional 
contexts  which  include  the  pastoral  ministry,  human  resources 
management  and  training,  large  computer  system  integration  management 
for  Federal  and  Fortune  100  clients,  as  well  as  real  time  astronomy 
satellite  operations  and  research  center  management. 

To  work  through  this  large  trove  of  admittedly  anecdotal  data,  we 
have  derived  a  small  number  of  ideas  from  the  basic  principle  of  ev  psych. 
In  traditional  academia,  ev  psych  is  used  to  explain  various  phenomena  in 
humans  based  on  the  premise  that  most  of  these  behaviors  occur  as 
services  to  the  genes  carried  by  animals’  bodies  and  were  optimized  for 
conditions  one  to  three  million  years  ago  to  assure  the  transmission  of 
these  genes.  For  instance,  about  five  years  ago  numerous  news  magazines 
and  nightly  news  television  shows  picked  up  findings  about  human  mate 
selection  strategies  based  on  a  few  ev  psych  studies.  Men,  it  was 
determined,  try  to  engage  as  many  mates  as  possible  in  order  to  spread 
their  genes  around  in  the  greatest  numbers  possible.  Women,  who  by  the 
nature  of  things  can’t  arbitrarily  throw  their  genes  around,  tend  to  select 
well-healed  mates  to  assure  that  the  small  number  of  offspring  they 
produce  will  live  to  sexual  maturity,  start  their  own  reproductive  activities, 
and  therefore  keep  their  genes  moving  more  broadly  into  the  ecosystem. 

As  astrophysicists  (turned  high  tech  business  people),  with  a 
somewhat  more  cosmological  view  of  things,  we  find  this  “selfish  gene 
worldview”  somewhat  silly,  and  certainly  overly  anthropomorphic.  Even 
if  there  is  a  form  of  geno-mechanics  that  facilitates  a  functional 
“selfishness”  in  genes,  if  the  environment  says  a  certain  combination  of 
genes  loses  the  game  of  natural  selection,  then  that  combination  loses, 
plain  and  simple.  It  doesn’t  matter  how  selfish  genes  are,  they  don’t  come 
near  to  having  the  final  say  in  the  course  of  evolution. 

Our  work,  on  the  other  hand,  attempts  to  reassert  the  importance  of 
environment  in  understanding  human  thought  and  behavior  as  a  product  of 
natural  selection.  This  is  an  explicitly  macroscopic  view  of  natural 


Winter  2006 


56 


selection.  Humans  have  changed  the  environmental  pressures  upon 
themselves  faster  than  natural  selection  can  keep  up,  and  consequently  the 
information  systems  currently  installed  in  human  animals  are  tuned  to  a 
different  set  of  pressures  than  humans  currently  face  in  modern  work 
places.  Our  model  of  human  information  processing  sub-systems, 
produced  via  the  method  of  system  reverse  engineering  (and  presented  in 
the  Fall,  20051  issue  of  this  journal)  suggests  epochs  for  the  appearance  of 
each  of  these  sub-systems  scattered  back  through  several  hundred  million 
years  of  natural  history.  Our  argument  is  if  a  sub-system  emerged  a 
million  years  ago,  it  is  probably  tuned  to  conditions  at  that  place  and  time 
much  better  than  it  is  tuned  to  current.  First  World,  knowledge-based  work 
environments. 

The  aim  of  this  interpretive  work  is  to  generate  practical 
applications  from  ev  psych.  Our  ambition  is  to  derive  repeatable  human 
capital  management  techniques  that  will  improve  business  performance  in 
knowledge-based  companies  or  the  performance  of  any  service-oriented 
organization.  The  human  animal  is  built  for  best  performance  in 
environments  other  than  the  modem,  knowledge  industry  work  place. 
How  can  managers  adapt  themselves  and  their  work  forces  to  overcome 
and/or  take  advantage  of  this  reality? 

By  examining  our  anecdotal  dataset,  collected  over  50  years  of 
highly  successful  managerial  experience  with  the  interpretive  tools  made 
available  by  evolutionary  psychology  and  information  system  engineering, 
we  hope  eventually  to  answer  that  question.  In  the  present  paper,  we  offer 
some  observations  about  the  effects  of  modern  work  environments  on 
human  emotion  as  distilled  from  our  managerial  experience  by  this 
interpretive  approach.  The  observations  suggest  a  few  high-level  strategies 
for  modifying  work  environments  to  achieve  better  performance  from 
people  working  outside  of  “their  naturally  selected  comfort  zones.” 


Natural  Selection,  Information  Requirements  for  Large  Mammals, 
and  the  Channels  Currently  Utilized 

j 

As  noted  in  the  previous  paper  in  this  sequence  ,  the  following 
drives  serve  as  guiding  principles  for  animal  success  in  general,  and  as 
system  design  requirements  for  animal  information  processing  in 
particular: 

•  The  drive  to  eliminate  or  avoid  all  forms  of  pain  or  discomfort. 


Washington  Academy  of  Sciences 


57 


•  The  drive  to  have  sex. 

•  The  drive  to  nurture  offspring  to  self-sufficiency  in  the  shortest 
time  possible. 

In  this  macroscopic  context,  all  living  things  are  highly  dependent 
on  information  collection  and  response  to  play  the  game  of  natural 
selection.  However,  the  information  requirements  for  animals  are  huge 
when  compared  to  the  members  of  other  living  kingdoms.  This  becomes 
ever  more  clear  when  we  attempt  to  map  out  the  channels  carrying 
information  into  an  animal  body.  One  also  begins  to  appreciate  the 
tremendous  power  of  human  sensory  processing  capability  when  you  take 
a  close  look  at  those  channels.  These  channels  look  extremely  wide-band 
from  one  perspective,  but  most  of  them  are  not.  Human  eyes,  for  example, 
can  collect  data  from  only  about  one  octave  out  of  the  electromagnetic 
spectrum  (being  sensitive  from  roughly  350  nanometers  to  700 
nanometers:  doubling  the  wavelength  gives  you  an  octave).  That  is  a  mere 
sliver  of  the  entire  range  of  energies  photons  can  deliver.  Yet,  even  within 
this  narrow  range  of  photon  energies,  the  eye  and  brain  divide  those 
energies  into  separate  channels  that  we  experience  as  individual  color 
sensations. 


The  Channels  and  Detectors 

Contrary  to  subjective  impression,  human  sensory  organs  are 
detecting  stimuli  on  a  24x7  basis.  Furthermore,  all  sensory  organs  are 
functioning  simultaneously.  “Back-office”  processing  of  this  information 
changes  with  the  24  hour  diurnal  cycle.  It  changes  when  various 
stimulation  thresholds  are  breached.  It  changes  more  drastically  when 
various  internal  conditions  fall  more  greatly  out  of  balance,  and  the  body 
begins  to  seek  out  resources  for  rebalance.  Pre-processing  in  the  sensory 
organs,  as  well  as  back-office  processing  in  the  brain,  change  even  more 
quickly  when  threatening  conditions  are  detected. 

Let’s  list  and  briefly  work  through  the  various  channels  which 
deliver  information  to  the  human  body.  The  channels  under  consideration 
here  are  limited  to  those  that  provide  information  about  the  external 
environment  with  which  the  body  is  not  in  direct  physical  contact,  as 
required  for  touch  and  taste.  However,  the  channels  that  remain  go  beyond 
sight,  sound  and  smell. 

•  Infrared  radiation  channel 


t 


Winter  2006 


58 


Infrared  radiation  is  experienced  as  heat  or  warmth.  It  is  sensed 
mostly  through  the  skin,  and  its  detection  tells  us  whether  we  need  to  find 
a  warmer  place  or  a  colder  place,  how  many  clothes  to  put  on,  or  how 
many  to  take  off  While  your  skin  isn’t  able  to  produce  images,  it  can  do 
pretty  well  at  detecting  the  direction  of  a  source  of  heat,  like  the  sun  or  a 
large  fire  in  the  distance. 

The  detection  of  infrared  radiation  by  your  skin  also  affects  your 
body  chemistry,  and  a  variety  of  physiological  conditions  change  with  the 
level  of  heat  your  skin  detects.  Extremes  of  heat  or  cold  often  lead  to 
strong  emotional  responses  such  as  fear. 

•  Visible  spectrum  light  channel 

This  is  the  light  detected  by  properly  functioning  eyes.  Even 
though  the  eyes  only  detect  approximately  one  octave  of  the  entire 
electromagnetic  spectrum,  it  is  possible  that  this  represents  the  most  data- 
dense  channel  that  human  beings  use.  The  imaging  and  resolution 
capabilities  of  human  eyes  provide  greatly  detailed  information  about  the 
environment  extending  for  several  miles,  and  very  much  useful 
information  for  as  far  as  the  air  is  clear.  The  combination  of  two  eyes  also 
provides  very  good  information  about  distances,  sizes,  changes,  speeds, 
and  many  other  quantities. 

The  information  from  the  eyes  is  also  color-coded.  Color  is  an 
information  enhancement  artifact  generated  by  eyes  and  brain;  colors  as 
perceived  are  arbitrary  from  a  universal  standpoint,  but  the  colors  do 
convey  real  information.  They  help  to  interpret  the  content  of  the 
environment.  Green  is  a  soothing  color,  presumably  because  it  represents 
locations  where  food  and  water  can  be  obtained.  Likewise,  sparkling 
things  attract  our  attention,  probably  suggesting  the  presence  of  open 
water. 

There  is  also  a  type  of  geometric  coding  that’s  important  for 
vision.  Human  eyes  are  part  of  an  exceptionally  complex  pattern 
recognition  system,  and  the  patterns  they  are  tuned  to  recognize  are  fractal 
patterns.  These  are  the  patterns  in  the  shapes  of  trees,  or  the  shapes  of 
clouds,  or  river  streams,  or  even  the  textures  observed  in  a  field  of  grass. 
This  is  the  geometry  of  the  natural  environment  to  which  all  human 
information  processing  systems  are  tuned. 

When  an  individual  is  deprived  of  exposure  to  natural  colors,  like 
the  greens  of  healthy  vegetation,  and  deprived  of  exposure  to  settings 
dominated  by  fractal  geometry,  many  information  processing  loops  in  the 


Washington  Academy  of  Sciences 


59 


body  go  unclosed.  This  is  interpreted  by  the  body’s  systems  as  though  it  is 
in  a  resource-poor  environment  and  thus  creates  any  of  a  number  of 
anxieties. 

•  Ultraviolet  radiation  channel 

Ultraviolet,  or  UV,  radiation  is  also  mainly  detected  by  the  skin. 
Its  most  obvious  effect  on  the  skin  is  tanning  in  the  sun.  You  can  even 
create  simple,  contact  images  on  your  skin  if  you’re  patient  enough  (or 
have  a  “properly  vented”  swimsuit).  But  tanning  is  only  one  of  a  family  of 
chemical  reactions  in  the  skin  that  strongly  affect  mood.  As  noted  above, 
there  is  a  strong  correlation  between  exposure  to  UV  sunlight  and 
emotional  states. 

•  Audio  signal  channel 

Audio  signals  received  by  the  ears  represent  perhaps  the  second 
densest  channel  of  information  into  your  body.  You  don’t  get  quite  the 
detail  that  you  do  with  vision,  but  information  about  location,  distance, 
speed,  and  size  can  be  obtained.  Like  color  in  vision,  audio  signals  are  also 
coded  in  pitch.  And  also  like  vision,  your  sense  of  hearing  is  equipped 
with  pattern  recognition  capabilities  that  look  for  fractal  geometry  patterns 
with  respect  to  time.  Rhythmic  patterns  in  music  tend  to  be  fractal, 
produced  by  the  continued  halving  of  the  durations  of  notes  and  the 
inteijection  of  percussive  events  splitting  the  time  between  two  other 
simple  musical  events. 

Sound  also  tells  you  how  rich  in  resources  your  environment  is. 
Most  people  enjoy  the  sounds  of  running  water,  like  streams  or  waves  on 
the  beach.  The  sounds  of  birds  are  also  pleasing  to  most  people.  Why? 
They  are  clear-cut  clues  that  you  can  find  food  and  water  nearby.  When 
your  sense  of  hearing  is  deprived  of  these  natural  sounds,  you  begin  to 
experience  anxieties  out  of  concern  for  a  lack  of  resources  to  keep  alive. 

•  Infrasonic  signal  channel 

The  deep  bass  of  thunder,  waves,  and  earthquakes  are  perceived  as 
pressure  waves  on  the  body  as  a  whole.  These  waves  are  perceived  like 
touch  instead  of  like  sound.  The  thumping  feet  of  large  animals  may  also 
be  detected  in  this  way,  providing  something  of  a  warning  of  approaching 
predators.  High  pressure,  low  frequency  acoustic  waves  elicit  a  variety  of 
emotional  responses  depending  on  the  perceived  source  of  the  stimulus. 

•  Chemical  detection  channel 


Winter  2006 


60 


Perhaps  the  most  under  appreciated  channel  affecting  mental 
health  is  the  sense  of  smell.  Compared  to  vision,  the  sense  of  smell  doesn’t 
offer  a  wealth  of  detail.  About  all  you  can  tell  when  you  smell  something 
is  that  there’s  a  source  of  the  smell  somewhere  nearby.  You  can  also  tell 
whether  or  not  the  smell  is  likely  to  be  good  for  you  in  greater 
concentrations. 

But  here’s  the  key.  Everything  in  your  body,  including  all  of  your 
information  processing  systems,  is  run  on  biochemistry.  Life  operates  on 
the  basis  of  the  chemical  resources  that  an  animal  body  can  find.  If  you 
can’t  find  food  and  water  you’re  dead.  That’s  why  being  able  to  detect 
vital  chemicals  in  your  environment  is  so  important. 

The  chemical  detectors  of  the  nose  are  the  only  sensory  detectors 
directly  wired  to  the  brain.  What’s  fascinating  is  that  your  nose  is  busy 
detecting  chemicals  that  you  don’t  even  sense  as  smell  Yet,  when  your 
nose  detects  key  chemicals  in  the  air,  it  signals  the  brain  that  the 
appropriate  resources  are  close  by. 

Conversely,  when  your  environment  is  filled  with  filtered  air,  your 
nose  is  deprived  of  evidence  that  you  are  living  in  a  resource-rich 
environment.  For  lack  of  chemical  evidence  detected  by  your  nose,  you 
start  to  become  anxious. 

Let’s  collect  together  the  channels  for  easier  reference. 

•  Infrared  radiation  channel 

•  Visible  spectrum  light  channel 

•  Ultraviolet  radiation  channel 

•  Audio  signal  channel 

•  Infrasonic  signal  channel 

•  Chemical  detection  channel 

The  reason  we  call  these  channels  (aside  from  the  fact  that  that’s 
what  they  are)  is  to  evoke  a  specific  image.  The  “transmitter”  of  these 
channels,  if  you  will,  is  the  environment  itself.  These  channels  create  a 
connection  between  the  environment  and  the  individual  human  attempting 
to  succeed  in  it.  We  can  then  think  of  the  human  being  as  a  mobile 
computing  unit  operating  in  a  vast  network  of  data  sources. 

What  makes  this  a  point  of  criticality  is  that  the  mobile  computing 
unit  (i.e.,  the  human  being)  is  equipped  with  extensive  arrays  of  pre¬ 
initialized  condition  expectations.  These  amount  to  the  triggers  that  make 
us  feel  good  or  feel  bad  under  any  given  set  of  circumstances.  For  large 


Washington  Academy  of  Sciences 


61 


mammals  in  particular,  the  deviations  from  these  environmental 
expectations  (or  deviations  from  internal  “specifications”  as  well)  are 
extremely  closely  monitored  by  the  systems  that  create  our  emotions.  As 
noted  in  the  previous  paper1,  emotions  exist  to  function  as  condition 
assessments,  alerts,  and,  when  appropriate,  the  signals  that  a  drive 
satisfaction  behavior  has  succeeded. 

Let’s  look  at  this  more  closely  in  the  next  section. 


Naturally  Selected  Sensory  Targets,  Emotions,  and  Drive  Satisfaction 

Strategies 

In  the  previous  section  we  listed  the  various  channels  that  your 
body,  as  a  node  on  the  environment’s  “wireless  network,”  is  connected  to. 
These  wireless  channels  provide  data  to  your  body  about  the  state  of  your 
environment,  and  whether  it  can  sustain  your  life  easily  or  not. 

If  you  were  to  remove  this  node  (your  body)  from  this  wireless 
network  of  environmental  data  resources,  it  would  receive  MUCH  LESS 
information  to  use  in  determining  how  well  you  can  live  where  you  are. 
The  flow  of  data  from  the  natural  environment  would  drop  to  nearly  zero, 
which  basically  tells  your  body  you  are  living  in  a  completely  barren 
location  without  food  and  water.  Over  a  long  enough  time,  this  elevates  a 
cluster  of  negative  emotions  into  the  awareness  of  your  inner  dialog, 
where  you  start  trying  to  figure  out  what’s  driving  you  crazy.  This 
becomes  increasingly  worse,  because  you  can’t  figure  out  why  you  feel 
this  way. 

Working  in  an  office  in  front  of  office  machinery  is  one  way  to 
remove  a  node  from  the  natural  environment.  You’re  working  in  a 
perfectly  comfortable  place,  are  perfectly  well  fed  and  watered,  and  the 
job  is  going  great.  Yet  you  still  feel  antsy  enough  to  run  out  of  your  office 
in  a  panic  (or,  at  least,  a  good  number  of  people  experience  vaguely 
unsettled  or  unsatisfactory  emotional  states).  This  doesn’t  come  about 
from  “an  urge  to  be  free.”  It  comes  about  because  your  body  arouses  itself 
to  seek  an  environment  that  more  obviously  and  assuredly  will  support  its 
life.  Most  minor  workplace  anxieties  would  evaporate  for  people  if  they 
simply  got  some  serious  exposure  to  active  natural  settings  everyday,  a 
couple  times  a  day. 

In  the  previous  paper  outlining  the  primary  EvolvingSuccess 
model  of  human  thought  and  behavior1,  we  made  use  of  the  following 


Winter  2006 


62 


diagram  to  describe  how  an  emotional  experience  evolves  with  time  while 
a  given  drive  goes  unbalanced  or  unsatisfied. 


Figure  1:  Time  Dependence  of  Subjective  Emotional  Experience  when 
a  Drive  or  Data-handling  Function  is  out  of  Balance 


Every  one  of  the  data  input  ports  in  the  human  body  is  followed 
by,  or  attached  to,  some  cluster  of  data  evaluation  systems.  These 
evaluative  functions  could  be  hosted  on  a  chemically-based  process,  a 
neurological  process,  or  (most  commonly)  a  combined  assessment 
process.  They  store  the  needed  standards  by  which  to  judge  the  incoming 
data  as  containing  “good”  information  (yielding  a  positive  emotional 
experience,  no  matter  how  fleeting),  or  “bad”  information  (initiating  the 
downward  sloping  curve  of  a  negative  emotion  that  will  persist  until  the 
information  changes,  or  time  reduces  its  significance). 

The  vast  majority  of  these  evaluative  systems,  and  there  are  many 
of  them,  do  not  generate  information  that  creates  an  inner  dialog  event, 
meaning  they  do  not  generate  thoughts.  They  aren’t  strongly  connected  to 
information  processing  systems  that  operate  through  symbols  that  human 
beings  typically  use  to  understand  and  communicate  their  experiences,  as 
illustrated  in  Figure  2  below.  They  usually,  if  not  always,  generate  an 
emotional  response  that  only  rarely  breaks  through  the  subliminal  into  the 
inner  dialog  with  a  label  (like  anxiety). 


Washington  Academy  of  Sciences 


63 


Figure  2.  Schematic  Representation  of  the  Four  Information  Sub¬ 
systems  at  Work  in  the  Human  Body,  with  emphasis  on  the  poor 
connection  between  the  dominant  sub-system  of  emotions  and  the  less 
well  developed  problem  solving  sub-system. 


Symbol-based 
Problem  Solving 
Sub-system 


Pre-tuned  Condition  Alert 
Sub-system  (Emotions) 


Sensory  Data  Collection 
and  Response  Sub-system 


Chemical-based  Information  Processing  Sub-system 


There  are  two  things  about  this  emotional  response.  Of  secondary 
interest  to  this  article  is  the  fact  that  the  experience  of  an  emotion  is 
actually  only  a  bi-product  of  processes  which  are  preparing  the  body  to 
work  on  re-balancing  one  of  the  three  drives.  Of  primary  importance  is 
that  all  we  become  aware  of  in  our  conscious  experience  is  a  feeling,  and 
that  feeling  is  often  difficult  to  hook  to  a  clear  event  in  the  environment. 

What  does  this  mean  for  the  everyday  experience  of  life  in  an 
ordinary  work  environment?  It  means  that  most  of  our  feelings  at  any 
given  time  are  the  product  of  a  vast  number  of  environment  assessment 
processes  over  which  we  have  very  little  direct  awareness  and  no  direct 
control  whatsoever. 

There  is  another  factor  for  emotional  comfort  implied  by  the  curve 
in  Figure  1:  the  factor  of  time.  Each  of  the  data  evaluation  systems 
includes  a  time  dependence.  The  deadlines  for  behavioral  responses  for 
virtually  all  of  them  are  very  short  when  compared  to  the  length  of  the 
average  knowledge- work  project.  For  example,  the  time  it  takes  to  flee 
from  a  predator,  or  conversely  to  chase  down  a  running  animal,  is 
measured  in  seconds  on  the  short  side,  and  one  or  two  minutes  on  the  long 


Winter  2006 


64 


side.  If  a  satisfactory  conclusion  to  the  event  occurs,  then  all  of  the 
evaluation  systems  reset  to  an  emotionally  neutral  state.  If  they  don’t, 
well,  if  you’re  prey  you’re  eaten,  and  if  you’re  predatory  you’re  still  on 
the  prowl.  If  you  don’t  find  food  for  several  days  you  continue  on  your 
quest  until  you  succeed  or  you  starve  (a  couple  of  weeks  tops). 

Knowledge  work,  on  the  other  hand,  often  entails  projects  that  run 
for  months  or  even  years.  Even  if  the  job  is  piece  work,  like  data  entry,  the 
sedentary  nature  of  the  job  and  the  confinement  to  a  single  desk  or  cube 
means  that  an  individual  is  not  functioning  properly,  from  a  system-design 
point  of  view. 

We  wrote  early  in  this  paper  that  physical  activity  affects 
emotional  states,  where,  within  certain  bounds,  great  physical  exertion 
tends  to  create  positive  emotions.  Sedentary  lifestyles,  in  general,  lead  to 
more  negative  moods,  especially  where  there  is  also  a  lack  of  mental 
stimulation  such  as  reading  or  problem  solving.  This  indicates  that  there 
are  also  many  data  assessment  processes  that  are  monitoring  internals 
states  in  the  body.  As  a  self-regulating  system,  the  human  body  is  full  of 
data-driven  control  loops  that  motivate  the  animal  to  exert  itself  in  the 
satisfaction  of  its  drives. 

When  there  is  a  lack  of  bodily  activity,  many  of  these  control 
feedback  loops  remain  out  of  balance  in  the  manner  illustrated  in  Figure  1 
above.  “Out-of-balance,  open,  or  otherwise  unsatisfied  control  loops 
initiate  and  maintain  downward  sloping  emotional  sensations.  Why  should 
this  be?  The  figure  suggests  a  naturally  selected  advantage  for  active 
animals,  or  human  beings.  In  nature  an  animal  has  to  exert  itself  physically 
to  satisfy  drives.  But  modem  knowledge  work  is  sedentary.  Successful 
drive  satisfaction  behaviors  in  modern  life  do  not  “trip  the  switches” 
indicating  drive  satisfaction  success  as  currently  configured  by  natural 
selection.  These  unsatisfied  control  loops  will  start  to  trigger  unpleasant 
emotions  whether  the  worker  is  successful  in  his  or  her  career  or  not. 

One  may  ask,  “What  is  actually  experienced  when  these  unfulfilled 
data  evaluation  functions  fire  off7”  The  answer  depends,  of  course,  on 
how  long  they  go  unfulfilled.  The  longer  they  go  unfulfilled,  the  more 
likely  a  person’s  awareness  of  them  will  eventually  make  an  impression 
on  his  or  her  inner  dialog.  But  prior  to  that  point,  here  are  some  anecdotal 
answers: 

•  “I  feel  a  little  off,  like  something  might  be  wrong,  but  I  just 
can’t  tell.”  (More  commonly  from  women  than  men.) 


Washington  Academy  of  Sciences 


65 


•  “Boy,  I  feel  restless.” 

•  “My  shoulders  are  so  stiff” 

•  A  person  might  become  surly  or  otherwise  difficult  to  deal  with 
for  no  apparent  reason  (more  commonly  men  than  women). 

What  is  it  about  knowledge  business  workplaces  that  create  these, 
and  more  serious,  emotional  and  behavioral  phenomena  in  human  beings? 


The  Modern  Workplace:  As  Alien  to  the  Human  Body  as  a  Flying 

Saucer 

Most  modern  living  and  work  environments  are  devoid  of  natural 
sights,  sounds,  colors,  smells,  solar  radiation,  fractal  patterns,  and  other 
clues  indicating  that  they  might  be  able  to  support  human  life  for  the 
duration  of  a  human  lifespan.  For  a  human  body  that  is  specifically  tuned 
by  millions  of  years  of  evolution  to  function  in  a  natural  environment,  the 
lack  of  these  stimuli  triggers  a  large  number  of  subliminal  danger  signals. 
The  emotions  these  danger  signals  generate  appear  to  follow  the  curve  in 
Figure  1,  as  indicated  by  numerous  coaching  sessions  we’ve  had  with 
employees  over  the  years. 

Add  to  this  lack  of  direct  sensory  stimulation  the  extended  periods 
of  time  that  sedentary  humans  perform  physically  inert  work  and  you  have 
a  recipe  for  numerous  subliminal  distractions  that  will  impact  the 
productivity  of  a  knowledge  workforce  in  increasingly  negative  ways.  The 
self-regulatory  mechanisms  of  the  body  assess  the  sedentary  state  as 
negative  and  act  to  push  the  human  animal  into  action.  The  restlessness 
many  people  experience  is  completely  real;  their  bodies  are  attempting  to 
get  them  up,  out  of  their  chairs,  and  into  fresh  air  and  sunlight.  Over 
extended  periods  of  time  these  unattended  subliminal  distractions  turn  into 
various  levels  of  anxiety,  depression,  frustration,  and  even  aggressive 
behavior. 

To  put  the  sharpest  of  points  on  this,  these  negative  responses  are 
exactly  what  the  human  body  is  built  to  do  when  conditions  are 
detrimental  to  personal  health  and  safety.  Before  the  human  inner  dialog 
came  into  existence,  emotional  programming  pushed  hominids  to  act  on 
behalf  of  their  own  welfare.  The  ability  to  think  using  symbols  to  model 
and  simulate  other  conditions  and  possible  solutions  was  not  needed  for 
the  human  body  to  meet  the  design  specifications  of  natural  selection.  The 
human  body  is  designed  to  have  a  negative  emotional  response  to 


Winter  2006 


66 


environments  which  make  living  difficult,  and  to  enjoy  environments 
which  are  rich  in  the  basic  resources  required  to  sustain  animal  existence. 
Or,  to  put  it  more  colloquially,  human  bodies  don’t  “know”  when  their 
knowledge-worker  owners  have  it  so  good. 

None  of  these  data  evaluation  routines  in  the  body  “understand” 
trading  time  for  money  as  a  drive  satisfaction  strategy.  The  smell  of 
money  doesn’t  really  elicit  a  gut-level  response,  unless  it’s  just  been  dug 
out  of  rich  soil.  Only  a  very  small  information  system  in  your  body 
understands  the  concept  of  gainful  employment,  and  it’s  not  big  enough  to 
change  thousands  of  small  responses  your  body  has  to  an  environment  ill- 
suited  to  human  living. 

However,  success  in  guiding  troubled  workers  through  emotional 
difficulties  suggests  that  the  system  that  understands  the  concept  of  gainful 
employment  can  also  be  taught  how  to  manage  those  thousands  of  small 
responses  to  avoid  negative  emotional  states. 


Managing  Subliminal  Emotional  Responses 
by  “Re-engineering”  Workplaces  and  Workflows 

The  most  empowering  way  to  deal  with  subliminal  emotional 
responses  is  to  systematically  retrain  a  worker’s  inner  dialog  to  deal  with 
them.  That,  however,  is  well  beyond  the  scope  of  this  article.  On  the  other 
hand,  we  have  employed  techniques  in  the  workplace  that  can  “fool”  the 
data  evaluation  functions  of  the  body  into  assessing  the  environment  as 
“good,  safe,  and  life-sustaining.” 

We  are  not  going  to  detail  a  wide  range  of  specifics.  There  are  many 
practical  reasons  why  workspaces  and  workflows  can  not  be  “completely” 
engineered  to  cater  to  more  primitive  animal  responses  in  human  beings. 
However,  with  a  knowledge  of  the  data  channels  emanating  from  the 
natural  environment,  we  can  perhaps  experiment  with  various  work 
environment  conditions  to  remediate  negative  worker  experiences. 
Likewise,  with  a  knowledge  of  the  time  dependence  of  these  subliminal 
emotional  experiences,  workflow  can  be  staged  differently  to  provide  both 
points  of  success  that  satisfy  them,  and  to  allow  time  for  exposure  to 
natural  environments  through  the  day. 


Washington  Academy  of  Sciences 


67 


Reconnecting,  Workers  to  the  Environment 's  Data  Outflow  Channels: 

Workspace  Design  Supporting  Positive  Workplace  Emotions 

From  an  architectural  standpoint,  many  builders  and  designers 
have  attempted  to  create  workspaces  which  emulate  certain  natural 
conditions.  These  include  the  extensive  use  of  windows  to  bring  in  natural 
light,  natural  colors  (various  greens  and  earth  tones),  and  plants.  Some 
locations  make  use  of  fountains  and  other  forms  of  moving  water  to  create 
a  little  light-play  and  to  produce  natural  sounds.  Let’s  look  at  these  in 
more  detail. 

The  use  of  sunlight  in  workplaces  makes  obvious  sense  for  a  lot  of 
reasons.  It  provides  many  key  forms  of  visual  stimulation.  However,  glass 
blocks  ultraviolet  radiation,  the  most  important  part  of  the  solar  spectrum 
for  creating  positive  moods  in  people.  People  need  to  get  outside  for 
proper  exposure  (cancer  risks  and  all)  to  the  sun. 

The  color  palette  for  workplaces  is  probably  important,  but  the  key 
is  the  patterns  with  which  they  are  applied.  Common  experience  indicates 
that  large  areas  of  unchanging  visual  stimulation  are  uninteresting.  Vast 
sheets  of  light  green  or  neutral  beige  are  not  as  soothing  as  natural  colors 
applied  in  some  form  of  natural  looking  fractal  pattern.  There  are  other 
visual  stimuli  that  can  be  built  into  a  workspace  that  evoke  specific  kinds 
of  natural  settings,  which  can  include  everything  from  large  format 
photographs  and  murals  to  large  portions  of  the  building  built  with  glass 
that  actually  face  an  attractive  natural  setting,  like  nearby  woods, 
farmland,  or  parks. 

Our  model  suggests  that  plants  are  most  likely  a  positive  addition 
to  workplace  decor.  In  larger  spaces  larger  plants  and  small  trees  strongly 
evoke  a  more  life-friendly  sense  of  the  environment.  They  can  also  add 
important  biochemical  molecules  to  the  air  to  improve  subliminal 
emotional  states.  The  olfactory  sensory  system  provides  a  direct, 
biochemically  driven  informational  link  between  the  human  and  external 
conditions  (Amen  19984). 

Decorative  water  displays  obviously  have  to  be  proportionate  to 
the  area  within  which  they  operate.  However,  small  table-top  “meditation” 
fountains  have  become  popular  in  workplaces,  and  perhaps  these  sounds 
can  be  integrated  into  larger,  more  complicated  spaces. 


Winter  2006 


: 


68 


Other  acoustic  effects  could  be  built  into  a  workplace  using  various 
mechanical  or  electronic  technologies.  Wind  sounds  could  be  produced, 
rustling  leaves,  or  even  the  noises  small  animals  make  under  the  brush. 

Our  model  also  suggests  that  the  least  attended-to  set  of  stimuli  are 
the  airborne  ones.  Office  air  is  typically  highly  filtered  and  de-humidified. 
Worse,  the  latent  scents  from  various  cleaners  and  the  outgassing  of  many 
synthetic  products  fill  the  air  with  what  the  brain  perceives  as  dangerous 
smelling  chemicals.  Remember,  the  body  is  built  to  seek  out  an 
environment  containing  optimal  biochemical  resources  without  needing 
to  think  about  it.  Office  air  is  full  of  strong,  unsubtle  clues  that  the  office 
is  NOT  an  optimal  source  of  biochemical  resources.  It  smells  like  a  vat  of 
poisons  to  the  brain,  and  that  creates  a  lot  of  subliminal  anxiety. 

Obviously,  things  need  to  be  kept  clean.  We  use  advanced 
technology  with  its  many,  unintended  by-products.  That  cannot  be 
avoided.  Can  an  office  building’s  air  be  re-filtered  and  stocked  with  trace 
chemicals  of  a  natural  kind9 


If  it  can’t,  then  the  solution,  again,  is  to  get  the  worker  outside  for 
some  part  of  the  day.  If  a  human  body  can’t  be  in  fresh,  naturally 
perfumed  air  all  of  the  time,  at  least  it  can  be  exposed  to  enough  good  air 
(current  pollution  problems  aside)  to  create  the  subliminal  “belief’  that 
good  air  is  quickly  and  easily  available.  Of  course,  getting  outside  also 
exposes  the  human  body  to  all  of  the  other  stimuli  for  which  it  is  so  well 
tuned  to  collect,  assess,  and  appreciate.  In  our  many  years  managing,  we 
often  have  heard  people  comment  on  their  refreshed  states  and  improved 
moods  after  a  walk  outside. 

In  fact,  the  best  of  workplace  designs  probably  doesn’t  involve 
interior  design  and  architecture.  Our  model  of  the  human  being  as  a  node 
in  the  environment’s  network  of  information  resources  suggests  that  the 
best  of  workplace  designs  may  be  to  build  extensive  garden  parks  around 
the  campus  of  a  major  business  district,  and  fill  them  with  as  many 
naturally  stimulating  features  as  they  can  contain. 


Synchronizing  with  Workers  ’  Internal  Clocks:  Workflow  Design 
Supporting  Positive  Workplace  Emotions 

Apart  from  migrations  and  other  long  travels  in  search  for  food  and 
water,  few  behaviors  in  the  entire  animal  kingdom  last  more  than  a  few 
minutes.  Virtually  all  of  the  self-regulating  controls  of  the  human  body  are 
built  on  these  same  behavioral  time  scales. 


Washington  Academy  of  Sciences 


69 


Workplace  reality,  however,  is  that  nothing  takes  a  few  minutes. 
In  the  very  least,  most  people  are  expected  to  sit  dutifully  at  their  desks  for 
every  minute  they  expect  to  get  paid.  People  don’t  sit  there  like  that  for 
many  of  the  reasons  we’ve  been  discussing,  but  the  person  signing  the 
paychecks  would  prefer  things  that  way. 

For  many  knowledge  workers,  though,  work  projects  run  for 
months  and  sometimes  years  with  virtually  no  short  timescale  successes  to 
satisfy  a  worker’s  subliminal  control  loops.  Very  few  things  occur  “by 
human  hand”  in  business  in  a  matter  of  minutes.  As  it  turns  out,  our 
problem  solving  capabilities  give  us  enough  logic  to  keep  us  plugging 
away  day  after  day,  but  that  merely  locks  the  animal  part  of  human 
experience  in  a  cage  that  is  more  than  merely  psychological. 

Things,  of  course,  have  to  get  done,  and  they  have  to  get  done  in 
the  workplace.  They  have  to  be  done  in  environments  and  on  timescales 
that  strongly  conflict  with  the  naturally  selected  characteristics  of  the 
typical  human  being.  How  does  a  manager  use  this  knowledge  to  avoid 
loss  of  productivity  in  his  or  her  shop? 

For  major  projects,  built  on  tasks,  goals,  objectives,  and  so  forth, 
the  current  trend  is  to  design  the  workflow  against  cost  barriers  and  time 
deadlines.  That’s  probably  not  changeable  for  practical  business 
performance  reasons.  However,  the  granularity  of  the  workflow  design 
CAN  be  optimized  for  human  mental  health  and  best  performance  levels. 
How  might  a  different  granularity  of  task  be  used  to  re-design  work? 

In  nature,  animal  behavior  is  dominated  by  three  types  of  activity 
with  fairly  immediate  rewards  if  they  succeed: 

1 .  An  animal  searches  for  food  and  water,  and  if  it  succeeds  it 
obtains  enjoyment. 

2.  An  animal  engages  in  a  mating  activity. 

3.  An  animal  flees  a  predator,  and  if  it  succeeds  it  enjoys  a  sense 
of  relief. 

For  most  of  human  history,  success  has  been  immediately  followed 
by  reward.  By  business  standards,  each  of  these  three  activities  is  a  small 
scale  effort.  Obviously  a  business  can’t  drop  a  cupcake  on  everyone  each 
time  they  successfully  complete  15  minutes  of  work,  nor  can  you  bring  in 
the  dancing  girls  or  Chip’n’Dales.  And  merely  getting  the  boss  (the 
predator)  off  your  back  for  a  short  period  of  time  isn’t  adequate 
compensation  either. 


Winter  2006 


70 


The  point  is  that  there  is  a  granularity  of  effort  and  a  system  of 
rewards  that  can  be  found  to  optimize  human  performance  and  maintain 
good  mental  health  in  the  workforce.  It  takes  a  committed  and  engaged 
management  team  to  experiment  and  find  it.  The  EvolvingSuccess  team 
has  been  very  successful  in  identifying  the  appropriate  granularity  of  task 
to  get  the  most  productivity  out  of  project  participants  while  keeping  them 
happy  on  the  job.  Interestingly,  the  more  difficult  and  esoteric  the  project, 
the  more  successful  our  productivity  techniques  tend  to  be. 

On  the  basis  of  our  ev  psych  model  of  human  thought  and 
behavior,  we  have  composed  an  idealized  daily  workflow  schedule 
designed  to  keep  tasks  short,  get  a  modicum  of  exercise,  as  well  as  get 
exposure  to  natural  settings  should  they  be  close  to  the  work  place.  Again, 
this  is  highly  idealized.  Our  work  teams  have  only  approximated  it  in 
practice.  However,  our  experience  in  staging  work  with  a  time-granularity 
that  aims  at  this  ideal  has  been  very  effective. 

Table  1:  Daily  Schedule  to  Optimize  Knowledge  Worker  Performance 
and  Maintain  Good  Mental  Health 

1.  Take  the  low-stress  route  to  work,  regardless  of  how  long  it 
takes. 

2.  Once  you  get  to  work,  take  care  of  2  to  4  tasks  taking  an 
average  30  minutes  each. 

3.  Get  outside  and  get  30  to  60  minutes  of  exercise  at  the  highest 
level  of  exertion  you  can  reasonably  work  up. 

4.  Get  something  to  eat,  but  just  enough  to  settle  the  hunger. 

5.  Take  care  of  another  2  to  4  tasks  averaging  30  minutes  each. 

6.  Get  outside  for  a  30  minute  walk. 

7.  Have  a  healthy,  reasonably  sized  lunch. 

8.  Take  care  of  another  2  to  4  tasks  averaging  30  minutes  each. 

9.  Get  outside  and  get  30  to  60  minutes  of  exercise  at  the  highest 
level  of  exertion  you  can  reasonably  work  up. 

10.  Get  another  light  snack. 

11.  Take  care  of  another  2  to  4  tasks  averaging  30  minutes  each. 

12.  Take  the  low-stress  route  back  home,  and  don’t  be  in  a  hurry. 

13.  Get  in  another  30  minute  walk  before  supper. 

This  schedule  is  designed  to  fool  the  body’s  various  environment 
assessment  systems  and  self-regulating  controls  into  assessing  the 
situation  as  good.  The  order  of  events  in  this  schedule  presumes  the  “ short 


Washington  Academy  of  Sciences 


71 


effort,  immediate  gratification”  model  of  animal  behavior  in  natural 
settings.  By  business  standards  this  is  an  extravagant  use  of  time  during 
work  hours.  In  our  experience,  where  this  ideal  has  been  attempted  but  not 
met,  this  schedule  avoids  burnout,  maintains  mental  freshness,  and 
engenders  productivity  levels  that  pay  for  themselves. 


Conclusions 

The  human  body  is  clearly  pre-tuned  to  receive  specific  data  inputs 
from  the  environment.  When  those  inputs  are  not  received,  the  body 
generates  motivations  to  get  out  of  the  resource-poor  or  dangerous 
environment,  and  move  into  a  life-friendly  environment.  The  longer  it 
takes  the  individual  to  succeed  in  moving  to  a  better  environment,  the 
more  deeply  negative  are  the  emotions  that  that  individual  experiences. 

The  body  is  also  extremely  well  tuned  to  the  state  of  its  internal 
conditions.  Many  people  often  experience  urges  to  get  a  little  food,  water, 
or  other  relief  without  a  conscious  thought  to  do  so.  If  the  body  is 
experiencing  conditions  outside  of  its  “safe  operating  specifications,” 
negative  emotions  crop  up  to  motivate  the  individual  to  take  care  of  the 
need.  The  longer  it  takes  to  satisfy  the  need,  the  more  extreme  the 
emotional  sensation  becomes. 

Modern  living  and  working  conditions  place  barriers  between  the 
human  being  and  the  data  sources  his  or  her  body  expects  to  detect.  Long¬ 
term  disconnection  from  these  sources  appears  to  be  creating  chronic 
mental  health  problems  in  the  American  workforce.  This,  of  course,  has 
various  impacts  on  business  productivity. 

We  have  suggested  principles  by  which  to  guide  a  manager  in  the 
re-engineering  of  workspaces,  including  the  landscape  architecture 
surrounding  work  locations.  We  have  also  supplied  a  rationale  to  guide  in 
the  restructuring  of  long  term  work  efforts.  Lastly,  we  provided  an 
idealized  work  schedule  aimed  at  structuring  time  utilization  on  the  day- 
to-day  level.  This  schedule  maps  to  the  basic  psychology  built  into  human 
beings  by  natural  selection  to  optimize  their  mental  health  and  business 
productivity. 


Notes 

1  Thomas  Meylan,  “Using  Evolutionary  Psychology  and  Information 
Systems  Engineering  to  Understand  Workplace  Patterns  of  Thought  and 


Winter  2006 


72 


Behavior:  An  Empirical  Model  of  Human  Information  Processing,”  I 
Autumn,  2005,  Journal  of  the  Washington  Academy  of  Sciences. 

2  Martin  E.P  Seligman,  Ph  D.,  The  Optimistic  Child,  Harper  Perennial,  pp. 
37-42,  1995. 

3  Robert  Wright,  The  Moral  Animal,  Pantheon,  1994. 

4  Daniel  G.  Amen,  M.D.,  Change  Your  Brain,  Change  Your  Life,  Three 
Rivers  Press,  pp.  37-43,  1998. 


Washington  Academy  of  Sciences 


73 


BOOK  REVIEW 


The  Best  American  Science  Writing  2006 ,  edited  by  Atul  Gawande 
(Harper  Collins,  Publishers,  2006,  362  pp.) 


IN  SPITE  OF  THE  TITLE,  this  almost  certainly  is  not.  “The  best 
science  writing”  of  any  year  should  surely  be  factually  sound,  balanced, 
informative,  and  enjoyable  to  scientists  and  non-scientists  alike.  The  21 
articles  reprinted  here  for  the  most  part  meet  those  criteria.  Beyond  that, 
however,  surely  the  best  science  writing  would  present  material  that  has 
not  already  been  repeatedly  reported,  or  provide  new  perspectives  on 
familiar  material,  or  introduce  provocative  and  fruitful  interpretations  of 
scientific  findings.  By  and  large,  these  articles  do  not  meet  that  challenge. 
Little  of  the  material  here  is  new  even  to  general  readers. 

One  might  suspect  just  from  the  list  of  original  sources  of  these 
articles  that  they  might  not  constitute  “the  best”  science  writing  of  the 
year.  One  would  not,  it  is  true,  expect  the  best  to  be  drawn  from  premier 
scientific  journals  such  as  Science  or  the  New  England  Journal  of 
Medicine,  where  original  research  results  are  presented  by  scientists  for 
other  frontline  scientists,  in  what  is  to  non-scientists  usually  mind- 
numbing  detail.  But  of  the  21  articles  collected  here,  no  fewer  than  6  are 
drawn  from  The  New  Yorker  (is  it  a  coincidence  that  the  editor  is  a  staff 
writer  for  that  magazine  as  well  as  a  surgeon  at  Brigham  and  Women’s 
Hospital?).  Four  more  come  from  Harper's  or  The  Atlantic  Monthly. 
Another  4  first  appeared  in  The  New  York  Times  or  its  Sunday  Magazine. 
Only  two  originated  in  first  tier  scientific  magazines  directed  at  the 
general  educated  public.  Sigma  Xi’s  American  Scientist  and  Scientific 
American.  Another  two  were  from  Discover  and  one  from  Wired.  There 
were  none  from  IEEE’s  Spectrum ,  none  from  Mind  or  MIT’s  Technology > 
Review,  all  of  which  feature  first-class  science  writers.  That  does  not 
mean,  of  course,  that  the  specialized  science-oriented  magazines  have  a 
lock  on  the  best  science  writing;  but  it  makes  one  wonder  about  the  scope 
of  the  Editor’s  reading. 

The  ten  articles  drawn  from  the  literary  magazines,  by  the  way, 
reflect  the  characteristics  common  to  articles  in  those  publications — they 
are  so  excessively  long  as  to  try  many  readers’  patience  and  needlessly 
confuse  the  information  they  are  meant  to  convey. 


Winter  2006 


74 


But  if  one  sets  aside  reservations  about  “the  best”  and  is  willing  to 
merely  enjoy  “very  good”  science  writing,  this  is  a  book  to  appeal  to  most 
readers  with  an  interest  in  a  wide  range  of  scientific  activities,  the  people 
who  carry  them  out,  and  the  implications  for  the  rest  of  us. 

The  topics  range  from  the  ordinary  and  personal  (Why  are  chess¬ 
playing  computers  getting  better  at  it?  What’s  the  best  strategy  for  a  music 
lover  who  is  gradually  losing  his  hearing?  Is  obesity  as  much  of  a  threat  to 
health  as  we  are  being  led  to  believe?)  to  the  sublime  (  What  is  the  nature 
of  time?  And  is  the  idea  of  God  an  evolutionary  artifact,  or  perhaps  an 
accidental  outcome  of  the  way  we  think  about  ourselves?).  The  rehashing 
of  the  debate  about  the  origin  of  progressive  supranuclear  palsy  on  Guam, 
or  the  closer-to-home  but  even  more  bitterly  conflicted  debate  about 
childhood  vaccinations  as  a  possible  cause  of  autism  seem  stale  in  terms 
of  2006,  but  the  implications  of  a  possible  avian  flu  epidemic  are,  to  use  a 
phrase  from  a  TV  series,  “ripped  from  the  headlines.” 

It  would,  in  short,  be  a  very  demanding  reader  who  will  not  find 
something  to  enjoy  and  something  to  ponder  in  this  collection. 

— Vary  Coates, 
vcoates@mac.  com 


Washington  Academy  of  Sciences 


75 


NEWS  OF  MEMBERS,  FELLOWS,  AND  AFFILIATED 

SOCIETIES 


Daryl  Chubin,  WAS  Fellow,  has  been  selected  a  Sigma  Xi  (Scientific 
Research  Society)  Distinguished  Lecturer,  2007-2009.  Daryl  recently 
published,  with  S.M.  Malcolm,  “The  New  Backlash  on  Campus,”  in 
College  and  University >  Journal ,  Fall  2006.  In  September  he  spoke  at  the 
Math  Alignment  and  Transition  Conference  at  Southern  Connecticut  State 
University,  on  “Why  Take  More  Math?  A  National  Policy  Response.” 
Daryl  has  received  an  NSF  grant  on  Building  Community  Resources  for 
the  NSF  Graduate  Teaching  Fellows  in  the  K-12  Education  Program. 

Mark  Holland,  WAS  Vice  President  for  Affiliated  Societies,  received  a 
Faculty  Appreciation  Award  from  the  Alumni  Association  of  Salisbury 
University.  The  award  is  given  irregularly  to  faculty  members 
spontaneously  nominated  and  elected  by  alumni. 

Alain  Touwaide,  President  Elect  of  WAS,  has  recently  published  14 
entries  on  the  history  of  ancient  medicine  and  pharmacology  in  Medieval 
Science ,  Technology,  and  Medicine,  An  Encyclopedia,  ed.  by  T.  Glick,  S. 
Livesey,  &  F.  Wallis,  2005;  as  well  as  entries  on  botany  and  horticulture 
in  J.W.  Meri,  ed..  Medieval  Islamic  Civilization.  He  also  wrote  on 
“Byzantine  Hospitals  Manuals  as  a  Source  for  the  Study  of  Therapeutics,” 
in  B  Bowers,  ed..  The  Medieval  Hospital  and  Medical  Practice  (Avista 
Studies  in  the  History  of  Medieval  Technology,  Science  and  Art).  In 
November  Alain  presented  three  Lansdowne  Lectures  at  the  University  of 
Victoria  in  Canada;  in  December  he  gave  the  inaugural  lecture  at  the  5th 
Conference  of  the  Pan-Hellenic  Society  for  History  of  Medicine  in 
Thessaloniki,  Greece. 

Jodi  Wesemann  (WAS  Board  member-at-large)  has  been  elected  president 
of  the  DC  Metropolitan  Area  chapter  of  American  Women  in  Science,  and 
Ester  Sztein  (also  a  WAS  member)  is  the  new  Vice  President  for 
Programs. 

Vary  Coates,  editor  of  the  WAS  Journal,  has  been  chosen  as  a  Fellow  of 
the  American  Association  for  the  Advancement  of  Science,  “in 


Winter  2006 


76 


recognition  of  her  early  and  continuing  support  of  the  concept  of 
technology  assessment.” 

Dr.  Edward  O.  Haenni,  a  longtime  member  of  the  Academy,  died  on 
August  28,  2006,  in  Sanibel,  Florida,  at  age  99. 

The  DC  Council  of  Engineering  and  Architectural  Societies  (DCCEAS) 
will  hold  an  Engineers  Week  Lunch  on  February  21,  at  the  Pier  7 
Restaurant,  and  an  Awards  Banquet  on  February  24  at  the  Crowne  Plaza 
Hotel  in  Silver  Spring.  Sajjad  Durrani,  a  WAS  Fellow,  is  President  of  the 
DC  Council.  DCCEAS  promotes  science  and  technology  in  high  schools 
and  colleges,  and  sponsors  a  Student  Paper  Competition  in  local 
universities.  DCCEAS  has  as  affiliates  37  local  chapters  of  engineering 
and  architectural  societies. 

The  Philosophical  Society  of  Washington,  the  area’s  oldest  scientific 
society,  has  scheduled  eight  public  lectures  by  prominent  scientists  from 
February  2  to  May  11;  topics  include  the  scourge  of  malaria,  astrobiology, 
the  global  positioning  system,  obesity,  chemistry  against  crime,  Benjamin 
Franklin’s  experiments,  technologies  of  the  future,  and  “the  dawn  of  the 
universe.”  All  lectures  are  free  and  are  held  at  the  Powell  Auditorium, 
2170  Florida  Ave,  NW  (Dupont  Circle  Metro  stop).  See  www.philsoc.orR 
for  the  exact  schedule. 

The  World  Future  Society’s  annual  meeting  will  be  held  July  29-31  in 
Minneapolis,  and  will  offer  multiple  sessions  on  technology,  health, 
governance,  education,  values,  and  social  trends.  It  will  be  preceded  by 
short  courses  on  futures  research  techniques  and  followed  by  a 
Professional  Members’  Forum  on  August  1.  For  more  information  see  the 
web  site,  www.wfs.oru. 

The  Washington  area  sections  of  the  IEEE  and  the  IEEE’s 
Communications  Society  will  host  the  2007  Global  Communications 
Conference  (N ov. 25-Dec.  1,  at  the  Washington  Hilton  Hotel).  The  web  site 
already  has  information  about  the  program,  exhibits,  and  activities: 
http  ://www.comsoc.  oru/confs/ul obecom/2007/i ndex , html .  Jerry  Gibbon,  a 
past  president  of  WAS,  is  the  General  Chairman  of  the  conference. 

The  Chesapeake  Section  of  the  American  Association  of  Physics  Teachers 
held  its  annual  meeting  October  27-28  at  James  Madison  University,  and 
elected  the  following  officers:  President,  Brett  Taylor  (Radford 


Washington  Academy  of  Sciences 


77 


University);  Vice  President,  Deonna  Woolard  (Randolph-Macon  College); 
Secretary,  David  Wright  (Tidewater  Community  College);  Treasurer,  Eric 
Kearsley  (High  Point  High  School,  Beltsville);  Vice-President  for 
Communications,  Rhett  Herman  (Radford  University);  Section 
Representative,  David  Wright  (Tidewater  Community  College).  In 
addition,  the  following  prizes  were  awarded: 

Frank  R.  Haig  Prize  (best  paper  from  a  4  year  college):  a  tie:  Joseph  W. 
Rudmin  (James  Madison  University)  and  Brett  Taylor  (Radford 
University) 

James  Newman  Prize  (best  paper  from  a  high  school):  a  tie:  Saharsha 
Nambiar  (Millbrook  High  School)  and  Michael  Pagel  (Collegiate  School) 
David  Wright  Prize  (best  paper  from  a  two  year  college):  James 
O’Connell  (Frederick  Community  College) 


Winter  2006 


AFFILIATED  INSTITUTIONS 


The  National  Institute  for  Standards  and  Technology 
Meadowlark  Botanical  Gardens 
The  John  W.  Kluge  Center  of  the  Library  of  Congress 
Potomac  Overlook  Regional  Park 


Washington  Academy  of  Sciences 


79 


WASHINGTON  ACADEMY  OF  SCIENCES 
MEMBERSHIP  DIRECTORY  2006 

M=Member;  F=FeIlow;  LF=Life  Fellow;  LM=Life  Member;  EM=Emeritus 
Member;  EF=Emeritus  Fellow 

ABDULLA YER,  KENZHE  (M) 

ABDULNUR,  SUHEIL  F.  (Dr.)  5715  Glenwood  Road,  Bethesda  MD 
20817(F) 

ABELSON,  PHILIP  H  10528  Georgia  Ave.,  Silver  Spring  MD  20902  (F) 
ALLEN,  J.  FRANCES  (Dr.)  The  Southerlands,  Apt  213,  600  Mount  View 
Street,  Front  Royal  VA  22630  (EF) 

ANASTAS,  PAUL  T  (Mr.)  217  E.  Bellefonte  AveSt,  Alexandria  VA 
22301-1351  (M) 

APPETITI,  EMANUELA  PO  Box  25805,  Washington  DC  20027  (M) 
ARSEM,  COLLINS  (Mr.)  3144  Gracefield  Rd  Apt  117,  Silver  spring  MD 
20904-5878  (EM) 

ARVESON,  PAUL  T.  (Mr.)  6902  Breezewood  Terrace,  Rockville  MD 
20852-4324  (F) 

BAILEY,  R.  CLIFTON  (Dr.)  6507  Divine  Street,  Mclean  VA  22101-4620 
(LF) 

BARBOUR,  LARRY  L.  (Mr.)  Pequest  Valley  Farm,  585  Townsbury 
Road,  Great  Meadows  NJ  07838  (M) 

BARWICK,  W.  ALLEN  (Dr.)  13620  Maidstone  Lane,  Potomac  MD 
20854-1008  (F) 

BASILI,  VICTOR  R.  (Dr.)  A.V.  Williams  Building,  University  of 
Maryland,  College  Park  MD  20742  (F) 

BEACH,  LOUIS  A.  (Dr.)  1200  Waynewood  Blvd.,  Alexandria  VA  22308- 
1842  (EF) 

BEAM,  WALTER  R.  (Dr.)  4804  Wellington  Farms  Drive,  Chester  VA 
23831 (F) 

BEMENT,  ARDEN  (Dr.)  National  Science  Foundation,  4201  Wilson 
Boulevard,  Arlington,  Virginia  22230  (F) 

BERG,  RICHARD  E.  (Dr.)  8308  Quill  Point  Dr.,  Bowie  MD  20720  (F) 
BERGMANN,  OTTO  (Dr.)  1039  South  19th  St.,  Arlington  VA  22202- 
1611  (EF) 

BERMAN,  BARRY  L.  (Prof.)  Department  of  Physics,  George 
Washington  University,  Washington  DC  (M) 

BERRY,  JESSE  F.  (Mr.)  2601  Oakenshield  Drive,  Rockville  MD  20854 
(M) 

BIBERMAN,  LUCIEN  M.  (Mr.)  3731  Glen  Eagles  Drive,  Silver  Spring 
MD  20906  (F) 


Winter  2006 


80 


BIONDO,  SAMUEL  J.  (Dr.)  10144  Nightingale  St.,  Gaithersburg  MD 
20882  (F) 

BLUNT,  ROBERT  F.  (Dr.)  541 1  Moorland  Lane,  Bethesda  MD  20814- 
13335  (F) 

BODSON,  DENNIS  (Dr.)  233  N.  Columbus  Street,  Arlington  VA  22203 

(F) 

BOGNER,  MARILYN  SUE  (Dr.)  9322  Friars  Road,  Bethesda  MD  20817- 
2308 (LF) 

BOYER,  WILLIAM  (Mr.)  3725  Alton  PI,  N.W.,  Washington  DC  20016 
(M) 

BRANCATO,  EMANUEL  L.  (Dr  )  7370  Hallmark  Road,  Clarksville  MD 
21029  (EF) 

BRIMMER,  ANDREW  F.  (Dr.)  Suite  302,  4400  MacArthur  Blvd.,  NW, 
Washington  DC  20007  (F) 

BRISKMAN,  ROBERT  D.  (Mr.)  61  Valerian  Court,  North  Bethesda  MD 
20852 (F) 

BROWN,  ELISE  A  B  (Dr.)  6811  Nesbitt  Place,  Mclean  VA  22101-2133 
(LF) 

BURNS,  EDGAR  JOHN  415  Lincoln  Ave,  Avon  NH  07717  (EF) 
BUTTERMORE,  DONALD  O.  (Mr.)  34  West  Berkeley  St,  Uniontown 
PA  15401-4241  (LF) 

CAMPBELL,  FRANCIS  J.  (Mr.)  Apt  113,  7406  Spring  Village  Dr, 
Springfield  VA  22150  (EF) 

CERF,  VINTON  G.  (Dr.)  1435  Woodhurst  Blvd.,  McLean  VA  22102- 
2234  (F) 

CHANDLER,  PH  D.,  JERRY  837  Canal  Drive,  Mclean  VA  22102-1407 

(F) 

CHERESHNEV,  VALERIY  A  (Mr.)  91,  Pervomayskya,  Yekaterinburg 
6202 19,  Russia  (F) 

CHRISTMAN,  GERARD  (Mr.)  1 16  Tanley  Rd,  silver  Spring  MD  20904 
(M) 

CHUBIN,  DARYL  E.  (Dr.)  1200  New  York  Ave,  NW,  Washington  DC 
20005 (F) 

CLINE,  THOMAS  LYTTON  (Dr.)  13708  Sherwood  Forest  Drive,  Silver 
Spring  MD  20904  (F) 

CLORE,  GIDEON  MARIUS  (Dr.)  Lab  of  Chemical  Physics,  Bldg  5,  Rm 
B 1-301,  NIDOK,  National  Institutes  of  Health,  Bethesda  MD  20892  (F) 
COATES,  JOSEPH  F.  (Mr.)  Apt.  401  Tilden  Gardens,  3930  Connecticut 
Ave.  NW,  Washington  DC  20008  (F) 

COATES,  VARY  T.  (Dr.)  Apt.  401  Tilden  Gardens,  3930  Connecticut 
Ave.  NW,  Washington  DC  20008  (F) 


Washington  Academy  of  Sciences 


81 


COFFEY,  TIMOTHY  P.  (Dr.)  976  Spencer  Rd.,  McLean  VA  22102  (F) 
COHEN,  MICHAEL  P.  (Dr.)  1615  Q.  St  NW  T-l,  Washington  DC 
20009-6310  (LF) 

COHEN,  ROBERTA  (Ms.)  The  Brookings  Institution,  1775 
Massachusetts  Avenue  NW,  Washington  DC  20036  (F) 

COLE,  JAMES  H.  (Mr.)  9404  Fairpine  Lane,  Great  Falls  VA  22066  (M) 
CONLEY,  ROBERT  E.  (Dr.)  Conley  &  Associates,  Inc.,  9001  Saunders 
Lane,  Bethesda  MD  208 1 7  (F) 

COOPER,  KENNETH  W.  (Dr.)  4497  Picacho  Drive,  Riverside  CA 
92507-4873  (EF) 

COSTRELL,  LOUIS  (Mr.)  Apartment  640,  1801  East  Jefferson  St, 
Rockville  MD  20852  (EF) 

CREVELING,  CYRUS  R.  (Dr.)  4516  Amherst  Lane,  Bethesda  MD  20814 

(F) 

CURRIE,  S.J.,  C.  L.  (Rev.)  Pres.,  Assn  of  Jesuit,  Colleges  &  Universities, 
One  Dupont  Circle  NW  #405,  Washington  DC  20036  (F) 

DAVIS,  DANIEL  (Dr.)  6324  N.  24th  St.,  Arlington  VA  22207  (M) 
DAVIS,  ROBERT  E.  (Dr.)  1793  Rochester  Street,  Crofton  MD  21 1 14  (F) 
DEDRICK,  ROBERT  L.  (Dr.)  1633  Warner  Avenue,  Mclean  VA  22101 
(EF) 

DEAN,  DONNA  (Dr.)  29  Eldwick  Court,  Potomac  MD  20854-2027  (F) 
DENG,  Francis  M.  (Dr.)  (F) 

DEUTSCH,  STANLEY  (Dr.)  7109  Laverock  Lane,  Bethesda  MD  20817 
(EF) 

DOCTOR,  NORMAN  (Mr.)  6  Tegner  Court,  Rockville  MD  20850  (EF) 
DONALDSON,  EVA  G.  (Ms.)  3941  Ames  St  Ne,  Washington  DC  20019 
(F) 

DONALDSON,  JOHANNA  B.  (Mrs.)  3020  North  Edison  Street, 

Arlington  VA  22207  (EF) 

DUBEY,  SATYA  D.  (Dr.)  7712  Groton  Road,  West  Bethesda  MD  20817 
(EF) 

DUNCOMBE,  RAYNOR  L.  (Dr )  1804  Vance  Circle,  Austin  TX  78701 

(F) 

DUPONT,  JOHN  E.  (Mr.)  P.O.  Box  358,  Newtown  Square  PA  19073  (F) 
DURRANI,  SAJ  (Dr.)  17513  Lafayette  Dr,  OLNEY  MD  20832  (EF) 
EDINGER,  STANLEY  EVAN  (Dr.)  Apt  #1016,  5801  Nicholson  Lane, 
North  Bethesda  MD  20852  (F) 

EISNER,  MILTON  PHILIP  (Dr.)  1565  Hane  Street,  Mclean  VA  22101- 
4439 (F) 

EL  KHADEM,  HASSAN  (Dr.)  Dept,  of  Chemistry,  American  University, 
Washington  DC  20016-8014  (EF) 


Winter  2006 


82 


ENDO,  BURTON  Y.  (Dr.)  1010  Jigger  Court,  Annapolis  MD  21401-6886 
(EF) 

ETTER,  PAUL  C.  (Mr.)  16609  Bethayres  Road,  Rockville  MD  20855- 
2043  (F) 

FAULKNER,  JOSEPH  A  (Mr  )  2  Bay  Drive,  Lewes  DE  19958  (F) 
FAUST,  WILLIAM  R.  (Dr.)  2940  Karen  Dr,  Chesapeake  Beach  MD 
20732-3845  (F) 

FAY,  ROBERT  E.  (Dr.)  7252  Greentree  Rd,  Bethesda  MD  20817  (F) 
FINKELSTEIN,  ROBERT  (Dr.)  1 1424  Palatine  Drive,  Potomac  MD 
20854-1451  (M) 

FLOURNOY,  NANCY  (Dr.)  3 1 05  Trailside  Dr.,  Columbia  MO  65203- 
5817(F) 

FOCKLER,  HERBERT  H.  (Mr.)  10710  Lorain  Avenue,  Silver  Spring  MD 
20901  (EF) 

FORZIATI,  ALPHONSE  F.  (Dr.)  65  Heritage  Dr,  Unit  6,  Cleveland  GA 
30528 (EF) 

FRANKLIN,  JUDE  E.  (Dr.)  7616  Carteret  Road,  Bethesda  MD  20817- 
2021  (F) 

FREEMAN,  ERNEST  R.  (Mr.)  5357  Strathmore  Avenue,  Kensington  MD 
20895-1160  (EF) 

FREEMAN,  HARVEY  1 1  South  Eutaw,  Apt  1302,  Baltimore  MD  21201 

(F) 

GAUNAURD,  GUILLERMO  C.  (Dr.)  4807  Macon  Road,  Rockville  MD 
20852-2348  (F) 

GEBBIE,  KATHARINE  B.  (Dr.)  Physics  Laboratory,  National  Institute  of 
Standards  and  Technology,  100  Bureau  Drive,  MS  8400,  Gaithersburg 
MD  20899-8400  (F) 

GIBBON,  JOROME  (Mr.)  311  Pennsylvania  Avenue,  Falls  Church  VA 
22046  (F) 

GIBBONS,  JOHN  H.  (Dr.)  Resource  Strategies,  P.0  Box  379,  The  Plains 
VA  20198  (F) 

GIBSON,  DOUGLAS  963 1  Boyett  Ct,  Fairfax  VA  22032  (M) 

GIFFORD,  PROSSER  (Dr.)  540  N.  St.  SW59  Penzance  Rd,  Woods  Hole 
MA  02543  (F) 

GLASER,  HAROLD  (Dr.)  1902  Berryman  Street,  Berkeley  CA  94709- 
1919  (EF) 

GLAZE,  JOHN  (Mr.)  658  E  St.,  S.E.,  Washington  DC  20003  (F) 
GLUCKMAN,  ALBERT  G.  (Mr.)  Institute  for  Physical  Science  and 
Technology,  University  of  Maryland,  College  Park,  MD  20742  (EF) 
GOOD  ALL,  JANE  (Dr.)  The  Jane  Goodall  Institute,  4245  Fairfax  Dr  Ste 
600,  Arlington  VA  22203-1698  (F) 


Washington  Academy  of  Sciences 


83 


GORDON,  NANCY  M  Associate  Director  for  Demographic  Programs, 
US  Census  Bureau,  Washington  DC  20233  (F) 

GOULD,  RICHARD  G.  Telecommunications  Systems,  3643  Upton 
Street,  NW,  Washington  DC  20008  (F) 

GRAY,  JOHN  E.  (Mr.)  PO  Box  489,  Dahlgren  VA  22448-0489  (M) 
GRAY,  MARY  (Professor)  Department  of  Mathematics,  Statistics,  and 
Computer  Science,  American  University,  4400  Massachusetts  Avenue 
NW,  Washington  DC  20016  (F) 

GREENOUGH,  M.  L.  (Mr.)  Greenough  Data  Assoc.,  616  Aster  Blvd., 
Rockville  MD  20850  (EF) 

GUDE,  GILBERT  (The  Honorable)  541 1  Duvall  Drive,  Bethesda  MD 
20816-1871  (F) 

GUPTA,  PRADEEP  KUMAR  (Dr.)  8301  Arlington  Blvd.  #405,  Fairfax 
VA  22182  (F) 

GUTERMUTH,  PAUL-GEORG  (Dr.)  IM  Wingert  28,  53604  Bad  Honnef 
,  Germany  (EF) 

HACK,  HARVEY  (Dr.)  Ocean  Systems,  Northrop  Grumman  Corp.,  POP 
Box  1488,  MS  9105,  Annapolis  MD  21404  (F) 

HACSKAYLO,  EDWARD  (Dr.)  7949  N  Sendero  Uno,  Tucson  AZ 
85704-2066  (EF) 

HAIG,  SJ,  FRANK  R.  (Rev.)  Loyola  College,  4501  North  Charles  St, 
Baltimore  MD  21210  (F) 

HANEL,  RUDOLPH  A.  (Dr.)  3881  Bridle  Pass,  Ann  Arbor  MI  481  OS- 
2264  (EF) 

HAYNES,  ELIZABETH  D  (Mrs.)  7418  Spring  Village  Dr.,  Apt  CS  422, 
Springfield  VA  22150-4931  (M) 

HAZAN,  PAUL  14528  Chesterfield  Rd,  Rockville  MD  20853  (F) 
HEANEY,  JAMES  B  6  Olive  Ct,  Greenbelt  MD  20770  (M) 

HERBST,  ROBERT  L.  (Mr.)  4109  Wynnwood  Drive,  Annadale  VA 
22003  (LF) 

HEYER,  W.  RONALD  (Dr.)  MRC  162,  PO  Box  37012,  Smithsonian 
Institution,  Washington  DC  20013-7012  (F) 

HIBBS,  EUTHYMIA  D  (Dr.)  7302  Durbin  Terrace,  Bethesda  MD  20817 
(M) 

HILL,  Christopher  T.  (Dr.)  George  Mason  Univ.  Original  Bldg.  Rm.  236, 
Mail  Stop  3B1,  3401  Fairfax  Dr.  Arlington,  VA  22030 
HERSHON,  Bob  (Mr.)  Directorate  for  Human  Resources  Programs, 
AAAS,  1200  New  York  Ave.  NW  Washington,  DC  20005 
HOFFELD,  J.  TERRELL  (Dr.)  11307  Ashley  Drive,  Rockville  MD 
20852-2403  (F) 


Winter  2006 


84 


HOLLAND,  PH  D.,  MARK  A.  201  Oakdale  Rd.,  Salisbury  MD  21801 
(M) 

HOLLINSHEAD,  ARIEL  (Dr.)  23465  Harbor  View  Rd  #622,  Punta 
Gorda  FL  33980-2162  (EF) 

HONIG,  JOHN  G  (Dr.)  7701  Glenmore  Spring  Way,  Bethesda  MD 
20817 (LF) 

HOOVER,  LARRY  A.  (Mr.)  1541  Stableview  Drive,  Gastonia  NC  28056- 
1658  (M) 

HOROWITZ,  EMANUEL  (Dr.)  Apt  618,  3100  N.  Leisure  World  Blvd, 
Silver  Spring  MD  20906  (EF) 

HOWARD,  SETHANNE  (Dr.)  5526  Dory  Lane,  Columbia  MD  21044 
(M) 

HOWARD-PEEBLES,  PATRICIA  (Dr.)  1457  Cattle  Baron  Court, 
Fairview  TX  75069  (EF) 

HUDSON,  COLIN  M.  (Dr.)  107  Lambeth  Drive,  Asheville  NC  28803- 
3429 (EF) 

HUMMEL,  LANI  S.  (Ms.)  PO  Box  3520,  Annapolis  MD  21403-0520  (M) 
HURDLE,  BURTON  G.  (Dr.)  6222  Berkley  Road3440  south  Jefferson  St, 
Falls  Church  VA  22041  (F) 

HUTTON,  GEORGE  L.  (Mr.)  1086  Continental  Avenue,  Melbourne  FL 
32940  (EF) 

IKOSSI,  KIKI  (Dr.)  6275  Gentle  LN,  Alexandria  VA  22310  (M) 

JACOX,  MARILYN  E.  (Dr.)  10203  Kindly  Court,  Montgomery  Village 
MD  20886-3946  (F) 

JARRELL,  H.  JUDITH  (Dr.)  9617  Alta  Vista  Ter.,  Bethesda  MD  20814 

(F) 

JENSEN,  ARTHUR  S.  (Dr.)  Chapel  Gate  1 104,  Oak  Crest,  8820  Wather 
Blvd,  Parkview  MD  21234-9022  (LF) 

JOHNSON,  EDGAR  M.  (Dr.)  1384  Mission  San  Carlos  Drive,  Amelia 
Island  FL  32034  (LF) 

JOHNSON,  GEORGE  P.  (Dr.)  3614  34th  Street,  N.W.,  Washington  DC 
20008  (EF) 

JOHNSON,  JEAN  M.  (Dr.)  3614  34th  Street,  N.W.,  Washington  DC 
20008  (EF) 

JOHNSON,  PHYLLIS  T.  (Dr.)  833  Cape  Drive,  Friday  Harbor  WA  98250 
(EF) 

JONG,  SHUNG-CHANG  (Dr.)  8892  Whitechurch  Ct,  Bristow  VA  20136- 
2005  (LF) 

JORDANA,  ROMAN  DE  VICENTE  (Dr.)  Batalla  De  Garellano,  15, 
Aravaca,  28023,  Madrid,  Spain  (EF) 


Washington  Academy  of  Sciences 


85 


JULIENNE,  PAUL  S.  (Dr.)  100  Bureau  Drive,,  Stop  8423,  Atomic 
Physics  Division,  National  Institute  of  Standards  and  Technology, 
Gaithersburg  MD  20899  (F) 

KAHN,  ROBERT  E.  (Dr.)  909  Lynton  Place,  Mclean  VA  22102  (F) 
KAPETANAKOS,  C.A.  (Dr.)  4431  MacArthur  Blvd,  Washington  DC 
20007  (EF) 

KATZ,  ROBERT  (Dr.)  Omega-3  Research  Institute  Inc.,  Suite  700,  3 
Bethesda  Metro  Center,  Bethesda  MD  20814  (F) 

KAY,  PEG  (Ms.)  Vertech  Inc.,  61 1 1  Wooten  Drive,  Falls  Church  VA 
22044  (LF) 

KEEFER,  LARRY  (Dr.)  7016  River  Road,  Bethesda  MD  20817  (F) 
KEISER,  BERNHARD  E.  (Dr.)  2046  Carrhill  Road,  Vienna  VA  22181 
(F) 

KIPSHIDZE,  NICHOLAS  (Dr.)  Cardiovascular  Research  Foundation,  55 
East  59th  St.  6th  floor.  New  York  NY  10022-1 1 12  (F) 

KIRKBRIDE,  JR.,  JOSEPH  H.  (Dr.)  1001  Devere  Drive,  Silver  Spring 
MD  20903  (F) 

KLINGSBERG,  CYRUS  (Dr.)  1318  Deerfield  Drive,  State  College  PA 
16803  (EF) 

KLOPFENSTEIN,  REX  C.  (Mr.)  4224  Worcester  Dr.,  Fairfax  VA  22032- 
1140 (LF) 

KRUGER,  JEROME  (Dr.)  619  Warfield  Drive,  Rockville  MD  20850  (EF) 
LANHAM,  CLIFFORD  E.  (Mr.)  P.0  Box  2303,  Kensington  MD  20891 
(F) 

LASLO,  ZOHAR  (Dr.Prof.)  10  Haseora  Street,  Rehovot  76454  ,  Israel  (F) 
LAWSON,  ROGER  H.  (Dr.)  10613  Steamboat  Landing,  Columbia  MD 
21044  (EF) 

LEE,  YONG-SOK  (Dr.)  10991  Centrepointe  Way,  Fairfax  Station  VA 
22039 (F) 

LEIBOWITZ,  LAWRENCE  M.  (Dr.)  3903  Laro  Court,  Fairfax  VA  22031 
(LF) 

LEINER,  ALAN  L.  (Mr.)  Apartment  635,  850  Webster  Street,  Palo  Alto 
C A  94301-2837  (EF) 

LENTZ,  PAUL  LEWIS  (Dr.)  5  Orange  Court,  Greenbelt  MD  20770  (EF) 
LESHUK,  RICHARD  (Mr)  9004  Paddock  Lane,  Potomac  MD  20854  (M) 
LEWIS,  DAVID  C.  (Dr.)  609  Sideling  Court,  Vienna  VA  22180  (F) 
LEWIS,  E.  NEIL  (Dr.)  Spectral  Dimensions,  Inc.,  3416  Olandwood  Court, 
Olney  MD  20832  (F) 

LIBELO,  LOUIS  F.  (Dr.)  9413  Bulls  Run  Parkway,  Bethesda  MD  20817 
(LF) 


Winter  2006 


86 


LINDQUIST,  P  E.,  ROY  P  (Mr.)  4109  Fountainside  Lane,  Fairfax  VA 
22030-6097  (F) 

LING,  LEE  (Mr.)  1608  Bel  voir  Drive,  Los  Altos  CA  94024  (EF) 

LINK,  CONRAD  B  (Dr.)  407  Russell  Avenue,  #813,  Gaithersburg  MD 
20877  (EF) 

LIPSETT,  MORLEY  (Dr.)  1529  Whitesails  Drive,  RR1,  Z-62,  Bowen 
Island,  Be  VON  1G0  ,  Canada  (EF) 

LONDON,  MARILYN  (Ms.)  3520  Nimitz  Rd,  Kensington  MD  20895  (F) 
LONG,  BETTY  JANE  (Mrs.)  416  Riverbend  Road,  Fort  Washington  MD 
20744-5539  (F) 

LOOMIS,  TOM  H.  W.  (Mr.)  11502  Allview  Dr.,  Beltsville  MD  20705 
(EM) 

LOVEJOY,  THOMAS  E.  (Dr.)  The  H.  John  Heinz  III  Center  for  Science, 
Economics,  and  the  Environment,  1001  Pennsylvania  Ave.,  NW,  STE. 

735  South,  Washington  DC  20004  (F) 

LUTZ,  ROBERT  J.  (Dr.)  17620  Shamrock  Drive,  Olney  MD  20832  (F) 
LYON,  HARRY  B.  (Mr.)  7722  Northdown  Road,  Alexandria  VA  22308- 
1329  (M) 

LYONS,  JOHN  W.  (Dr.)  7430  Woodville  Road,  Mt.  Airy  MD  21771  (EF) 
MADHAVAN,  GURUPRSAD  State  University  of  New  York,  143 
Washington  St  #2f,  Binghamton  NY  13901-3108  (M) 

MALCOM,  SHIRLEY  M.  (Dr.)  12901  Wexford  Park  Court,  Clarksville 
Maryland  20005  (F) 

MANDERSCHEID,  RONALD  W.  (Dr.)  10837  Admirals  Way,  Potomac 
MD  20854-1232  (LF) 

MARTIN,  CHARLES  R.  (Dr.)  PO  Box  7,  Huntington  MD  20639  (F) 
MARTIN,  WILLIAM  F  9949  Elm  Street,  Lanham  MD  20706  (F) 
MARTIN,  P.E  BCEE,  EDWARD  J  (Dr.)  15366  Stillwell  Road, 
Huntsburg  OH  44046  (M) 

MARVEL,  KEVIN  B.  (Dr.)  American  Astronomical  Society,  Suite  400, 
2000  Florida  Ave  NW,  Washington  DC  20009  (M) 

MATHER,  John  (Dr.)  NASA  Goddard  Space  Flight  Center,  JWST  Project 
Office,  Mailstop  433.0,  Greenbelt  MD  20771  (F) 

MENZER,  ROBERT  E.  (Dr.)  90  Highpoint  Dr,  Gulf  Breeze  FL  32561- 
4014(F) 

MESSINA,  CARLA  G.  (Mrs.)  9800  Marquette  Drive,  Bethesda  MD 
20817  (F) 

METAILIE,  GEORGES  C.  (DR.)  18,  Rue  Liancourt,  75014  Paris  , 
FRANCE  (F) 

MEYLAN,  THOMAS  (Dr.)  3550  Childress  Terrace,  Burtonsville  MD 
20866  (M) 


Washington  Academy  of  Sciences 


87 


MILLER,  LANCE  A.  (Dr.)  7403  Buffalo  Avenue,  Takoma  Park  MD 
20912  (EF) 

MINTZ,  RAYMOND  D.  (Mr.)  815  Duke  Street,  Rockville  MD  20850  (F) 
MITTLEMAN,  DON  (Dr.)  Apartment  909,  5200  Brittny  Dr.  S,  St. 
Petersburg  FL  33715-1538  (EF) 

MOROWITZ,  HAROLD  J  (Dr  )  The  Krasnow  Institute  for  Advanced 
Study,  Mail  Stop  2A1,  George  Mason  University,  Fairfax  VA  22030  (M) 
MORRIS,  J.  ANTHONY  (Dr.)  4550  N  Park  Ave  Apt  104,  Chevy  Chase 
MD  20815-7234  (M) 

MORRIS,  P.E.,  ALAN  (Dr.)  4550  N.  Park  Ave.  #104,  Chevy  Chase  MD 
20815  (EF) 

MOUNTAIN,  RAYMOND  D.  (Dr.)  5  Monument  Court,  Rockville  MD 
20850 (F) 

MUMMA,  MICHAEL  J.  (Dr.)  210  Glen  Oban  Drive,  Arnold  MD  21012 

(F) 

MURDOCH,  WALLACE  P.  (Dr.)  65  Magaw  Avenue,  Carlisle  PA  17015 
(EF) 

NEKRASOV,  ARKADI  (Dr.)  Bldg.  1,  420  Flat,  House  4  Kuncevskaja  St, 
121351  Moscow  ,  Russia  CIS  (F) 

NOFFSINGER,  TERRELL  L.  (Dr.)  125  Echo  Valley  Road,  Auburn  KY 
42206  (EF) 

NORRIS,  KARL  H.  (Mr.)  1 1204  Montgomery  Road,  Beltsville  MD 
20705  (EF) 

O'HARE,  JOHN  J.  (Dr.)  108  Rutland  Blvd,  West  Palm  Beach  FL  33405- 
5057 (EF) 

OHRINGER,  LEE  (Mr.)  5014  Rodman  Road,  Bethesda  MD  20816  (EF) 
ORDWAY,  FRED  (Dr.)  5205  Elsmere  Avenue,  Bethesda  MD  20814- 
5732 (EF) 

OSER,  HANS  J.  (Dr.)  8810  Quiet  Stream  Court,  Potomac  MD  20854- 
4231 (EF) 

OSTENSO,  GRACE  (Dr.)  9707  Old  Georgetown  Rd  #2618,  Bethesda  MD 
20814-1763  (EF) 

OTT,  WILLIAM  R.  (Dr.)  Physics  Laboratory,  National  Institute  of 
Standards  and  Technology,  100  Bureau  Drive,  Stop  8400,  Gaithersburg 
MD  20899-8400  (F) 

PARASCANDOLA,  JOHN  (Dr.)  11503  Patapsco  Dr,  Rockville  MD 
20852  (M) 

PARR,  Albert  C.  (Dr.)  NIST,  100  Bureau  Dr.,  MS-8440,  Gaithersburg, 

MD  20877  (F) 

PATEL,  D.  G.  (Dr.)  1 1403  Crownwood  Lane,  Rockville  MD  20850  (F) 
PAYNE,  ZABORIAM  E.  (F) 


Winter  2006 


88 


PA Z,  ELVIRA  L.  (Dr.)  172  Cook  Hill  Road,  Wallingford  CT  06492  (EF) 
PERROS,  THEODORE  P.  (Dr.)  500  23rd  Str.  NW  B-606,  Washington 
DC  20037  (EF) 

PICKHOLTZ,  RAYMOND  L  (Dr )  3613  Glenbrook  Road,  Fairfax  VA 
22031-3210  (EF) 

POLAVARAPU,  MURTY  8610  Dellway  La,  Vienna  VA  22180  (F) 
POLLARD,  HARVEY  B  (Dr.)  Department  of  Anatomy,  Phsiology,,  and 
Genetics,  USUHS,  Naval  Medical  Center,  Bethesda  MD  20814  (F) 
PROCTOR,  JOHN  H.  (Dr.)  102  Moray  Firth,  Ford’s  Colony, 

Williamsburg  VA  23 1 88  (LF) 

PRYOR,  C.  NICHOLAS  (Dr.)  2299  Puppy  Creek  Rd„  Amherst  VA 
24591  (F) 

PRZYTYCKI,  JOZEF  M.  (Prof.)  10005  Broad  St,  Bethesda  MD  20814 

(F) 

PYKE,  JR,  THOMAS  N.  (Mr.)  4887  N.  35th  Road,  Arlington  VA  22207 
(F) 

QUIROZ,  RODERICK  S.  (Mr.)  4520  Yuma  Street,  N.W.,  Washington  DC 
20016 (EF) 

RADER,  CHARLES  A.  (Mr.)  1101  Paca  Drive,  Edgewater  MD  21037 
(EF) 

RAJAGOPAL,  A  K  Code  6860.1,  Naval  Research  Laboratory, 
Washington  DC  20375  (EF) 

RALL,  JOSEPH  EDWARD  (Dr.)  3947  Baltimore  Street,  Kensington  MD 
20895  (EF) 

RAMAKER,  DAVID  E  (Dr.)  6943  Essex  Avenue,  Springfield  VA  22150 

(F) 

RAMSEY,  NORMAN  F.  (Dr.)  Lyman  Physics  Laboratory,  Harvard 
University,  Cambridge  MA  02138  (LF) 

RAUSCH,  ROBERT  L.  (Dr.)  P  O.  Box  85447,  University  Station,  Seattle 
WA  98145-1447  (F) 

RAVITSKY,  CHARLES  (Mr.)  37129  Village  37,  Camarillo  CA  93012 
(EF) 

REDISH,  EDWARD  F.  (Prof.)  6820  Winterberry  Lane,  Bethesda  MD 
20817(F) 

REINER,  ALVIN  (Mr.)  1 1243  Bybee  Street,  Silver  Spring  MD  20902 
(EF) 

RHYNE,  JAMES  J.  (Dr.)  1830  Corona  Ave.,  Los  Alamos  NM  87544- 
5767 (F) 

RICKER,  RICHARD  (Dr.)  12809  Talley  Ln,  Darnestown  MD  20878- 
6108  (F) 


Washington  Academy  of  Sciences 


89 


RIDGELL,  MARY  P.0  Box  133,  48073  Mattapany  Road,  St.  Mary's  City 
MD  20686-0133  (LM) 

ROBERTS,  SUSAN  (Dr.)  Ocean  Studies  Board,  Keck  752,  National 
Research  Council,  500  Fifth  Street,  NW,  Washington  DC  20001  (F) 
ROBINSON,  MICHAEL  HILL  (Dr.)  8291  SW  Bent  Oak  Court,  Stuart  FL 
34997  (EF) 

ROESCH,  DARREN  M  (Dr.)  Unit  808,  7915  Eastern  Ave,  Silver  Spring 
MD  20910  (M) 

ROSE,  WILLIAM  K.  (Dr.)  10916  Picasso  Lane,  Potomac  MD  20854  (F) 
ROSENBLATT,  JOAN  R.  (Dr.)  Apartment  702,  2939  Van  Ness  Street. 
N.W,  Washington  DC  20008  (EF) 

SAENZ,  ALBERT  W.  (Dr.)  6338  Old  Town  Court,  Alexandria  VA  22307 

(F) 

SAFRANEK,  BARBARA  (Dr.)  1401  N.  Taft  Street  #526,  Arlington  VA 
22201  (M) 

SAMARAS,  THOMAS  T.  (Mr.)  1 1487  Madera  Rosa  Way,  San  Diego  CA 
92124  (M) 

SAVILLE,  JR,  THORNDIKE  (Mr.)  5601  Albia  Road,  Bethesda  MD 
20816-3304  (LF) 

SCHALK,  JAMES  M.  (Dr.)  267  Forest  Trl,  Isle  of  Palms  SC  29451-2518 
(EF) 

SCHINDLER,  ALBERT  I.  (Dr.)  6615  Sulky  Lane,  Rockville  MD  20852 

(F) 

SCHMEDDLER,  NEAL  F.  (Mr.)  Omni  Engr  &  Technology,  Inc, 
82200Greensboro  Dr  #900,  McLean  VA  22102  (F) 

SCHMIDT,  CLAUDE  H.  (Dr.)  1827  North  3rd  Street,  Fargo  ND  58102- 
2335 (EF) 

SCHROFFEL,  STEPHEN  A.  1860  Stratford  Park  PI  #403,  Reston  VA 
20190-3368  (F) 

SCRIBNER,  BOURDON  F.  (Mr.)  9109  River  Crescent  Dr.,  Annapolis 
MD  21401-7731  (EF) 

SEBRECHTS,  MARC  M.  (Dr.)  7014  Exeter  Road,  Bethesda  MD  20814 

(F) 

SEITZ,  FREDERICK  (Dr.)  Rockefeller  University,  1230  York  Avenue, 
New  York  NY  10021  (EF) 

SEVERINSKY,  ALEX  J  (M) 

SHAFRIN,  ELAINE  G.  (Mrs.)  800  4th  St  SW,  No.  N702,  Washington  DC 
20024  (EF) 

SHENGELIA,  RAMAZ  (Prof.)  Dean  of  the  Medical  Faculty,  University 
of  Tbilisi,  7  Asatiani  Street,  Tbilisi  0177,  GEORGIA  (F) 


Winter  2006 


90 


SHENGEL1A,  RAMAZ  M.  (Professor)  71  Asatiani  Street  Tbilisi  0177, 
Georgia  (F) 

SHETLER,  STANWYN  G.  (Dr.)  142  E  Meadowland  Ln,  Sterling  VA 
20164-1144  (EF) 

SHRIER,  STEFAN  (Dr.)  PO  Box  19139,  Alexandria  VA  22320-0139  (F) 
SFIROPSHIRE,  JR,  W.  (Dr.)  Omega  Laboratory,  P  O.  Box  189,  Cabin 
John  MD  20818-0189  (LF) 

SILBER,  CRISTINA  C.  7803  Beard  Ct,  Falls  Church  VA  22043  (M) 
SILVER,  DAVID  M.  (Dr.)  Applied  Physics  Laboratory,  1 1 100  Johns 
Hopkins  Road,  Laurel  MD  20723-6099  (M) 

SIMHA,  ROBERT  (Dr.)  Dept.  Macromolecular  Sci.,  Case-Western 
Reserve  University,  Cleveland  OH  44106-7202  (EF) 

SIMPSON,  MICHAEL  M.  (Dr.)  101  Independence  SE,  CRS  RSI  LM423, 
Washington  DC  20540-7450  (LM) 

SLACK,  LEWIS  (Dr.)  Carol  Woods  #1 1 14,  750  Weaver  Dairy  Road, 
Chapel  Hill  NC  27514-1441  (EF) 

SMITH,  THOMAS  E.  (Dr.)  Dept  of  Biochemistry  &  Molecular  Biol., 
College  of  Medicine,  Howard  University,  520  W.  Street,  NW, 
Washington  DC  20059  (LF) 

SODERBERG,  DAVID  L.  (Mr.)  403  West  Side  Dr.  Apt.  102, 
Gaithersburg  MD  20878  (M) 

SOLAND,  RICHARD  M.  (Dr.)  SEAS,  George  Washington  Univ., 
Washington  DC  20052  (LF) 

SOLDIN,  STEVEN  J.  (Dr.)  6308  Walhonding  Road,  Bethesda  MD  20813 

(F) 

SOUSA,  ROBERT  J.  (Dr.)  168  Wendell  Road,  Shutesbury  MA  01072 
(EF) 

SPANO,  MARK  (Dr.)  239  Chestertown  Street,  Gaithersburg  MD  20878 

(F) 

SPARGO,  WILLIAM  J  (Dr.)  9610  Cedar  Lane,  Bethesda  MD  20814  (F) 
SPILHAUS,  JR,  A  F  (Dr.)  10900  Picasso  Lane,  Potomac  MD  20854  (F) 
STEGUN,  IRENE  A.  (Ms.)  93  Park  Ave  #1406,  Danbury  CT  06810-7625 
(F) 

STERN,  KURT  H.  (Dr.)  103  Grant  Avenue,  Takoma  Park  MD  20912- 
4328 (EF) 

STIEF,  LOUIS  J.  (Dr.)  332  N  St.,  SW„  Washington  DC  20024  (EF) 
STRAUSS,  SIMON  W.  (Dr.)  4506  Cedell  Place,  Camp  Springs  MD 
20748 (LF) 

SYKES,  ALAN  O.  (Dr.)  304  Mashie  Drive,  Vienna  VA  22180  (EM) 
SZTEIN,  ESTER  (Dr.)  8509  Cottage  st.,  Vienna  VA  22180  (M) 


Washington  Academy  of  Sciences 

_ - 


91 


TABOR,  HERBERT  (Dr.)  NIDDK,  LBP,  Bldg  8,  Rm  223,  National 
Institutes  of  Health,  Bethesda  MD  20892-0830  (M) 

TAMARGO,  JUAN  (Dr.)  Guzman  El  Bueno  100,  3  A,  28003  Madrid  , 
Spain  (F) 

TAUBENBERGER,  Jeffery  Karl  (Dr.)  F 

TAYLOR,  P.E.,  WILLIAM  B.  (Mr.)  4001  Belle  Rive  Terrace,  Alexandria 
VA  22309  (M) 

TEICH,  ALBERT  H  (Dr )  Science  &  Policy  Programs,  American 
Association  for  the  Advancement  of  Science,  1200  New  York  Avenue, 
N.W.,  Washington  DC  20005  (F) 

THOMPSON,  F.  CHRISTIAN  (Dr.)  661 1  Green  Glen  Ct,  Alexandria  VA 
22315-5518  (LF) 

TIMASHEV,  SLAVA  A.  (Mr.)  3306  Potterton  Dr.,  Falls  Church  VA 
22044-1603  (F) 

TOMLINSON,  KEITH  PHILLIP  3235  Doctors  Crossing  Road, 

Chari otteville  VA  22911  (F) 

TOUWAIDE,  ALAIN  Department  of  Botany  -  MRC  166,  National 
Museum  of  Natural  History,  PO  Box  37012,  Smithsonian  Institution, 
Washington  DC  20013-7012  (LF) 

TOWNSEND,  LEWIS  R  (Dr.)  8906  Liberty  Lane,  Potomac  MD  20854 
(M) 

TOWNSEND,  MARJORIE  R.  (Mrs  )  3529  Tilden  Street,  NW, 
Washington  DC  20008-3194  (LF) 

TYLER,  PAUL  E.  (Dr.)  1023  Rocky  Point  Ct.  N.E.,  Albuquerque  NM 
87123-1944  (EF) 

UBELAKER,  DOUGLAS  H.  (Dr.)  Dept,  of  Anthropology,  National 
Museum  of  Natural  History,  Smithsonian  Institution,  Washington  DC 
20560-01 12(F) 

UHLANER,  J  E.  (Dr.)  5  Maritime  Drive,  Corona  Del  Mar  CA  92625  (EF) 
UMPLEBY,  STUART  (Professor)  Department  of  Management  Science, 
The  George  Washington  University,  Washington  DC  20052  (F) 

VAN  FLANDERN,  TOM  (Dr.)  Meta  Research,  994  Woolsey  Ct,  Sequim 
WA  98382-5058  (EF) 

VAN  TUYL,  ANDREW  (Dr.)  1000  W.  Nolcrest  Drive,  Silver  Spring  MD 
20903  (EF) 

VANE  III,  RUSSELL  RICHARDSON  (Dr.)  2102  Capstone  Circle, 
Herndon  VA  20170  (M) 

VARADI,  PETER  F.  (Dr.)  Apartment  1606W,  4620  North  Park  Avenue, 
Chevy  Chase  MD  20815  (EF) 

VAVRICK,  DANIEL  J.  (Dr.)  10314  Kupperton  Court,  Fredricksburg  VA 
22408  (F) 


Winter  2006 


92 


VIZAS,  CHRISTOPHER  (Dr.)  504  East  Capitol  Street,  NE,  Washington 
DC  20003  (M) 

WALDMANN,  THOMAS  A.  (Dr.)  3910  Rickover  Road  Silver  Spring 
MD  20902  (F) 

WALLER,  JOHN  D.  (Dr.)  5943  Kelley  Court,  Alexandria  VA  22312- 
3032  (M) 

WARD,  SHERRY  L  (Dr.)  6710  Meadowlawn  Circle,  New  Market  MD 
21774  (M) 

WAYNANT,  RONALD  W.  (Dr  )  6525  Limerick  Court,  Clarksville  MD 
21029(F) 

WEBB,  RALPH  E.  (Dr.)  21-P  Ridge  Road,  Greenbelt  MD  20770  (F) 
WEGMAN,  EDWARD  J.  (Dr.)  368  Research  Bldg,  Center  Computer 
Statistics  MS  6A2,  George  Mason  University,  Fairfax  VA  22030  (LF) 
WEISS,  ARMAND  B.  (Dr.)  6516  Truman  Lane,  Falls  Church  VA  22043 
(LF) 

WERGIN,  WILLIAM  P  (Dr.)  1  Arch  Place  #322,  Gaithersburg  MD 
20878  (EF) 

WESEMANN,  JODI  L  (Dr)  (M) 

WIESE,  WOLFGANG  L.  (Dr.)  8229  Stone  Trail  Drive,  Bethesda  MD 
20817 (EF) 

WINKLER,  STANLEY  (Dr.)  6413  Earlham  Dr,  Bethesda  MD  20817  (F) 
WINTERS,  WILLIAM  W.  6825  Capri  Place,  Bethesda  MD  20817-4209 
(LM) 

WITHERSPOON,  F  DOUGLAS  National  Capital  Society  of  American 
Foresters,  1 1316  Smoke  Rise  Ct.,  Fairfax  Station  VA  22039  (M) 

WULF,  WILLIAM  A.  (Dr.)  National  Academy  of  Engineering, 
Washington  DC  20001  (F) 


Washington  Academy  of  Sciences 


DELEGATES  TO  THE  WASHINGTON  ACADEMY  OF  SCIENCES 
REPRESENTING  AFFILIATED  SCIENTIFIC  SOCIETIES 


Acoustical  Society  of  America 

American/Intemational  Association  of  Dental  Research 
American  Association  of  Physics  Teachers 
American  Ceramics  Society 
American  Fisheries  Society 
American  Institute  of  Aeronautics  and  Astronautics 
American  Institute  of  Mining,  Metallurgy  &  Exploration 
American  Meteorological  Society 
American  Nuclear  Society 
American  Phytopathological  Society 
American  Society  for  Cybernetics 
American  Society  for  Microbiology 
American  Society  of  Civil  Engineers 
American  Society  of  Mechanical  Engineers 
American  Society  of  Plant  Physiology 
Anthropological  Society  of  Washington 
ASM  International 

Association  for  Women  in  Science  (AWIS) 

Association  for  Computing  Machinery 
Association  for  Science,  Technology,  and  Innovation 
Association  of  Information  Technology  Professionals 
Biological  Society  of  Washington 
Botanical  Society  of  Washington 
Chemical  Society  of  Washington 
District  of  Columbia  Institute  of  Chemists 
District  of  Columbia  Psychology  Association 
Eastern  Sociological  Society 
Electrochemical  Society 
Entomological  Society  of  Washington 
Geological  Society  of  Washington 
Historical  Society  of  Washington,  DC 
History'  of  Medicine  Society 
Human  Factors  and  Ergonomics  Society 

Institute  of  Electrical  and  Electronics  Engineers,  Washington  Section 
Institute  of  Electrical  and  Electronics  Engineers,  Northern  Virginia  Section 
Institute  of  Food  Technologies 
Institute  of  Industrial  Engineers 
Instrument  Society  of  America 
Marine  Technology  Society 
Mathematical  Association  of  America 
Medical  Society  of  the  District  of  Columbia 
National  Capital  Astronomers 
National  Geographic  Society 
Optical  Society  of  America 
Pest  Science  Society  of  America 
Philosophical  Society  of  Washington 
Society  of  American  Foresters 
Society  of  American  Military  Engineers 
Society  of  Experimental  Biology  and  Medicine 
Society  of  Manufacturing  Engineers 
Soil  and  Water  Conservation  Society 
Technology  Transfer  Society 
Washington  Evolutionary  Systems  Society 
Washington  History  of  Science  Club 
Washington  Chapter  of  the  Institute  for  Operations 
Research  and  Management  Science 
Washington  Paint  Technology  Group 
Washington  Society  of  Engineers 
Washington  Statistical  Society 
World  Future  Society 


Paul  Arveson 
J.  Terrell  Hoffeld 
Frank  R.  Haig,  S.J. 

VACANT 
Ramona  Schreiber 
David  W.  Brandt 
Michael  Greeley 
Kenneth  Carey 
Steven  Arndt 
Kenneth  L.  Deahl 
Stuart  Umpleby 
VACANT 
Kimberly  Hughes 
Daniel  J.  Vavrick 
Mark  Holland 
Marilyn  London 
Toni  Marechaux 
Emanuela  Appetiti 
Lee  Ohringer 
F.  Douglas  Witherspoon 
Barbara  Safranek 
VACANT 
Alain  Touwaide 
James  J.  Zwolenik 
James  J.  Zwolenik 
David  Williams 
Ronald  W.  Mandersheid 
Robert  L.  Ruedisueli 
F.  Christian  Thompson 
Bob  Schneider 
VACANT 
Alain  Touwaide 
Douglas  Griffith 
Gerard  Christman 
Murty  Polavarapu 
Isabel  Walls 
Russell  Wooten 
Hank  Hegner 
Judith  T.  Krauthamer 
Sharon  K.  Hauge 
Duane  Taylor 
Jay  H.  Miller 
VACANT 
Jim  Cole 
VACANT 
Vary  T.  Coates 
G.  Foster 
VACANT 
Darren  Roesch 
VACANT 
Bill  Boyer 
Clifford  Lanham 


Jerry  L.R.  Chandler 
Albert  G.  Gluckman 
Russell  Wooten 

VACANT 
Alvin  Reiner 
Michael  P.  Cohen 
Russell  Wooten 


Washington  Academy  of  Sciences 
Room  637 

1200  New  York  Ave.  NW 
Washington,  DC  20005 
Return  Postage  Guaranteed 


Media  l\ 

U.S.  Pos 

PAID 

Merrifield.  VA 
Permit  NO.  1502 


ERNST  MAYR  LIBRARY 
MUSEUM  COMP  ZOOLOGY 
HARVARD  UNIVERSITY 
26  OXFORD  ST 
CAMBRIDGE.  MA  02138-2902