Skip to main content

Full text of "Information storage and neural control; tenth annual scientific meeting of the Houston Neurological Society jointly sponsored by the Dept. of Neurology, Baylor University College of Medicine, Texas Medical Center, Houston, Texas"

See other formats


Marine  Biological  Laboratory  Library 

Woods  Hole,  /Massachusetts 


Presented  by  t.ie  i-lBL  Associates-1971  Gift 


INFORMATION  STORAGE  AND  NEURAL  CONTROL 


^-^ 


^^  ^^-1 


INFORMATION  STORAGE 

and 

NEURAL  CONTROL 


Tenth  Annual  Scientific 
Meeting  of  the  Houston 
Neurological  Society 

Jointly  Sponsored  by  the 
Department  of  Neurology 
Baylor  University 
College  of  Medicine 
Texas  Medical  Center 
Houston,  Texas 


Compiled  and  Edited  by 
WILLIAM  S.  FIELDS,  M.D. 

Professor  and  Chairman 

Department  of  Neurology 

Baylor  University  College  of  Medicine 

and 

WALTER  ABBOTT,  Ph.D. 

Assistant  Professor  of  Epidemiology 
Director,  Biomathematics  Research  Laboratory 
Baylor  University  College  of  Medicine 


CHARLES  C  THOMAS  .  PUBLISHER 

Springfield   •    Illinois    •    U.S.A. 


Published  and  Distributed  Throughout  the  World  by 

CHARLES  C  THOMAS  •  PUBLISHER 

Bannerstone  House 
301-327  East  Lawrence  Avenue,  Springfield,  Illinois,  U.S.A. 


This  book  is  protected  by  copyright.  No 
part  of  it  may  be  reproduced  in  any  manner 
without  written  permission  from  the  publisher. 


©  1963,  by  CHARLES  C  THOMAS  •  PUBLISHER 
Library  of  Congress  Catalog  Card  Number:  62-20580 


With  THOMAS  BOOKS  careful  attention  is  given  to  all  details  of 
manufacturing  and  design.  It  is  the  Publisher's  desire  to  present  books 
that  are  satisfactory  as  to  their  physical  qualities  and  artistic  possibilities 
and  appropriate  for  their  particular  use.  THOMAS  BOOKS  will  be  true 
to  those  laws  of  quality  that  assure  a  good  name  and  good  will. 


Printed  in  the  United  States  of  America 


CONTRIBUTORS 

Gregory  Bateson,  M.A.:  Ethnologist,  Veterans  Administration 
Hospital,  Palo  Alto,  California;  Professoi  (Visiting)  Depart- 
ment of  Anthropology,  Stanford  University,  Stanford,  California. 

Mary  A.  B.  Brazier,  D.Sc.:  National  Institutes  of  Health  Career 
Professor  at  the  Brain  Research  Institute,  University  of  Cali- 
fornia, Los  Angeles,  California. 

Neil  R.  Burch,  M.D.:  Associate  Professor  of  Psychiatry,  Baylor 
University  College  of  Medicine,  Houston,  Texas. 

Harold  E.  Childers:  Assistant  Professor  of  Biophysics,  Baylor  Uni- 
versity College  of  Medicine,  Houston,  Texas. 

James  E.  Darnell,  Jr.,  M.D.:  Department  of  Biology,  Division  of 
Microbiology,  Massachusetts  Institute  of  Technology,  Cam- 
bridge, Massachusetts. 

Harrison  Echols,  Ph.D.:  Assistant  Professor,  Department  of  Bio- 
chemistry, College  of  Agriculture,  The  University  of  Wisconsin, 
Madison,  Wisconsin. 

Ralph  W.  Gerard,  M.D.,  Ph.D.:  Director  of  Laboratories,  Mental 
Health  Research  Institute,  The  University  of  Michigan,  Ann 
Arbor,  Michigan 

Robert  T.  Gregory,  Ph.D.:  Associate  Professor  of  Mathematics; 
Senior  Research  Mathematician,  Computation  Center,  The 
University  of  Texas,  Austin,  Texas. 

E.  Roy  John,  Ph.D.:  Professor  and  Director,  Center  for  Brain 
Research,  The  University  of  Rochester,  College  of  Arts  and 
Science,  Rochester,  New  York. 

Saul  Kit,  Ph.D.:  Biochemist,  Department  of  Biochemistry;  Head, 
Section  of  Nucleoprotein  Metabolism,  The  University  of  Texas 
M.  D.  Anderson  Hospital  and  Tumor  Institute,  Houston,  Texas. 

Robert  K.  Lindsay,  Ph.D.:  Assistant  Professor  of  Psychology; 
Research  Scientist,  Computation  Center,  The  University  of 
Texas,  Austin,  Texas. 


vi  Contributors 

Warren  S.  McCulloch,  M.D.:  Head,  Neurophysiology  Group, 
Division  of  Sponsored  Research,  Research  Laboratory  of  Elec- 
tronics, Massachusetts  Institute  of  Technology,  Cambridge, 
Massachusetts. 

James  G.  Miller,  M.D.:  Director,  Mental  Health  Research  In- 
stitute, The  University  of  Michigan,  Ann  Arbor,  Michigan. 

Frank  Morrell,  M.D.:  Professor  of  Neurology,  Stanford  University 
School  of  Medicine,  Palo  Alto,  California. 

Bernard  C.  Patten,  Ph.D.:  Associate  Professor  of  Marine  Science, 
Virginia  Institute  of  Marine  Science,  College  of  William  and 
Mary,  Gloucester  Point,  Virginia. 

Bernard  Saltzberg:  Senior  Scientist,  The  Bissett-Berman  Cor- 
poration, Santa  Monica,  California. 


FOREWORD 


X 


HIS  volume  entitled  Information  Storage  and  Neural  Control  is 
compiled  from  the  proceedings  of  the  Tenth  Annual  Scientific 
Meeting"  of  the  Houston  Neurological  Society.  This  meeting,  like 
its  predecessors,  was  concerned  with  the  exploration  of  a  specific 
area  of  current  biomedical  investigation.  For  some  of  those  persons 
who  may  have  occasion  to  read  the  contributions  presented  here 
by  scientists  in  various  disciplines,  there  may  be  little  that  is  im- 
inediately  applicable  in  clinical  medicine.  Many  of  the  concepts 
and  techniques  which  are  described  relate  at  this  time  only  to 
fundamental  research,  but  there  is  no  doubt  that  in  the  future  a 
better  appreciation  of  these  facts  will  be  exceedingly  important 
to  clinicians. 

Progress  in  the  biological  sciences  has  been  impeded  to  a  con- 
siderable extent  by  our  inability  to  obtain  objective  quantitative 
data  in  many  critical  areas  of  research.  Biostatisticians  and 
geneticists  were  among  the  first  to  recognize  this  serious  defect 
and  to  inake  attempts  to  fill  in  the  gaps.  The  concept  of  vary- 
ing information  content  was  introduced  when  I — the  informa- 
tion value  of  a  group  of  observations — was  defined  as  the  re- 
ciprocal of  the  variance  of  the  data.  At  first  glance,  this  concept 
appears  to  be  in  direct  conflict  with  the  modern  idea  of  high 
information  content  for  a  low  probability  datum.  This  need 
not,  necessarily,  be  the  case,  since  a  narrow  range  of  variation 
implies  inclusion  of  low  probability  observations  from  the  extremes 
of  the  normal  curve  with  the  correspondingly  high  information 
content  of  these  low  probability  observations.  The  next  iinportant 
forward  step  resulted  from  the  application  in  the  biological  sciences 
of  physical  and  chemical  laws  derived  from  the  exact  sciences. 
This  period  began  with  the  publication  of  A.  J.  Lotka's  Elements 
of  Physical  Biology  in  1925,  in  which  the  theoretical  concepts  of 
modern  mathematics,  physics,  and  physical  chemistry  were  applied 
rigorously  to  models  of  biological  systems.  Many  of  the  laws  and 


viii  Foreword 

theories  now  widely  accepted  in  various  special  areas  of  research 
can  be  traced  to  this  basic  work.  For  example,  the  studies  of 
Gause  and  Witt  on  competitive  action  in  biological  systems  con- 
stitute experimental  verification  of  several  of  the  models  described 
by  Lotka. 

The  modern  era  of  physical  biology,  or  biological  physics, 
cannot  be  dated  precisely,  but  great  impetus  was  given  to  this 
field  of  investigation  by  the  publication  of  Shannon's  work.  The 
Mathematical  Theory  of  Communication,  in  1948.  The  full  impact  of 
this  monumental  contribution  is  only  now  beginning  to  be  realized. 

The  symposium  from  which  these  papers  were  compiled  was 
organized  for  the  specific  purpose  of  presenting  to  both  basic 
scientists  and  clinicians  a  spectrum  of  applications  of  information 
theory  in  biology.  The  audience,  as  well  as  the  contributors, 
represented  a  diversity  of  disciplines  including  mathematics, 
physics,  chemistry,  virology,  ecology,  physiology,  and  several 
fields  of  clinical  medicine,  such  as  neurology,  psychiatry,  and 
internal  medicine.  It  is  hoped  that  this  volume  will  serve  as  a 
source  of  reference  for  clinicians  and  basic  scientists  alike. 

We  wish  to  acknowledge  the  continued  support  of  Dr.  Hampton 
C.  Robinson,  whose  financial  aid  has  made  possible  the  presen- 
tations of  these  symposia.  Assistance  in  underwriting  publication 
costs  of  the  proceedings  has  been  given  to  us  by  the  M.  B.  and 
Fannie  Finkelstein  Foundation. 

We  also  wish  to  express  our  appreciation  to  Dr.  Wayne  H. 
Holtzman,  Director  of  The  Hogg  Foundation  for  Mental  Health, 
Austin,  Texas,  for  his  helpful  suggestions  in  the  formulation  of 
the  program. 

We  are  grateful  for  the  wonderful  cooperation  given  us  by  the 
contributors  to  this  volume  and  for  the  editorial  assistance  of 
Thelma  Armstrong  and  Joan  Chambers. 

W.  S.  F. 
W.  A. 


CONTENTS 


Page 


Contributors 
Foreword 


V 

vii 


Part  I — Introduction 
Moderator  William  S.  Fields,  M.D. 
Chapter 

I.   What  Is  Information  Theory? — Bernard  Saltzberg      ...        5 

Discussion  of  Chapter  1 17 

II.   Binary  Representation  of  Information — Robert  T.  Gregory    .      27 

III.  Information  Processing  Theory — Robert  K.  Lindsay         .      .      34 

Part  II — Information  in  Biological  Systems 
Moderator:  Heather  D.  Mayor,  Ph.D. 

IV.  Genetic  Control  of  Protein  Synthesis — Harrison  Echols    .      .      59 
Discussion  of  Chapter  IV 72 

V.    Coding    by    Purine    and    Pyrimidine    Moieties    in   Animals, 

Plants,  and  Bacteria — Saul  Kit 76 

Discussion  of  Chapter  V 120 

VI.   Virus  Action  and  Replication — James  E.  Darnell,  Jr.        .      .123 

Discussion  oi  Ch3.\)ie.r  VI 139 

VII.  The  Information  Concept  in  Ecology:  Some  Aspects  of  In- 
formation-Gathering  Behavior   in   Plankton — Bernard    G. 

Patten 140 

Discussion  of  Chapter  VII 171 

VIII.    Exchange  of  Information  About  Patterns  of  Human 

Behavior — ^Gregory  Bateson     „ 173 

Discussion  oi  Chdipi^r  Will 184 


X  Contents 

Chapter  Page 

Part  III — Neurophysiological  Aspects  of 
Information  Storage  and  Transfer 
Moderator:  Hebbel  E.  HofF,  M.D.,  Ph.D. 

IX.    Information  Storage  in  Nerve  Cells — Frank  Morrell  .      .      .  189 
X.   How   Can   Models   From    Information  Theory  Be   Used   in 

Neurophysiology? — Mary  A.  B.  Brazier 230 

Discussion  of  Chapter  X 241 

XI.   Neural  Mechanisms  of  Decision  Making — E.Roy  John    .       .  243 

Discussion  of  Chapter  XI         278 

XII.   Anastomotic  Nets  Combating  Noise — Warren  S.  McCulloch  .  283 

Discussion  of  Chapter  XII 296 

Part  IV — The  Human  Nervous  System 
Moderator:  Wayne  H.  Holtzman,  Ph.D. 

XIII.  The    Individual    as    an    Information    Processing    System — 

James  G.  Miller 301 

XIV.  Information  Processing  in  the  Time  Domain — Neil  R.  Burch 

and  Harold  E.  Childers 329 

Discussion  of  Chapter  XIV 349 

Part  V — Summary  and  General  Discussion 
Moderator:  Ralph  W.  Gerard,  M.D.,  Ph.D. 

XV.   Summary— Ralph  W.  Gerard 353 

General  Discussion 367 

Appendix  A 
Introduction— Michael  H.  Arbib 377 

A  Logical  Calculus  of  the  Ideas  Immanent  in  Nervous 

Activity— Warren  S.  McCulloch  and  Walter  H.  Pitts   .      .    379 

Index 401 


INFORMATION  STORAGE  AND  NEURAL  CONTROL 


PART  I  — INTRODUCTION 

Moderator:  William  S.  Fields,  M.D. 


CHAPTER 
I 

WHAT  IS  INFORMATION  THEORY? 

Bernard  Saltzberg 

^^  INTRODUCTION 

XHE  purpose  of  this  paper  is  to  describe  the  principles  under- 
lying" the  quantitative  aspects  of  the  storage  and  communication 
of  information  so  that  a  better  understanding  may  be  gained  of 
the  nature  of  efficient  information  storage,  with  its  attendant 
implications  in  coding  and  control  processes,  including  neural 
control.  Insofar  as  possible,  the  discussion  will  avoid  abstract 
mathematical  arguments  and  will  be  directed  to  those  with  little  or 
no  previous  acquaintance  with  probability  or  information  theory. 

INFORMATION  MEASURE 

Although  information  theory  is  an  essentially  mathematical 
subject,  a  basic  understanding  of  the  underlying  principles  can 
be  acquired  without  resorting  to  complex  mathematical  argu- 
ments. In  simple  qualitative  terms,  information,  as  defined  by 
Shannon,  *  is  merely  a  measure  of  how  much   uncertainty  has 


*The  formal  development  of  information  theory  originated  in  the  work  of  Claude 
E.  Shannon  of  Bell  Telephone  Laboratories  who  published  his  fundamental  paper, 
"The  Mathematical  Theory  of  Communication,"  in  1948.  In  this  paper  he  set  up 
a  mathematical  scheme  in  which  the  concepts  of  the  production  and  transmission 
of  information  could  be  defined  quantitatively.  Historically  however,  Shannon's 
work  stems  from  certain  early  basic  observations  in  theoretical  physics  concerning 
entropy.  Boltzman  (1894)  observed  that  entropy  is  related  to  "missing  information," 
inasmuch  as  it  is  related  to  the  number  of  alternatives  which  remain  possible  to  a 
physical  system  after  all  the  macroscopically  observable  information  concerning  it 
has  been  recorded.  Leo  Szilard  (1925)  extended  this  idea  to  a  general  discussion  of 
information  in  physics,  and  von  Neumann  (1932)  treated  information  in  quantuin 
mechanics  and  particle  physics.  Information  theory,  as  developed  by  Shannon,  con- 


6  Information  Storage  and  Neural  Control 

been  removed  by  the  receipt  of  a  message.  For  example,  if  you 
are  told  that  the  baby  Dr.  Jones  delivered  today  is  a  boy,  then 
you  have  been  given  one  bit  of  information*  (by  definition  one 
bit  is  the  amount  of  information  necessary  to  resolve  two  equally 
likely  alternatives).  If  the  uncertainty  is  greater,  the  amount  of 
information  necessary  to  remove  it  is  greater.  Therefore,  a  message 
which  identifies  one  of  32  equally  likely  alternatives  contains  more 
information  (five  bits  as  we  shall  explain  later)  than  a  message 
which  resolves   16  equally  likely  alternatives   (four  bits). 

Some  elementary  examples  to  convey  the  basic  notions  of 
information  measure  may  be  helpful  in  developing  some  qualitative 
insights.  Consider  a  simple  game  in  which  you  are  asked  to  guess 
a  number  with  possible  values  from  one  to  eight.  With  no  a  priori 
knowledge  of  which  of  these  numbers  is  the  correct  choice,  the 
probability  of  guessing  the  correct  number  is  1/8.  In  the  language 
of  information  theory,  this  situation  might  be  described  as  follows: 
A  system  is  in  one  of  eight  equally  probable  states,  and  the  state 
of  the  system  is  completely  unknown  to  the  receiver.  It  is  appro- 
priate to  ask  how  much  information  is  conveyed  to  the  receiver 
by  completely  resolving  the  uncertainty  of  the  receiver's  knowledge. 
Let  us  designate  the  eight  equally  probable  states  of  the  system 
by  the  numbers  from  one  to  eight.  Assume  that  you  are  to  ask 
only  binary  questions,  i.e.,  questions  which  admit  only  of  a  yes 
or  no  answer,  in  an  attempt  to  determine  the  state  of  this  system. 
It  is  a  simple  matter  to  discover  that  the  minimum  number  of 
such  questions  certain  to  establish  the  state  of  the  system  is  three. 
In  this  simple  illustration  we  have  introduced  the  basic  concepts 
from  which  the  quantitative  definition  of  information  can  be 
formulated:    namely,   the   number  of  equally  probable  states   of 

nects  more  directly  with  certain  ideas  generated  about  thirty  years  ago  by  H.  Nyquist 
and  R.  V.  L.  Hartley,  both  of  Bell  Telephone  Laboratories.  Professor  Norbert  Weiner's 
work  in  the  study  of  Cybernetics,  which  deals  mainly  with  the  use  of  information  to 
effect  certain  control  actions,  has  been  a  major  impetus  in  applying  information 
theory  to  biological  and  central  nervous  system  phenomena. 

*Information  theory  does  not  deal  with  the  importance  of  the  information  in  a 
message.  For  example,  the  information  in  the  message,  "the  baby  is  a  boy,"  is  one  bit 
independent  of  whether  you  are  the  father.  This  comment  is  made  in  order  to  em- 
phasize the  fact  that  information  theory  docs  not  deal  with  the  subjective  value  of 
information,  which  falls  more  properly  into  the  domain  of  semantics,  but  rather  with 
objective  measures  of  information. 


JVhat  is  Information  Theory?  7 

a  system  (in  this  case  eight),  the  number  of  ahernatives  resolved 
by  each  question  (two,  because  of  the  binary  nature  of  the  ques- 
tion), and  the  minimum  number  of  questions  necessary  to  de- 
termine tlie  state  of  the  system  (three  in  this  case).  It  is  easily 
seen  that  tiie  relationship  between  these  numbers  is  2'^  =  8.  In 
the  vernacular  of  information  theory,  we  say  that  three  bits  of 
information  are  necessary  to  determine  the  state  of  such  a  sys- 
tem; i.e.,  three  appropriately  chosen  questions,  each  of  which 
resolves  two  alternatives,  usually  designated  as  1  or  0,  correspond- 
ing" to  yes  or  no,  are  all  that  is  necessary  to  reduce  indeterminacy 
to  certainty.  The  problem  of  choosing  the  appropriate  questions 
is  analogous  to  that  of  choosing  a  good  code.  For  example,  asking 
the  question:  "Is  the  number  3?"  would  correspond  to  very 
inefficient  coding  of  information.  Phrasing  or  coding  questions  in 
this  way  would  require  that  you  be  allowed  to  ask  eight  questions 
in  order  to  be  certain  to  determine  the  state  of  the  system.  In  this 
illustration  a  correct  way  of  coding  or  phrasing  the  c^uestions 
would  be  as  follows:  Question  1:  "Is  the  number  greater  than  4?" 
If  yes,  then  ask  Question  2:  "Is  the  number  greater  than  6?" 
If  no,  then  ask  Question  3:  "Is  the  number  5?"'  If  no,  then  the 
system  must  be  in  state  6. 

As  previously  mentioned,  the  probability  of  having  guessed  the 
correct  state  before  receiving  these  three  bits  of  information  was 
1/8  in  the  example  used.  After  the  first  bit  of  information  is  re- 
ceived the  probability  of  guessing"  correctly  is  increased  from 
1/8  to  1/4,  after  the  second  bit  from  1/4  to  1/2,  and  after  the 
third  bit  from  1  ;'2  to  1 .  Thus,  each  successive  bit  received  has 
reduced  our  uncertainty  as  to  the  state  of  the  system  until  all  the 
uncertainty  is  removed.  In  this  example,  the  receipt  of  any  more 
information  is  unnecessary  or  redundant.  However,  as  we  shall 
discuss  later,  the  redundancy  may  be  useful  in  correcting  errors 
due  to  noise  in  the  communication  channel. 

In  order  to  use  a  more  general  illustration  which  is  not  restricted 
to  a  system  with  equally  probable  states,  let  us  consider  the  game 
of  Twenty  Questions.  In  most  situations  the  probabilities  of  some 
states,  i.e.,  the  possible  set  of  objects  to  be  identified,  are  higher 
than  the  probabilities  of  others.  A  good  information  theorist  with 
some  a  priori  knowledge  of  the  probabilities  of  these  states  would 


8  Information  Storage  and  Neural  Control 

ask  questions  in  accordance  with  his  a  priori  knowledge  of  these 
probability  states.  Information  theory  as  well  as  intuition  tells  us 
that  a  good  strategy  would  be  to  inquire  about  the  most  likely 
probability  states  first.  This  point  might  be  more  clearly  illustrated 
by  considering  the  information  storage  problem,  which  is  equiva- 
lent in  principle. 

INFORMATION  STORAGE 

Mathematically,  there  is  no  important  difference  between  the 
application  of  information  theory  to  communications  systems 
through  which  information  flows  continuously  and  to  static  sys- 
tems used  for  storing  information.  The  problem  of  storing  infor- 
mation is  essentially  one  of  making  a  representation.  The  repre- 
sentation can  take  any  form  as  long  as  the  original  or  something 
equivalent  to  it  can  be  reconstructed  at  will.  It  is  clear  for  example, 
that  even  though  information  exists  as  sound,  there  is  no  need 
to  store  it  acoustically.  There  is  no  objection  to  the  use  of  a  re- 
versible code  since  information  is  invariant  under  such  a  trans- 
formation and,  therefore,  can  be  stored  equally  well  electrically 
or  magnetically;  as  for  example  on  a  recording  tape.  We  simply 
have  to  insure  that  every  possible  event  to  be  recorded  can  be 
represented  in  the  store.  This  implies  that  an  empty  store  must 
merely  be  capable  of  being  put  into  different  states  and  that  the 
precise  nature  of  these  states  is  quite  immaterial  to  the  question 
of  how  much  information  can  be  stored.  Thus,  the  capacity  of 
an  empty  information  store  depends  only  on  the  total  number 
of  distinguishable  states  of  which  it  admits.  Hence,  the  larger  the 
number  of  states,  the  larger  the  capacity. 

If  a  storage  unit  such  as  a  knob  with  click  positions  has  n  possible 
states,  then  two  such  units  provide  altogether  n''  states.  From  this 
it  is  clear  that  duplication  of  the  basic  units  is  a  powerful  way  to 
increase  storage  capacity.  Since  physically,  it  is  generally  easier 
to  make  two  ^/-state  devices  than  one  single  device  with  n''  states, 
practical  storage  systems  will  generally  be  found  to  consist  of  a 
multiplicity  of  smaller  units.  Thus,  1000  two-state  devices  can 
provide  a  total  of  2'""''  possible  states. 

The  exponential  dependence  of  the  number  of  states  on  the 
number  of  units  immediately  suggests  a  logarithmic  measure  of 


What  is  Information  Theory?  9 

information  capacity  and,  in  fact,  the  information  capacity  of  a 
storage  system  is  defined  by  the  equation, 

C  =  log  n, 
where  n  is  tlie  number  of  distinguishable  states.  This  makes  the 
capacity  of  a  compound  storage  system  equal  to  the  capacity  of  a 
basic  storage  unit  multiplied  by  the  number  of  units  in  the  system. 
If  the  logarithm  is  taken  to  the  base  2,  then  C  is  the  equivalent 
number  of  binary  storage  units  (bits);  and  if  the  logarithm  is 
taken  to  base  10,  then  the  information  capacity  is  given  in  units 
called  Hartleys.  For  example,  the  capacity  of  a  knob  with  32 
click  positions  is  equal  to  that  of  five  two-position  switches  (five 
bits).  A  ten-position  knob,  on  the  other  hand,  has  a  capacity  of 
one  Hartley,  and  two  ten-position  knobs  capable  of  being  placed 
in  100  different  states  have  an  information  capacity  of  two  Hart- 
leys. Since  storage  elements  which  are  binary  in  nature  (two 
positions)  are  much  less  susceptible  to  error  and  are  easier  to 
mechanize,  it  is  more  common  to  deal  with  binary  units  (bits)  of 
information  than  with  decimal  units  of  information  (Hartleys). 

So  far  we  have  discussed  information  storage  and,  correspond- 
ingly, information  capacity.  There  is  an  important  distinction, 
however,  between  information  capacity  and  information  content. 
The  information  content  of  a  message  may  be  defined  as  the ! 
minimum  capacity  required  for  storage.  To  illustrate  this  impor- 
tant point,  consider  a  two-state  message  such  as  a  reply  to  some 
question  which  admits  only  yes  or  no.  If  someone  in  this  audi- 
torium is  asked,  "Are  you  a  doctor?",  then  a  reply  admits  of  two 
possible  message  states  and  it  will  certainly  be  possible  to  store 
the  reply  in  one  binary  storage  unit.  Intuition  tells  us  that  the 
message  contains  one  bit  of  information,  for,  by  itself,  it  cannot 
be  stored  any  more  efficiently.  However,  our  previous  discussion 
has  demonstrated  that  a  bit  of  information  should  substantially 
reduce  uncertainty.  In  view  of  the  fact  that  most  of  the  people 
in  this  auditorium  are  doctors,  I  could  simply  guess  "yes"  for 
each  person  questioned  and  be  correct  most  of  the  time.  Thus, 
one  would  expect  the  average  information  per  question  to  be 
less  than  one  bit,  as  indeed  it  is. 

Using  a  numerical  example  from  Woodward,  suppose  that  128 
people   in   this   auditorium   are   questioned    and    the    128   binary 


1 0  Information  Storage  and  Neural  Control 

messages  have  to  be  stored  in  a  system  of  binary  storage  units. 
(It  will  be  assumed  that  we  are  interested  in  preserving  the  exact 
order  of  the  replies  and  not  simply  in  counting  the  number  of 
yeses  and  noes).  Proceeding  in  the  most  obvious  manner  and  using 
one  storage  unit  for  each  message,  we  should  set  down  a  sequence 
such  as  this: 

YYYYYYYYNYYYYYYYYYYYYYYYNYYY  .  .  . 

The  question,  "Are  you  a  doctor?",  expects  the  answer  yes  from 
this  medical  group,  and  of  the  128  messages  there  will  be  only 
one  or  two  no  states.  Therefore,  it  would  be  more  economical  to 
store  the  positions  of  the  noes  in  the  sequence  and  convert  the 
numbers  9,  25,  corresponding  to  the  noes  in  the  above  sequence, 
into  the  binary  form  as  0001001  and  0011001.  Seven  digits  are 
allowed  because  there  are  2  messages  altogether.  Thus,  the  se- 
quence could  be  coded  into  the  sequence 

00010010011001  .  .  . 

This  makes  use  of  binary  storage  units  just  as  in  the  original 
sequence,  but  a  much  smaller  number  of  them.  It  is  understood, 
as  part  of  the  code,  that  decoding  proceeds  in  blocks  of  seven. 
This  avoids  violating  the  binary  form  of  marking  off  groups  of 
digits.  The  preceding  code,  which  is  only  one  of  many  that  could 
be  devised,  shows  that  a  set  of  two-state  messages  can  sometimes 
be  stored  in  such  a  way  that  each  message  occupies,  on  the  average, 
less  than  one  bit  of  storage  capacity. 

From  such  considerations,  the  following  definition  of  information 
content  is  suggested: 

n 

I  =  -2  ?^.Tog/>i 

whei'e  p.  =  probability  of  the  ith.  state 

n  =  total  number  of  states. 
If  all  n  states  are  equally  probable,  then  it  follows  that  pi  =  \/  n 
for  all  values  of  /.  Thus,  substituting  pi  =  1/n  into  the  expression 
for  /,  we  note  that 

/   =   log  71. 

From  this  it  follows  that  if  all  n  states  are  equally  probable,  the 
information  content  is  exactly  equal  to  the  information  capacity 


What  is  Information  Theory?  11 

of  a  store  with  n  states.  In  other  words,  if  ah  the  states  are  equally 
probable,  then  it  is  not  possible  to  store  the  information  any  more 
efficiently  than  one  bit  of  information  per  message,  on  the  average. 
If  there  are  some  preferred  states,  i.e.,  if  the  pi  are  not  equally 
probable,  then  it  can  be  shown  that  the  average  information  per 
message  can  range  from  zero  to  one  bit.  Zero  information  cor- 
responds to  the  condition  where  a  single  state  has  unity  probability 
and  all  the  other  states  have  probability  zero.  As  stated  before,  the 
other  extreme  is  attained  when  all  the  states  are  equally  probable. 
In  other  words,  on  the  average,  one  must  receive  more  information 
to  resolve  fully  the  states  of  a  completely  random  system  (all 
states  equally  probable)  than  to  resolve  the  states  of  a  less  random 
system  (all  states  not  equally  probable). 

COMMUNICATION  OF  INFORMATION 

Let  us  now  consider  information  theory  as  it  pertains  to  the 
communication  of  information.  For  this  purpose,  we  define  infor- 
mation received  as  the  diff'erence  between  the  state  of  knowledge 
of  the  recipient  before  and  after  the  communication.  In  more 
precise  terms,  information  received  is  given  by: 


where 


/  =  log 


I  =  information  received 
Pa  =  probability  of  the  event  at  the  receiver  after 

the  message  is  received 
Pb  =  probability  of  the  event  at  the  receiver  before 
the  message  is  received. 

In  receiving  a  message  regarding  the  sex  of  a  baby,  for  example, 
this  expression  implies  that  if  the  receiver  does  not  know  the 
baby's  sex,  then 

1 
Vb  =  2' 

and  if  you  (the  receiver)  receive  a  signal  that  "the  baby  is  a  boy," 
then 

Pa  =  I  (provided  the  message  is  not  noisy) 


1 2  Information  Storage  and  Neural  Control 

and,  therefore, 

/  =  log  jy^  =  log  2  =  1  bit. 

If  the  message  were  a  noisy  one,  then  you  might  not  be  quite 
certain  that  you  received  the  signal  for  "boy"  correctly.  You  may 
nevertheless  be  willing  to  give  four  to  one  odds  that  it  is  a  boy, 
based  on  the  noisy  signal  you  received.  In  this  case 

Va    =     .8 

and,  thus 

I  =  log  Yjx  =  log  1.6  =  .68  bits, 

1/  Z 

which  demonstrates  the  quantitative  reduction  in  information  due 
to  noise. 

In  the  case  of  no  noise  it  is  clear  tlie  pa  is  always  unity  and 

/  =  -log  pb. 

The  important  problems  in  tlie  communication  of  information  are, 
however,  concerned  with  the  effects  of  noise.  The  maximum 
amount  of  information  that  can  be  sent  through  a  communication 
channel  in  the  presence  of  noise  is  a  topic  of  particular  usefulness 
whicli  we  shall  examine  briefly. 
Getting  back  to  the  expression 

it  is  interesting  to  note  the  implications  of  this  definition.  If,  for 
example,  a  communication  system  is  so  noisy  that  the  message 
has  not  reduced  the  receiver's  uncertainty  as  to  the  event  (i.e., 
p^  =  p^)^  then  /  =  log  1=0  and  no  information  has  been  received. 
Thus,  it  is  seen  that  a  communication  does  not  necessarily  convey 
any  information.  The  communication  must  reduce  the  recipient's 
uncertainty  as  to  the  events  in  question  in  order  to  convey  infor- 
mation. The  mathematical  definition  I  =  log  Pa/Pb  is,  therefore, 
consistent  with  intuitive  requirements  for  a  measure  of  information. 
One  of  the  important  problems  in  communication  theory  has 
to  do  with  the  maximum  rate  at  which  information  can  be  sent 
over  a  communication  channel  which  is  disturbed  by  random 
noise.  This  problem  has  fundamental  implications  for  information 


What  is  Information  Theory?  13 

transfer  rates  in  biological  systems  as  well  as  for  the  neurophysio- 
logical  aspects  of  information  transfer,  which  are  to  be  treated  in 
later  papers  at  this  symposium.  In  order  to  discuss  this  problem, 
it  is  necessary  to  define  a  few  terms,  namely: 

B  =  the  bandwidth  of  the  communication  channel  (this 
defines  the  range  of  frequencies  which  can  pass  through 
a  system) 

S  =  received  effective  signal  power 
N  =  received  effective  noise  power. 

In  any  communication  system,  the  message  from  which  the 
recipient  derives  information  is  a  combination  of  signal  plus  noise. 
It  can  be  shown  (not  without  some  mathematical  difficulty, 
however)  that  the  maximum  rate  at  which  information  can  be 
sent  through  a  channel — which  is  1)  signal  power  limited  by  .S', 
and  2)  disturbed  by  random  noise  of  power  N — is  given  by 

R  =  B\og(l  +  S/N). 
In  other  words,  the  maximum  information  that  can  be  sent  in 
a  time  T  is  RT  or 

I  =  BT  log  (1  +  S/N). 
The  important  implication  of  these  formulae  in  the  design 
of  communication  systems  resides  in  the  fact  that  S/N,  the  signal- 
to-noise  ratio,  is  a  function  of  B,  the  bandwidth  of  the  channel. 
Therefore,  if  one  determines  the  dependence  of  signal-to-noise 
ratio  on  bandwidth,  it  is  possible  to  achieve  a  tradeoff  between 
S/N  and  B,  which  optimizes  the  information  handling  capacity 
of  the  system. 

EQUIVOCATION 

This  leads  us  to  the  more  involved  concepts  of  equivocation 
and  channel  capacity  and  to  Shannon's  basic  theorems  on  error 
correction.  The  previously  mentioned  maximum  rate  at  which 
information  can  be  sent  through  a  channel,  usually  referred  to 
as  the  channel  capacity  C,  is  intimately  related  to  these  ideas  and, 
therefore,  requires  some  elaboration  and  clarification. 

As  Shannon  has  stated,  it  may  seem  surprising  that  we  should 
define  a  definite  capacity  C  for  a  noisy  channel,  since  we  can  never 


14  Information  Storage  and  Neural  Control 

send  certain  {i.e.,  probability  equal  one)  information  over  such 
a  channel.  It  is  clear,  however,  that  by  sending  the  information 
in  a  redundant  form,  the  probability  of  errors  can  be  reduced. 
For  example,  by  repeating  the  message  many  times  and  by  a 
statistical  study  of  the  different  versions  of  the  message,  the  prob- 
ability of  errors  can  be  made  very  small.  One  would  expect, 
however,  that  to  make  this  probability  of  errors  approach  zero, 
the  redundancy  of  the  encoding  must  increase  indefinitely  and 
the  rate  of  transmission  must  therefore  approach  zero.  This  is  by 
no  means  true.  If  it  were,  there  would  not  be  a  well-defined 
capacity,  but  only  a  capacity  for  a  given  frequency  of  errors  or  a 
given  equivocation,  the  capacity  going  down  as  the  error  require- 
ments are  made  more  stringent.  Actually,  the  capacity  C  defined 
earlier  has  a  very  definite  significance.  It  is  possible  to  send  infor- 
mation at  the  rate  C  through  the  channel,  with  as  small  a  fre- 
quency of  errors  or  equivocation  as  desired,  by  proper  encoding. 
This  statement  is  not  true  for  any  rate  greater  than  C.  If  an  attempt 
is  made  to  transmit  at  a  higher  rate  than  C,  then  there  will  neces- 
sarily be  an  equivocation  equal  to  or  greater  than  the  excess. 

To  clarify  the  concept  of  equivocation,  let  us  suppose  there  are 
two  possible  symbols,  0  and  1,  and  that  we  are  transmitting  at  a 
rate  of  1,000  symbols  per  second  with  probabilities  Po  =  Pi  =  1/2. 
Thus,  our  source  is  producing  information  at  the  rate  of  1,000 
bits  per  second  (Shannon  refers  to  this  as  the  entropy  of  the 
source).  During  transmission,  noise  introduces  errors  so  that,  on 
the  average,  one  symbol  in  100  is  received  incorrectly  (a  0  as  1, 
or  1  as  0).  What  is  the  rate  of  transmission  of  information?  Cer- 
tainly less  than  1,000  bits  per  second  since  about  one  per  cent 
of  the  received  symbols  are  incorrect.  Our  first  impulse  might  be 
to  say  the  rate  is  990  bits  per  second,  merely  subtracting  the 
expected  number  of  errors.  This  is  not  satisfactory  since  it  fails 
to  take  into  account  the  recipient's  lack  of  knowledge  of  where 
the  errors  occur.  We  may  carry  this  to  an  extreme  case  and 
suppose  the  noise  so  great  that  the  received  signals  are  entirely 
independent  of  the  transmitted  signals.  The  probability  of  receiving 
1  is  one-half  whatever  was  transmitted,  and  the  same  is  true  for 
zero.  Since  about  one-half  of  the  received  symbols  are  correct  due 
to  chance  alone,  we  could  give  the  system  credit  for  transmitting 


What  is  Iiijormatwn  Theory?  15 

500  bits  per  second  while  actually  no  information  was  being 
transmitted  at  all.  Equally  good  transmission  would  be  obtained 
by  dispensing  with  the  channel  entirely  and  flipping  a  coin  at 
the  receiving  end. 

The  proper  correction  to  apply  to  the  amount  of  information 
transmitted  is  the  uncertainty  of  what  was  actually  sent  after  we 
have  received  a  signal.  This  reduction  in  received  information 
is  the  conditional  entropy  of  the  message  and  is  called  the  equivo- 
cation. It  measures  the  average  ambiguity  of  the  received  signal 
or,  in  other  words,  the  average  uncertainty  in  the  message  when 
the  signal  is  known.  For  definiteness,  let  us  calculate  the  equivo- 
cation of  the  first  example.  In  this  example,  noise  caused  an  error 
in  about  one  out  of  each  1 00  symbols,  so  that  if  a  zero  was  received, 
the  a  posteriori  probability  that  a  zero  was  transmitted  is  .99  and 
that  a  1  was  transmitted  is  .01.  The  equivocation,  or  the  uncer- 
tainty associated  with  each  symbol,  is  exactly  the  entropy  associated 
with  these  concHtional  probabilities.  Thus, 

Equivocation  per  symbol  =  —[.99  log  .99  +  .01  log  .01] 

=  .081  bits. 

Since  the  source  is  producing  information  at  a  rate  of  1,000 
bits  per  second,  the  equivocation  rate  is  1,000  X  .081  =81  bits 
per  second.  Therefore,  we  may  say  that  the  system  is  transmitting 
at  a  rate  of  1,000  —  81  =  919  bits  per  second.  Again,  in  the 
extreme  case  where  a  0  is  equally  likely  to  be  received  as  a  0  or  1 
and  a  1  as  a  1  or  0,  the  a  posteriori  probabilities  are  ,1/2  and  1/2, 

Equivocation  =  —  U^  log  ~y  -{-  7,  log  -^ 

=  1  bit  per  symbol, 

or  1,000  bits  per  second.  The  rate  of  transmission  is  then  zero  as 
it  should  be. 

These  examples  have  demonstrated  that  noise  causes  a  reduction 
in  received  information  and  have  shown  precisely  how  this  loss  in 
information  is  measured.  Before  leaving  this  subject,  I  would  like 
to  quote  a  theorem  (due  to  Shannon)  which  emphasizes  why  this 
quantitative  measure,  called  the  equivocation,  is  so  important. 

Shannon  has  shown  that  (in  a  noisy  communication  system) 
if  a  correction  channel  is  added  which  has  a  capacity  equal  to 


1 6  Information  Storage  and  Neural  Control 

the  equivocation  of  the  system,  then  it  is  possilole  to  encode  the  cor- 
rection data  so  as  to  send  it  over  this  channel  and  correct  all  but 
an  arbitrarily  small  fraction  of  the  errors.  This  is  not  possible 
if  the  channel  capacity  is  less  than  the  equivocation. 

Roughly  then,  the  equivocation  may  be  considered  as  the 
amount  of  additional  information  that  must  be  supplied  per 
second  at  the  receiving  point  to  correct  the  received  message. 

CONCLUDING  REMARKS 

This  paper  has  dealt  primarily  with  some  of  the  basic  aspects 
of  the  statistical  theory  of  information.  Very  few  comments  have 
been  made  regarding  semantic  information,  not  because  this  sub- 
ject is  unimportant,  but  rather  because  there  is  at  present  no 
sound  quantitative  theory  for  treating  semantic  information.  In 
concluding,  however,  I  would  like  to  remark  that  statistical  infor- 
mation theory  has  relevance  to  semantics  insofar  as  it  tells  us  what 
confidence  we  can  place  in  the  accuracy  of  the  information  we  re- 
ceive as  opposed  to  the  information  sent.  The  significance  or  value 
of  the  information  to  the  recipient  does  not  fall  within  the  domain 
of  the  quantitative  measures  provided  by  information  theory. 

With  reference  to  the  theme  of  this  symposium,  one  might  say 
that  information  theory  provides  insight  for  analyzing  and  im- 
proving storage  and  communication  processes,  but  does  not  unravel 
the  bewildering  complexities  associated  with  significance,  meaning, 
or  value  judgments.  From  my  personal  experience  with  the  prob- 
lems of  physiological  signal  analysis,  this  fact  lies  at  the  core  of 
the  difliculties  which  the  life  sciences  face  in  applying  information 
theory  to  their  problems.  Finding  significant  factors  in  a  maze 
of  statistical  information  is  an  immensely  challenging  problem  in 
medicine  as  well  as  in  many  other  fields.  The  problems  require 
both  an  intelligent  application  of  information  theory  and  a 
thorough  knowledge  of  the  phenomena  being  studied  so  that  good 
questions  can  be  asked  in  the  right  way  to  enhance  the  probability 
of  getting  a  useful  answer. 

REFERENCES 

1.   Bell,   D.   A.:   Information    Theory  and  Its  Engineering  Applications.   New 
York,  Sir  Isaac  Pitman  and  Sons,  Ltd.,  1956. 


What  is  Information  Theory?  17 

2.  Cherry,  Colin:  On  Human  Communication — A  Review,  A  Survey,  and  A 

Criticism.  Massachusetts  Institute  of  Technology,  Technology  Press, 
1957. 

3.  Feinstein,    Amiel:     Foundations    of    Irformation     Theory.    New    York, 

McGraw-Hill,   1958. 

4.  Gabor,  D.:  Lectures  on  Communication  Theory.  Massachusetts  Institute 

of  Technology  Research  Laboratory  of  Electronics,  Technical  Re- 
port No.  238,   1952. 

5.  Goldman,  Stanford:  Information  Theory.  Englewood  Cliffs,  New  Jersey, 

Prentice-Hall,  Inc.,  1952. 

6.  Khinchin,  A.  I.:  Mathematical  Foundations  of  Information  Theory.  New 

York,  Dover  Puljlications,  Inc.,  1957. 

7.  Schwartz,  Mischa:  Information  Transmission  Modulation  and  Noise.  New 

York,  McGraw-Hill,  1959. 

8.  Shannon,  Claude  E.,  and  Weaver,  Warren:  The  Mathematical  Theory 

of  Communications.  Urbana,  The  University  of  Illinois  Press,  1949. 

9.  Stumpers,  F.  L.  H.  M.:  Interpretation  and  Communication  Theory.  Lab- 

oratoria  N.  V.  Philips  Gloeilampenfabreiken,  Eindhoven,  Holland, 
1959. 
10.  Woodward,   P.  M.:  Probability  and  Information   Theory,    With  Applica- 
tions to  Radar.  New  York,  Pergamon  Pixss,  1953. 

DISCUSSION  OF  CHAPTER  I 

Heather  D.  Mayor  (Houston,  Texas):  In  case  one  wishes  to 
draw  analogies  from  physics  rather  than  from  thermodynamics, 
can  you  clarify  something?  Could  we  equate  your  conditional 
entropy  with,  say,  the  Heisenberg  uncertainty  principle  and  your 
noise  ratio  with  the  perturbations  introduced  in  measuring  the 
system? 

Bernard  Saltzberg  (Santa  Monica,  California) :  Yes,  in  terms  of 
inforination  measure,  uncertainty,  or  conditional  entropy,  and  the 
noise  which  gives  rise  to  the  uncertainty  (/.^.,  equivocation)  are 
aspects  of  essentially  ec^uivalent  ideas. 

Mayor:  And  the  Bohr  generalized  complementary  principle  in 
biological  systems — would  that  fit,  too,  with  your  intrinsic  con- 
cepts? For  example,  if  we  can  find  the  exact  position  of  a  micro- 
organism, it  is  difficult  at  the  same  time  to  establish  with  certainty 
another  parameter,  such  as  its  size.  In  a  biological  system,  would 
this  approach  fit  with  your  generalized  entropy  concept? 


1 8  Information  Storage  and  Neural  Control 

Saltzberg:  The  principles  involved  in  applying  generalized 
entropy  concepts  or  information  theory  to  biological  systems  and 
quantum  mechanical  systems  are  not  altered  in  essential  ways. 
There  are  differences  in  nomenclature  which  sometimes  conceal 
these  basic  similarities. 

Mayor:  But  could  you  treat  them  the  same  way? 

Saltzberg:  Yes.  In  fact,  all  of  these  applications  have  an  amaz- 
ingly close  parallel  to  the  generalized  treatment  of  entropy  in 
thermodynamics. 

Robert  R.  Ivers  (Fargo,  North  Dakota):  Would  you  define 
a  little  better  the  term  noise  that  you  used  during  your  discussion? 
Is  this  interference  with  signals,  or  is  it  the  interposing  of  randoin 
signals  in  the  systein,  or  is  it  just  general  inaccuracy  of  the  system? 

Saltzberg:  You  have  asked  a  very  basic  question.  In  order  to 
avoid  confusion,  I  should  like  to  refer  to  noise  as  a  subclass  of  a 
larger  class  of  signals  called  undesired  signals.  Undesired  signals 
may  be  placed  in  three  categories:  namely,  (a)  noise,  (b)  inter- 
ference, and  (c)  distortion.  Noise  signals  may  be  defined  as  signals 
which  are  not  coherent  with  any  signals  to  which  meaning  is 
assigned.  Interference  may  be  defined  as  an  undesired  signal 
which  is  a  desired  signal  in  some  other  system  or  is  coherent  with 
desired  signals  of  some  other  systein.  Examples  are  cross-talk  and 
common  channel  interferences  in  broadcast  programs.  Distortion 
introduces  undesired  signals  due  to  effects  such  as  non-linearities 
or  non-flat  amplitude  vs.  frequency  transmission  characteristics. 
In  my  discussion  I  have  been  referring  to  the  first  of  these  undesired 
signals,   namely,  random  noise. 

Mayor:  In  actually  performing  measurements  on  your  system, 
you  no  doubt  introduce  additional  perturbations  which  could  be 
considered  as  "noise."  Would  you  consider  this  a  valid  parallel? 

Saltzberg:  This  sort  of  parallel  seems  reasonable  to  me.  If  you 
add  an  element  of  indeterminacy  to  the  state  of  the  system,  you  may 
consider  this  as  due  to  noise.  There  are  some  deep  questions  as 
to  what  constitutes  noise  in  systems,  and  these  cannot  be  treated 
in  cjualitative  terms  or  in  brief  comments. 

Arthur  Shapiro  (New  York,  New  York) :  Along  the  same  line, 
how  would  you  treat  what  happens  if  you  read  onto  a  transmission 
line  a  page  from  a  table  of  random  numbers  and  then  another 


What  is  Information  Theory?  19 

page  from   a   table   of  random   numbers?   Are   you   transmitting 
information,  and  how  would  you  determine  how  much? 

Saltzberg:  Yes.  You  are  always  transmitting  information  when- 
ever you  convey  a  message,  unless  the  noise  is  so  great  that  the 
equivocation  of  the  system  is  equal  to  the  information  content  of 
the  source.  Your  question  apparently  refers  to  tlie  importance  of 
the  information.  A  table  of  random  numbers  may  be  useless  to 
the  receiver,  but,  nevertheless,  statistical  information  has  been 
communicated . 

Shapiro:  Then  it  is  not  really  true,  as  you  started  out  by  saying, 
that  the  meaning  of  what  you  transmit  has  nothing  to  do  with  how 
much  information  is  transmitted.  The  meaning  apparently  has  a 
great  deal  to  do  with  how  much  information  is  transmitted. 

Saltzberg:  Apparently  I  have  caused  some  confusion.  Seman- 
tic meaning  or  the  importance  of  a  message  is  subjective  and  is 
not  part  of  statistical  information  theory.  The  previously  used 
example  applies  here.  A  message  announcing  the  birth  of  a  boy 
conveys  one  bit  of  information  to  an  unknowing"  receiver  inde- 
pendent of  whether  the  receiver  is  the  father  or  not.  Thus,  whether  a 
number  is  taken  from  a  table  of  random  numbers  or  a  table  of 
trigonometric  functions  has  no  bearing  on  the  information  received, 
providing  the  receiver  has  no  a  priori  knowledge  of  these  numbers. 

Walter  Abbott  (Houston,  Texas):  The  point  has  been  made 
that  the  information  content  of  any  datum  is  proportional  to  its 
surprise  value.  Does  this  get  involved  in  your  semantic  implications? 

Saltzberg:  Surprise,  as  used  in  this  context,  does  not  have  any 
semantic  implications.  If  a  datum  or  a  message  identifies  one  of  a 
thousand  possible  states,  then  it  has  surprise  value  in  the  sense  that 
you  would  have  been  extremely  surprised  to  have  guessed  the  state 
without  receipt  of  the  information  provided  by  the  message.  If  a 
system  had  only  two  possible  states,  then  you  would  not  be  so  sur- 
prised to  guess  the  correct  state. 

Mary  A.  B.  Brazier  (Los  Angeles,  California) :  I  believe  that 
by  surprise  value  Dr.  Abbott  means  an  event  of  low  probability. 

Myron  F.  Weiner  (Dallas,  Texas) :  How  much  must  be  known 
of  the  probabilities,  or  of  the  number  of  probabilities  of  different 
messages,  or  of  the  number  of  possible  different  messages  to  be 
conveyed   before  one  can  get  some  idea  of  what  a  message  is, 


20  Information  Storage  and  Neural  Control 

providing    one    has    previously    had    no    information    about    the 
system? 

Saltzberg:  If  you  knew  nothing"  about  the  probability  states  of 
the  messages,  then,  of  course,  you  would  have  very  little  engi- 
neering data  upon  which  to  base  an  optimum  design  for  a  receiving 
system.  This  question  may  pertain  to  the  a  priori  probabilities 
which  are  useful  in  choosing  an  appropriate  code.  This  is  analogous 
to  the  Twenty  Question  game  mentioned  previously.  If  the  ques- 
tioner has  some  a  priori  knowledge  of  the  probabilities,  he  can  ask 
questions  in  a  specific  order,  depending  on  the  probabilities,  and, 
on  the  average,  will  ask  fewer  questions  to  get  a  correct  answer 
than  will  someone  who  just  asks  questions  at  random. 

E.  Roy  John  (Rochester,  New  York):  It  seems  to  me  that 
there  is  a  large  class  of  messages  in  which  the  a  priori  probability 
cannot  be  evaluated  by  the  receiver.  One  can  think  of  messages 
in  which  the  rate  of  convergence  of  the  total  information  of  the 
message  is  not  linear  for  the  components  of  the  message  and  in 
which  the  rate  of  convergence  would  depend  upon  the  sequence 
of  the  components.  This  might,  as  a  matter  of  fact,  be  a  charac- 
teristic difference  between  certain  languages.  In  a  situation  where 
you  do  not  have  this  advantage  of  being  able  to  stipulate  prob- 
abilities— in  which  the  probability  of  a  given  event  is  affected  by 
the  preceding  sequence — it  seems  to  me  you  must  modify  your 
treatment  to  provide  an  argument  for  the  bit  function,  recog- 
nizing that  the  information  content  of  a  specific  event  depends  on 
preceding  events  or  context.  Could  you  say  something  about  how 
you  treat  this  kind  of  situation,  since  it  seems  to  be  much  nearer 
the  situation  in  which  we  frequently  find  ourselves  in  the  nervous 
system  than  does  the  starting  point  from  which  you  began  here. 

Saltzberg:  Your  question  is  a  good  one.  It  refers  to  the  effects 
of  inter-symbol  influence  on  information  content.  The  fact  that 
there  are  transitional  probabilities  which  have  to  be  taken  into 
account  in  determining  the  information  content  in  language,  for 
example,  is  included  in  the  mathematics  of  information  theory. 
These  transitional  probabilities  have  the  effect  of  making  the 
information  content  of  a  sentence  much  less  than  that  calculated 
by  assuming  that  the  sequences  of  letters  and  words  are  inde- 
pendent of  their  predecessors.  I  should  comment  at  this  point  on 


What  is  Information  Theory?  21 

another  aspect  of  information  theory  which  I  have  not  mentioned 
before.  Information  theory  is  concerned  with  the  properties  of 
ensembles  of  messages  or  objects.  One  of  the  properties  which  is 
quite  important  in  scientific  analysis  is  known  as  ergodicity. 
In  analyzing"  many  problems,  the  assumption  of  ergodicity  is  one 
that  is  a  practical  necessity  rather  than  a  statement  of  fact  relative 
to  the  nature  of  things.  However,  this  simplifies  analysis  in  that 
it  allows  one  to  examine  a  long  time  sample  of  one  of  the  mem- 
bers of  an  ensemble  and  to  conclude  from  this  that  he  knows 
something  about  the  statistics  of  the  ensemble.  This  is  not  always 
true  since,  for  example,  it  would  imply  that  the  statistics  associated 
with  the  EEG  record  of  a  single  individual  apply  equally  well  to 
another  subject.  If  this  were  the  case,  then  an  ensemble  of  messages 
composed  of  the  EEG's  of  many  subjects  would  be  an  ergodic 
ensemble.  In  testing  engineering  components,  one  ordinarily  takes 
a  single  component  and  tests  it  for  a  long  period  of  time  and  then 
draws  implications  about  the  behavior  of  all  similar  components. 
This  is  an  aspect  of  statistical  analysis  and  information  theory 
which,  when  applied  to  the  life  sciences,  creates  a  great  many 
problems  since  one  may  not  be  aware  that  this  assumption  may 
underly  the  mathematical  formulation  of  certain  problems. 

Herman  Blustein  (Chicago,  Illinois):  How  do  you  determine 
the  validity  of  the  samples  when  you  analyze  the  EEG's  in  this 
manner  and  make  a  generalization  from  them? 

Saltzberg:  The  validity  of  the  sample  is  not  the  question  here. 
For  example,  an  EEG  record  may  be  sufficiently  long  to  give  you 
a  good  estimate  of  its  properties  for  a  particular  individual.  How- 
ever, unless  EEG's  of  different  individuals  are  statistically  similar, 
or  ergodic,  this  does  not  allow  you  to  draw  any  conclusions  about 
the  properties  of  another  individual's  EEG  record. 

Harold  W.  Shipton  (Iowa  City,  Iowa) :  The  way  the  discus- 
sion is  going  means,  I  think,  that  we  have  to  say  a  little  more 
about  the  properties  of  noise,  because  when  we  deal  with  formal 
information  theory  we  use  "noise"  in  exactly  the  way  that  we 
used  to  use  the  phrase  "Brownian  movement."  This  is  quite 
acceptable.  However,  when  we  perform  an  experiment,  we  are 
dealing  with  band  limited  noise,  and  we  are  also  probably  dealing 
with  nonrandom  perturbations  in  the  system.  I  would  like  to  hear 


22  Information  Storage  and  Neural  Control 

from  Dr.  Saltzberg  whether  he  wishes  to  introduce  a  second 
term — noise  in  this  physical  sense — or  whether  he  would  also  Hke 
to  consider  things  wliich  are  not  related  to  the  required  signal 
over  a  short-time  epoch.  There  is  a  good  example  of  tliis  in  the 
field  of  EEG  analysis.  If  you  repeat  an  experiment  in  time,  you 
expect  the  signal-to-noise  ratio  to  go  up  as  -^/N  ,  but  in  almost 
any  biological  system  you  will  find  it  goes  up  by  rather  more  than 
this  simply  because  our  noise  is  not  "noisy,"  so  to  speak,  in  the 
sense  that  it  is  not  white. 

Saltzberg:  There  are  many  things  which  people  refer  to  as 
noise  that  are  quite  diff"erent  from  one  another.  The  different 
types  of  noise  have  considerably  different  effects  on  the  informa- 
tion content  of  systems.  For  example,  there  is  distortion  which, 
if  reversible,  does  not  reduce  the  information  content  of  a  message 
at  all.  Although  people  commonly  refer  to  this  type  of  distortion 
as  noise,  it  is  not  noise  in  the  context  of  information  theory.  You 
have  mentioned  white  noise,  which  is  a  special  type  of  random 
noise,  and  the  chscussion  on  maximum  rate  of  transmission  of 
information  in  the  presence  of  noise  is  applicable  to  this  lype  of 
noise.  It  is  important  to  distinguish  between  this  type  of  noise 
and  interference,  which  is  sometimes  referred  to  as  noise.  The 
basic  difference  is  that  random  noise  is  not  coherent  with  any 
signals  to  which  meaning  is  assigned,  while  interference  is  an 
undesired  signal  which  is  coherent  with  desired  signals  of  some 
otlier  system.  The  improved  signal-to-noise  ratios  that  you  men- 
tioned for  biological  systems  may  have  something  to  do  with  the 
ability  of  biological  systems  to  narrow  their  noise  bandwidths  by 
providing  certain  kinds  of  adaptive  filtering. 

Blustein:  Is  this  similar  to  a  television  signal  in  which  the 
audio  signal  is  intact  and  the  video  is  distorted,  and  yet  one  can 
receive  and  interpret  the  signal? 

Saltzberg:    I  am  not  sure  of  the  analogy.  I  have  to  beg  off  on  this. 

Blustein:  Does  the  system  have  to  filter  the  signals? 

Saltzberg:  If  the  receiver  has  some  a  priori  knowledge  of  what 
it  is  looking  for,  then  it  can  do  an  excellent  job  of  minimizing  the 
effects  of  noise.  One  of  the  simple  ways  this  is  accomplished  is 
by  means  of  frequency  filters  or  correlators.  If  the  information 
signals  occupy  a  narrow  bandwidth,  then  narrowing  the  accept- 


What  is  Information  Theory?  23 

ance  band  of  the  system  by  employing  filters  will  reduce  the 
amount  of  random  noise,  since  random  noise  occupies  all  parts 
of  the  spectrum.  The  object  is  to  use  spectrum  space  for  the  signal 
information,  not  the  noise. 

Shapiro:  Are  there  actually  two  kinds  of  information  involved 
here,  only  one  of  which  is  treated  in  this  way?  Perhaps  I  should  not 
use  the  word  information  for  the  other  kind,  but  a  priori  knowledge 
about  the  nature  of  the  system  which  the  receiver  may  have, 
whether  it  is  a  person  or  a  machine,  must  be  important.  For 
example,  if  the  machine  or  the  person  knows  that  nothing  is 
coming  over  this  channel  except  when  some  kind  of  an  event 
occurs,  then  when  a  lot  of  noise,  i.e.,  a  lot  of  signals,  comes  over, 
this  will  be  interpreted  as  meaning  a  lot  of  activity  in  the  trans- 
mitter. On  the  other  hand,  if  the  receiver  knows  from  past  experi- 
ence that  this  system  generates  its  own  noise,  then  when  a  lot  of 
noise  comes  over  the  channel,  the  receiver  says  it  does  not  know 
what  is  going  on.  Only  when  a  clear  individual  signal  comes  over 
will  it  be  interpreted  as  information.  I  think  that  there  is  another 
set  of  values,  which  I  suspect  is  what  you  mean  by  filter  theory. 

Saltzberg:  I  believe  your  comment  relates  to  filtering  in  the 
time  domain.  It  is  possible  to  design  a  system  which  simply  stops 
processing  information  when  the  signal  is  too  badly  corrupted  by 
noise.  If  cues,  which  are  essentially  a  priori  information,  are  avail- 
able, then  it  is  possible  to  use  various  types  of  time  filtering.  When 
you  do  not  have  any  a  priori  timing  cues  for  determining  when 
you  ought  to  process  information,  then  frequency  filtering  is  sug- 
gested, providing  you  know  something  about  the  spectral  region 
which  the  signals  occupy. 

Peter  Kellaway  (Houston,  Texas):  It  would  help  if  you  could 
tell  us  something  about  your  ow^n  results.  I  understand  you  are 
interested  in  analyzing  the  EEG.  What  sort  of  information  have 
you  obtained  by  applying  information  theory  and  technique  to 
this  type  of  analysis,  and  what  type  of  information  do  you  hope 
to  obtain? 

Saltzberg:  I  can  comment  on  some  of  the  analysis  of  EEG 
which  was  conducted  in  an  attempt  to  establish  how  much  infor- 
mation is  processed  by  one  technic^ue  of  EEG  analysis  as  compared 
with  the  amount  of  information  which  is  processed  using  another 


24  Information  Storage  and  Neural  Control 

technique  of  EEG  analysis.  The  objective  of  this  investigation  was 
to  evaluate  the  information  handling  capability  of  zero  crossing 
analysis  as  compared  to  frequency  analysis.  Conventional  fre- 
quency analyzers  were  compared  on  an  information  theoretical 
basis  with  the  zero  crossing  analyzers  which  Dr.  Burch  of  Baylor 
is  using  in  his  EEG  research.  The  particular  problem  treated  was 
concerned  with  how  much  information  is  abstracted  by  an  analyzer 
which  processes  only  time  point  data  associated  with  zero  crossings 
of  the  record  and  the  time  positions  of  its  peaks  and  points  of  inflec- 
tion. Knowing  the  time  resolution  which  could  be  achieved  with 
an  analyzer  of  this  type,  it  was  possible  to  establish  the  frequency 
resolution  that  would  be  required  of  a  frequency  analyzer  in  order 
to  provide  the  same  amount  of  information.  I  will  admit  that 
this  does  not  add  much  understanding  of  neurophysiological 
processes,  but  it  does  give  a  basis  for  some  engineering  decisions 
relative  to  whether  one  type  is  more  effective  in  abstracting 
information  than  another  type.  It  turned  out  that  the  resolution 
that  could  be  achieved  with  practical  period  analyzers  was  much 
greater  than  that  which  could  be  achieved  with  practical  fre- 
quency analyzers. 

Kellaway:  Would  it  matter  if  it  were  a  physiological  signal? 
The  signal  that  you  are  using  could  be  anything,  could  it  not? 

Saltzberg:  The  signal  could  well  be  anything.  However,  what 
one  would  like  to  achieve  is  a  signal  representation  and  a  cor- 
responding form  of  analysis  emphasizing  the  physiological  effects. 
For  example,  if  one  attempts  to  extract  information  by  analyzing 
the  coefficients  of  a  Fourier  series,  the  problem  may  be  impossible 
because  the  interesting  infoi^mation  is  contained  in  small  per- 
turbations in  many  amplitudes  of  many  frequencies.  Since  similar 
small  perturbations  can  be  caused  by  noise,  the  presence  of  noise 
would  invalidate  any  physiological  correlates  being  sought.  How- 
ever, if  some  other  parameter  associated  with  a  different  signal 
representation  were  measured,  it  is  possible  that  physiological 
effects  could  cause  this  parameter  to  change  grossly,  which  would 
lead  to  a  higher  probability  of  valid  pliysiological  correlates. 

Max  E.  Valentinuzzi  (Atlanta,  Georgia):  Gan  you  say  any- 
tliing  about  the  relation  between  information  and  organization? 
Gan  we  say  that  these  two  words  are  equivalent?  You  have  dealt 


What  is  Information  Theory?  25 

with  transmission  of  information  from  one  point  to  another. 
Suppose  that  now  we  are  not  transmitting  information,  but  are 
organizing  a  set  of  elements  in  a  particular  pattern  or  configura- 
tion. What  is  the  amount  of  information  obtained  by  the  system 
in  the  transition  from  one  state  to  the  other? 

Saltzberg:  Whether  we  talk  about  the  entropy  of  thermo- 
dynamics or  the  information  in  a  message,  we  are  in  principle 
talking  about  organization,  or,  more  precisely,  about  the  prob- 
abilities of  the  various  arrangements  of  the  component  parts  of 
the  system.  The  second  law  of  thermodynamics  states  that  entropy 
must  increase  or,  at  best,  remain  constant,  which  is  another  way 
of  saying  that  the  system  is  becoming  more  disorganized  or  ran- 
dom. In  communication  systems,  however,  upon  the  receipt  of 
information,  the  disorganized  or  uncertain  state  of  our  knowledge 
becomes  more  certain  or  better  organized;  therefore,  we  may 
consider  received  information  as  negative  entropy  since  it  increases 
the  organization  of  the  receiver. 

Gregory  Bateson  (Palo  Alto,  California) :  You  separated  rather 
clearly  the  notion  of  measuring  cjuantity  of  information  from  the 
notion  of  ""meaning''  of  the  information  measured,  but  it  appears 
to  me  that  this  becomes  difficult  when  we  have  a  secjuence  of 
items  comprising  a  total  message  and  so  related  that  some  of  these 
items  reflect  upon  the  significance  of  other  items  in  the  sequence. 
In  this  case,  the  meaning  of  these  meta  signals  is  a  very  important 
part  of  the  whole  economics  of  communication. 

Saltzberg:  I  think  your  question  refers  to  the  very  strong  con- 
straints which  may  exist  between  the  elements  of  a  signal.  These 
constraints  are  essentially  the  transition  probabilities  and  the 
intersymbol  influences  which  are  referred  to  in  information  theory. 
We  do  not  look  upon  a  knowledge  of  these  transition  probabilities 
as  implying  meaning  in  the  sense  that  we  talk  about  semantic 
meaning.  In  other  words,  the  constraints  between  a  sequence  of 
elements  comprising  a  signal  are  accounted  for  in  measuring 
information,  but  the  importance  of  the  signal,  i.e.,  its  semantic 
value,  is  not. 

Bernard  S.  Patrick  (Memphis,  Tennessee):  Can  you  speak 
for  just  a  moment  about  redundancy  or  the  use  of  redundancy  in 
communication  systems? 


26  Information  Storage  and  Neural  Control 

Saltzberg:  In  any  communications  system  which  places  a  very 
high  priority  on  accuracy,  redundancy  is  frequently  employed. 
In  situations  where  it  is  not  practical  to  increase  the  strength  of 
the  signal  in  order  to  improve  accuracy  by  increasing  signal-to- 
noise  ratio,  it  becomes  important  to  use  redundancy  to  correct 
the  errors  due  to  noise.  In  digital  communication  systems,  re- 
dundancy is  often  employed  in  the  form  of  error-correcting  codes. 
The  accuracy  of  language  communication  is  greatly  enhanced  by 
the  redundancy  of  language.  It  is  easy  to  get  a  qualitative  feeling 
for  the  amount  of  redundancy  in  the  English  language.  For 
example,  consider  a  situation  in  which  a  sentence  composed 
of  a  sequence  of  symbols  is  transmitted  and  the  first  symbol  re- 
ceived is  a  t,  the  second  a  k,  and  the  third  an  e.  You  would 
have  no  trouble  concluding  that  the  word  transmitted  is  "the" 
because  of  the  constraints  in  the  language.  The  language  structure 
tells  us  that  "tke"  must  be  in  error  since  "tke"  is  not  a  word. 
Further,  since  h  very  frequently  follows  t  in  the  English  language, 
there  is  not  much  doubt  that  the  first  word  in  the  sentence  is 
"the."  This  is  one  aspect  of  how  the  redundancy  of  the  language 
increases  the  accuracy  of  communication.  In  fact,  in  communica- 
tion systems  which  are  signal  power  limited,  it  is  necessary  to 
employ  redundancy  techniques  to  reduce  the  errors  in  com- 
munication due  to  noise. 

Ralph  W.  Gerard  (Ann  Arbor,  Michigan):  It  was  said  that 
maybe  an  example  would  illuminate  some  of  the  points  that 
would  come  up,  and  I  think  I  can  give  one  that  might  be  helpful. 
Limiting  the  spectrum  in  frequency  or  the  interval  of  time  in 
which  the  significant  signal  is  to  be  expected  is  helpful.  Gregory 
Bateson  referred  back  to  that  in  speaking  of  a  para-signal,  which 
tells  the  meaning  of  the  signal  itself  by  indicating  where  in  the 
total  space  you  must  look  for  the  signal.  I  wonder  how  many  in 
the  room  will  understand  the  statement,  ^^  Cur  antrum  santrum, 
ovidiim,  ovidum.''  Hands?  None.  That  is  because  you  thought  I  was 
talking  Latin,  whereas  I  was  actually  talking  German.  Now  I 
will  repeat  it  the  same  way.  "'/Tz//?  rant  rum,  Sand  rum,  ohwiedumm,  oh 
wie  dummy  Knowing  that  the  language  is  German  is  the  para- 
signal;  locating  the  message  in  the  total  space  is  what  is  meant  by 
having  the  proper  set. 


I 


CHAPTER 
II 

BINARY  REPRESENTATION  OF 
INFORMATION 

Robert  T.  Gregory,  Ph.D. 
INTRODUCTION 


T  IS  well  known  that  most  modern  electronic  digital  computers 
use  the  binary  representation  of  numbers  internally  rather  than 
the  more  familiar  decimal  representation,  although  sophisticated 
programming  systems  may  allow  the  computer  user  to  do  almost 
all  of  his  communicating  with  the  machine  in  decimal.  The 
reason  for  the  fact  that  binary  representation  is  in  common  use 
is  explained  in  the  next  section. 

It  is  tiie  purpose  of  this  paper  to  review  the  binary  representation 
of  numbers,  including  the  word  structure  for  a  typical  binary 
computer,  and  to  demonstrate  some  typical  machine  commands 
that  are  available  for  manipulating  patterns  of  binary  digits.  It 
is  hoped  that  this  will  provide  some  indication  of  the  extreme 
versatility  of  the  electr'onic  computer  as  an  information  processing 
instrument  and  will  encourage  those  who  have  not  yet  discovered 
its  usefulness  to  explore  its  potentialities. 

REPRESENTING  NUMBERS  INSIDE  A  COMPUTER 

It  is  well  known  to  those  who  design  the  basic  circuits  for 
digital  computers  that  the  optimum  number  base,  B,  for  representing 
numbers  inside  a  computer,  from  the  standpoint  of  economy  of 
electronic  hardware  needed,  is,  B  =  3.  To  verify  this  let  us  recall 
that  the  number  of  numbers  that  can  be  expressed  using  n  digits, 
base  B,  is  B".  For  example,  in  the  decimal  numeral  system  if 
n  =  3,  we  can  express  10^  numbers  000,  001,   .  .  .,  999. 

27 


28  Injormation  Storage  and  Neural  Control 

Assume  that  the  number  of  electronic  components  required  to 
represent  a  single  digit,  base  5,  is  approximately  proportional  to 
B.  Thus,  a  rough  estimate  of  the  number  of  electronic  components 
required  to  represent  n  digits,  base  B,  is 

TV  =  n  lu  B, 
where  A"i  is  a  constant. 

If  B"  =  P  is  held  fixed,  and  we  wish  to  find  the  value  of  B 
which  minimizes  the  number  of  electronic  components,  N,  needed 
to  represent  P  numbers,  we  proceed  as  follows:  Since 

i?"  =  P, 


we  can  write 


and  so 


n  In  B  =  In  P 
=  /v%. 


cJN  _  „ 
dB  -  ^^' 


In  B  • 
Thus,  the  number  of  electronic  components  needed  is 

KsB 

In  B  • 

DifTerentiating  with  respect  to  B  yields 

InB  -  r 

.   (In  BY'  _■ 

Setting  this  derivative  equal  to  zero  gives  us 

biB  =  1, 
or 

B  =  r 

=  2.71828  .... 

For  integer  values  of  B  the  minimum  occurs  when  B  =  3,  with 
slightly  greater  values  for  B  =  2. 

Since  tristable  devices  are  almost  nonexistent,  and  bistable  devices 
are  plentiful,  most  engineers  choose  the  binary  numeral  system 
rather  than  the  ternary  numeral  system  when  they  design  a 
machine. 


Binary  Representation  of  Information 


29 


BINARY  NUMBER  REPRESENTATION 

If  we  recall  the  definition  of  our  standard  positional  notation 
whereby  the  meaning  of  a  digit  depends  on  its  position  relative 
to  other  digits  in  the  number  representation,  then  we  note  that 
any  positive  integer  may  be  written 

(/„  .  .  .  r/.//ir/o  =  r/o/i"  +  (JiB'  +  f/,/i'  +  .  .  .  +  daB" 
where  i:?  >  1  is  the  base  of  the  number  representation  and  where 
0  ^d,<  B. 

For  example,  if  Z?  =  10  the  integer  "fifty-seven"  may  be  written 
57  =  7  .  10'  +  5  .  kV 
-  7  +  50. 
\i  B  =  1  then  "fifty-seven"  becomes 

Llll0011,,vo  =  [2'^  +  2' +  2' +  2i,,.„ 
=  [1  +  8  +  1(>  +  32],,„ 

where  the  subscript  "two"  indicates  that  binary  notation  is  used 
on  the  left  of  the  equal  sign  and  the  subscript  "ten"  indicates 
that  decimal  notation  is  used  on  the  right. 

Similarly,   any  positive  fraction   (less  than  one)   may  be  written 

OV/_if/_or/^3  .  .  .  d-,n  =  (UB-'  +  r/_o^-'  +  d.^B'^  +  .  .  .  +  r/_„,fi""', 
where  m  does  not  have  to  be  finite  and  0  ^  «'_,  <  B.  For  example, 
if  /?  =  2  then  fi\e-sixteenths  becomes 


[0.0101]uvo  = 


2"'  +  2- 
1  +  i,'       . 

.4  IbJten 


Since  positive  numbers  may  be  decomposed  into  an  integer  part 
and  a  fraction  part  we  may  consider,  as  a  more  general  example, 
the  binary  representation  of  thirty-seven  and  nine  sixty-fouiths. 

[I00101.0010011t„,,  =  [2"  +  2'  +  2' 


,„  +  [2-^  +  2^^J 


[1  +  4  +  82],,„  + 


1+^ 
L8       ()4J 


[37]ten  + 


!) 
L()4jt 


It  is  not  our  intention  to  go  into  great  detail  at  this  point  and 
discuss  the  procedures  for  converting  from  one  number  representa- 
tion to  another.  We  have  merely  tried  to  review,  by  means  of 


30  Information  Storage  and  Neural  Control 

three  examples,  the  well-known  fact  that  positive  numbers  are 
easily  represented  by  a  pattern  of  binary  digits,  that  is  to  say, 
by  a  pattern  of  "zeros"  and  "ones."  (Binary  digits  are  commonly 
called   bits,  for  short.) 

Before  continuing,  it  is  necessary  to  recall  that  negative  numbers 
also  have  representations  in  terms  of  a  pattern  of  bits.  To  demon- 
strate this  let  us  discuss  methods  for  negative  number  representa- 
tion inside  a  computer.  The  following  systems  are  currendy  used: 
[1]  Signed  absolute  values 

[2]   Complements  with  respect  to  some  integral  power  of  the 
base 

[3]   Complements  with  respect  to  one  less  than  some  integral 
power  of  the  base. 

The  first  method  is  simple — the  machine  contains  the  absolute 
value  of  each  number  stored,  with  an  indication  of  its  sign.  The 
second  and  third  methods  involve  number  representation  modulo 
B''  and  modulo  (^^-1),  respectively,  where  B  is  the  number  base, 
and  the  machine  registers  are  assumed  to  hold  k  digits. 

To  illustrate  system  [2],  let  /:  =  9  and  assume  a  decimal  machine. 
Thus,  if  we  use  the  symbol  =  to  mean  "is  represented  by,"  then 

126  «  000  000  126 
and 

-126  «  999  999  874, 
since 

-126  =  999  999  874  (mod  10'). 
Using  system  [3]  we  have 

-126  «  999  999  873, 
since 

10'  -  1    =  999  999  999, 
and 

-126  =  999  999  873  (mod  999  999  999). 

System  [3]  is  sometimes  called  the  "nines  complement"  system 
when  B  is  ten.  This  is  motivated  by  the  fact  that  one  merely  takes 
a  digitwise  complement  with  respect  to  nine  in  forming  999  999  873 
as  the  negative  representation  of  126.   System   [2]  is  called  the 


Binary  Representation  of  Information  31 

"tens  complement"  system  when  B  is  ten,  since  the  least  significant 
digit  is  actually  complemented  with  respect  to  ten. 

In  a  nine-digit  binary  machine  using  "ones  complements,"  we 
would  have 

[126]ten    «    001    111    no, 

and 

[-126]ten    «    110  000  001. 

Thus  we  have  reviewed,  by  means  of  examples,  how  both  positive 
numbers  and  negative  numbers  may  be  represented  easily  by  a 
pattern  of  bits. 

PROCESSING  BINARY  INFORMATION 
INSIDE  A  COMPUTER 

Let  us  consider  a  typical  modern  high-speed  computer  which 
is  designed  to  process  binary  information  in  blocks  of  48  bits. 
Such  blocks  are  called  words,  and  we  describe  such  a  computer 
as  having  a  48  bit  word  length.  These  words  may  be  bit  patterns 
representing  numbers  (we  discussed  binary  representations  of 
numbers  in  the  previous  section)  or  the  words  may  be  bit  patterns 
having  a  non-numerical  interpretation.  To  the  machine  this  is 
immaterial. 

The  repertoire  of  machine  commands  for  carrying  out  operations 
on  machine  words  includes  commands  for  performing  the  basic 
arithmetic  operations  of  addition,  subtraction,  multiplication,  and 
division.  More  complicated  mathematical  tasks,  such  as  the  ex- 
traction of  square  roots,  solving  algebraic  equations,  and  so  on, 
are  accomplished  by  using  a  combination  of  these  basic  commands. 

In  addition  to  the  commands  for  performing  basic  arithmetic 
operations,  the  machine  is  capable  of  executing  commands  which 
perform  operations  of  a  non-numerical  character.  These  are  the 
commands  that  make  the  modern  electronic  digital  computer  a 
versatile  information-processing  instrument  rather  than  just  a 
high-speed  computing  instrument. 

We  begin  our  discussion  of  non-numerical  type  commands 
(although  some  of  these  may  have  a  numerical  interpretation 
as  well)  by  mentioning  the  shift  cortimands.  Consider  the  bit  pattern 
consisting  of  ones  in  the  odd  numbered  positions  and  zeros  in 


32 


Information  Storage  and  Neural  Control 


the  even  numbered  positions.  We  shall  write  the  word  in  the  form 

101010  .  .  .   101010, 

where  the  meaning  of  the  three  dots  is  obvious,   and   their  use 
enables  us  to  avoid  writing"  all  48  bits. 

A  left  shift  of  n  bits  causes  the  individual  digits  to  be  shifted 
to  the  left  n  places  in  an  "end-around"  fashion,  which  means 
that  bits  shifted  off  the  left  end  are  carried  around  and  introduced 
into  the  right  end  of  the  word.  Thus,  if  n  is  an  even  integer,  the 
pattern  displayed  above  will  not  appear  to  have  changed  following 
the  shift.  On  the  other  hand,  if  n  is  odd,  the  pattern  will  have  the 
appearance 

010101   .  .  .  010101 
following  the  shift. 

This  shifting  operation  can  be  quite  useful.  To  illustrate  this 
we  need  to  mention  that  the  machine  is  capable  of  performing 
branchmg  operations,  i.e.,  it  can  be  made  to  do  one  thing  if  the  first 
bit  of  a  word  is  a  "one"  and  another  thing  if  the  first  bit  is  a 
"zero."  This  means  that  the  machine  is  capable  of  performing 
each  of  two  sequences  of  operations  depending  on  the  nature  of 
the  first  bit  of  a  word.  Figure  1  will  aid  us  in  this  discussion. 


Fig.  1 


If  lines  represent  sequences  of  operations  then  we  traverse  the 
path  AB  if  the  first  bit  of  our  word  is  a  "one"  and  AC  if  the  first 
bit  is  a  "zero."  If  we  assume  that  we  shall  begin  at  point  A  many 
times  and  if  we  desire  to  traverse  the  paths  AB  and  AC  on  alter- 
nate occasions,  then  we  can  use  the  word 


101010 


101010 


and  the  left  shifting  operation  to  do  this.  All  we  need  to  do  upon 
arrival  at  either  of  the  points  B  or  C  is  to  command  the  machine 


Binary  Representation  of  Information 


33 


to  perform  an  odd  number  of  left  shifts.  This  will  cause  the  first 
bit  of  our  word  to  be  alternately  "one"  and  "zero." 

Other  examples  of  useful  commands  include  commands  to  per- 
form several  logical  operations.  In  order  to  describe  a  few  of  these 
commands,  reference  will  be  made  to  Table  I. 

TABLE  I 

Logical  Operations 


O  Logical  Product 
10  1  =  1 
10  0  =  0 
0  0  1  =  0 
0  0  0  =  0 


©  Logical  Sum 
1  ®  1  =  1 
1  ©  0  =  1 
0  ©  1  =  1 
0  ©  0  =  0 


®  Exclusive  "Or' 
1  ®  1  =  0 
1  ®  0  =  1 
0  ®  1  =  1 
0  ®  0  =  0 


For  example,  if  we  have  the  two  words 


A  10101010 


and 


Q,  11001100 
we  can  generate  the  word 

M  10001000 


10101010 
11001100 

10001000 


by  performing  the  bit-by-bit  logical  product  of  the  two  words  A 
and  Q,,  that  is  to  say,  we  can  form  A  O  Q  =  M. 
If  we  have  the  two  words 


A  111000111 


000111000 


and 


M  101010101  .  .  .  010101010 
we  can  replace  A  by  M  ©  A  giving 

A  010010010  .  .  .  010010010 
and  as  a  final  example,  we  can  replace  A  by  A  ©  M  giving 

A  111010111   .  .  .  010111010 

These  examples  merely  illustrate  the  kinds  of  bit  manipulation 
that  are  possible,  and  no  attempt  has  been  made  to  be  exhaustive. 
As  one  gains  experience  in  using  such  commands  it  is  possible  to 
discover  how  versatile  this  new  instrument  is  as  an  information 
processor.  Research  workers  from  many  diverse  fields  are  con- 
stantly finding  ways  to  apply  such  machines  to  their  problems. 


CHAPTER 
III 

INFORMATION  PROCESSING  THEORY 

Robert  K.  Lindsay,  Ph.D. 


n 


'ESCARTES  is  usually  credited  with  introducing  the  mind- 
body  problem  to  psychology.  What  he  did  was  introduce  the  body 
to  psychology.  In  those  days,  of  course,  there  were  no  card-carrying 
psychologists,  but  there  were  many  people  who  were  interested 
in  the  huinan  thought  processes,  which  were  assumed  to  reside 
in  a  mysterious  nonentity  called  the  mind.  Descartes  wished  to 
show  how  the  mind  influenced  the  motions  of  the  body,  and  in 
so  doing  he  made  some  guesses  as  to  how  the  body  itself  might  have 
something  to  do  with  decision  making.  The  abstracted  description 
of  the  control  mechanism  which  Descartes  provided  sounded  much 
like  a  description  of  the  inechanical  statues  which  were  found  in 
the  gardens  of  his  day.  He  described  nerves  as  hollow  tubes 
through  which  ran  bell  ropes  of  the  sort  used  to  summon  servants. 
These  bell  ropes,  when  stimulated,  manipulated  valves  in  the  head 
which  directed  the  flow  of  animal  spirits  from  the  ventricles  of 
the  brain  to  the  inflatable  muscles.  The  expansion  of  the  muscles 
brought  about  movement.  The  mind  was,  in  this  model,  adjoined 
to  the  body  through  the  pineal  gland,  which  served  as  a  sort  of 
master  control  which  could  override  any  of  the  other  valves,  thus 
maintaining  the  integrity  of  the  free  will. 

Descartes'  system,  though  somewhat  obsolete  today,  was,  in  its 
time,  quite  ingenious.  Even  though  men  are  no  longer  profitably 
viewed  as  garden  decorations,  Descartes  and  his  notions  can  be 
credited  with  having  thrown  a  great  deal  of  light  on  the  working 
of  the  human  control  system. 

Although  advances  in  physiology  have  shown  the  preceding 
model  to  be  inadequate,  such  knowledge  has  not  eliminated  the 

34 


Injormation  Processing  Theory  35 

approach.  Students  of  behavior  still  exhibit  a  propensity  to  describe 
the  human  system  in  terms  of  the  engineer's  handiwork.  In  the 
first  half  of  this  century,  and  still  today,  psychological  models 
took  a  form  remarkably  similar  to  the  telephone  switchboard,  with 
incoming  signals  being  routed  through  connections,  strengthened, 
by  degrees,  through  use,  to  trigger  a  response,  with  scarcely  a 
"by-your-leave"  to  their  brothers  under  the  skin.  Psychology 
moved  back  the  boundaries  of  the  mind  as  emphasis  withdrew 
from  the  mechanical  procedures  which  performed  the  motions, 
and  moved  toward  the  control  procedures  which  decided  what 
motions  would  be  made.  The  mechanical  monster  seemed  too 
clumsy,  and  an  electrical  monster  was  substituted. 

As  the  preceding  papers  have  indicated,  the  recent  years  have 
seen  some  new  tools,  both  conceptual  and  actual,  added  to  the 
engineer's  gadget  bag.  These  years  have  also  seen  some  further 
friendly  borrowing  of  these  tools  by  students  of  biology  and 
behavior.  The  new  tool  with  which  I  am  most  impressed  and 
with  which  I  hope  to  impress  you  is  the  digital  computer.  A  great 
deal  of  work  has  been  directed  in  the  last  decade  toward  the 
understanding  of  the  neural  bases  of  the  control  processes  which 
interest  the  psychologist,  and  a  fair  number  of  psychologists  have 
decided  that  switchboards  are  perhaps  not  the  best  model  for 
neurological  processes.  So  now  we  hear  that  people  are  really 
like  electronic  monsters. 

We  are  not  quite  as  physiologically  naive  as  was  Descartes. 
We  know  that  humans  are  not  really  made  up  of  transistors, 
resistors,  or  even  electric  wire.  What,  then,  do  we  mean  when  we 
say  that  humans  are  like  computers? 

Humans  and  other  animals  make  decisions,  behave,  solve  prob- 
lems, and  learn.  Machines — digital  computers — also  do  these 
things.  Superficially,  at  least,  humans  are  like  machines.  C'an  we 
be  more  specific?  If  a  machine  does  the  same  things  that  a  human 
does  and  fails  in  the  same  things  in  which  humans  fail,  then 
machine  and  man  are  alike  at  a  more  basic  level  of  description. 
The  more  details  which  can  be  replicated  by  the  machine,  the 
closer  is  the  comparison. 

Once  the  basic  features  of  a  computer  are  pointed  out  to  some- 
one who  has  seriously  attempted  to  analyze  the  human  system, 


36  Information  Storage  and  Neural  Control 

some  similarities  are  obvious,  as  are  some  points  of  dissimilarity. 
A  legitimate  question  yet  remains:  How  does  the  existence  of  a 
potentially  remarkable  device  of  this  sort  aid  us  in  our  present 
work?  There  are  at  least  two  answers  to  this  question. 

The  first  answer  is  exemplified  by  the  work  of  those  who  have 
studied  the  brain  as  a  computing  machine.  Turing  (1)  proved 
that  a  very  simple  device  is  capable  of  computing  any  number 
which  a  reasonable  man  might  wish  to  call  computable.  In  a 
classic  paper,  McCulloch  and  Pitts  (2)  argued  that,  since  mathe- 
matical logic  has  been  stated  in  a  form  where  deductions  become 
a  form  of  computation,  a  device  of  no  greater  complexity  than  a 
Turing  machine  should  be  capable  of  performing  any  logical 
computation,  no  matter  how  complex.  They,  in  fact,  proved  that 
elements  no  more  complex  than  neurons  were  sufficient  for  this 
purpose.  That  is,  they  demonstrated  that  to  every  logical  proposi- 
tion there  corresponds  a  nerve  net  which  can  be  constructed  from 
idealized  neurons,  and  that  the  converse  is  also  true.  The  brain, 
thus,  is  not  just  in  some  vague  sense  like  a  computing  machine; 
the  brain  is  a  computing  machine.  The  important  activity  of  the 
brain  is  its  inputting,  processing,  and  outputting  of  information. 
Although  we  have  had  these  computing  machines — brains — around 
for  a  long  time,  only  recently  have  we  had  any  others  of  comparable 
complexity.  To  biology,  the  presence  of  the  digital  computer  has 
provided,  in  addition  to  a  new  source  of  interested  human  talent, 
a  manipulatable  device  which  can  be  studied  in  vivo  and  whose 
descriptors,  as  they  are  discovered,  might  profitably  be  applied 
to  the  human  machine.  Studies  of  electronic  systems,  and  of  sys- 
tems in  general,  have  provided  insights  into  some  important 
biological  questions.  To  mention  just  one  such  question  which 
has  received  a  lot  of  attention:  How  is  it  possible  to  construct  a 
reliable  system  out  of  billions  of  variable,  unreliable  parts?  This 
question  has  been  attacked  profitably  by  McCulloch  (3,  4)  and 
von  Neumann   (5),   among  others. 

The  second  answer  is  the  one  on  which  I  wish  to  dwell  more 
extensively.  It  is  frequently  the  case  that,  although  we  know  the 
properties  of  all  components  of  a  system,  we  are  unable  to  predict 
the  behavior  of  the  system  if  it  is  composed  of  many  components. 
It  is  true  that  not  all  of  the  relevant  properties  of  neurons  are 


Information  Processing  Theory  37 

known  in  sufficient  detail.  However,  some  of  their  basic  features — 
features  which  undoubtedly  are  critical  in  brain  function —  are 
well  established  and  can  be  described  accurately.  A  lot  of  talent 
has  gone  into  speculating  on  the  manner  in  which  neurons  inter- 
act to  perform  the  higher  functions.  One  such  theory  is  that  of 
the  psychologist  Hebb  (6),  who  attempted  to  explain  the  phe- 
nomenon of  memory  in  terms  which  were  physiologically  sound 
and  yet  psychologically  relevant.  Hebb  proposed  three  phases  in 
the  formation  of  memory  traces.  The  first  is  reverberation,  the 
persistence  of  nervous  activity  after  the  termination  of  the  initiating 
stimulus.  The  second  mechanism,  the  cell-assembly,  consists  of  a 
characteristic  pattern  of  firing  associated  with  a  particular  stim- 
ulus configuration  and  comes  into  being  upon  adequate  repetition 
of  the  stimulus.  To  account  for  this,  Hebb  postulated  that  if  one 
neuron  succeeded  in  firing  a  second,  the  synapse,  by  some  un- 
specified processes,  should  change  so  as  to  make  this  triggering 
more  probable  in  the  future.  The  third  mechanism,  evolving  from 
the  second,  amounts  to  the  passing  of  activity  from  one  cell-assem- 
bly to  another  as  a  result  of  the  repeated  temporal  sequencing  of 
the  corresponding  stimuli.  This  mechanism,  the  phase-sequence, 
is  the  primitive  basis  of  expectancy,  an  important  psychological 
concept. 

Although  the  neuronal  properties  which  Hebb  assumed  are 
well  established,  and  the  "growth"  hypothesis  is  almost  certainly 
correct,  it  is  not  an  easy  task  to  show  that  these  assumptions  are 
sufficient  to  cause  the  reverberation,  cell-assembly,  phase-sequence 
organization  postulated.  Rochester  et  al.  (7),  attempted  to  demon- 
strate the  sufficiency  of  Hebb's  assumptions  in  a  novel  way.  They 
instructed  a  digital  computer  to  behave  according  to  the  assump- 
tions, and  then  simply  observed  its  behavior.  It  is  interesting  to 
note  that  they  were  forced  to  make  some  additional  minor  assump- 
tions before  the  theory  was  specified  in  sufficient  detail  to  be 
realized.  But  more  important,  they  found  that,  although  rever- 
beration was  easily  achieved  and  cell-assemblies  formed  spon- 
taneously after  some  suitable  modification  of  the  theory,  phase- 
sequences  were  not  achieved.  This  work  has  given  some  important 
clues  as  to  what  is  lacking  in  the  theory,  and  some  specific  altera- 
tions have  been  proposed. 


38  Information  Storage  and  Neural  Control 

We  see  in  the  works  of  McCulloch  and  Rochester  a  feature 
which  distinguishes  them  from  many  other  efforts  at  describing 
brain  activity  and  behavior.  This  new  approach  may  be  con- 
trasted with  many  behavior  theories  which  describe  the  product 
or  output  of  the  system  rather  than  the  process  by  which  the 
output  is  obtained.  Akhough  it  is  perfectly  reasonable  to  develop 
a  science  of  psychology  from  product  models,  psychological 
theories  would  be  more  directly  useful  to  neurophysiology  if  they 
could  be  stated  as  process  models.  The  absence  of  analytic  tech- 
niques and  languages  for  describing  processes  has  until  recently 
blocked  any  rigorous  development  of  psychological  process  models. 
The  development  of  computer  sciences  offers  hope  of  removing 
these  blocks. 

In  order  to  provide  a  clearer  picture  of  what  is  meant  by  the 
infoimation  processing  theory  approach,  I  wish  to  contrast  a 
process  model  with  three  other  types  of  theoretical  descriptions. 
All  four  of  the  models  to  be  discussed  deal  with  human  language 
production.  The  three  non-process  theories,  in  fact,  purport  to 
explain  exactly  the  same  phenomenon;  unfortunately,  the  infor- 
mation processing  model  does  not.  However,  I  think  the  exposition 
will  not  suffer  excessively  from  this  lack  of  aesthetics. 

The  phenomenon  described  by  the  three  non-process  models 
is  the  strikingly  regular  statistical  distribution  of  words  produced 
in  speech  and  writing.  The  data  are  most  often  displayed  in  what 
is  called  the  standard  curve,  which  is  obtained  as  follows.  A 
passage  of  text  is  examined  to  determine  which  word  occurs  with 
greatest  frequency,  which  with  second  greatest  frequency,  and  so 
on.  A  graph  is  then  made,  with  this  rank  plotted  on  the  abscissa 
and  frequency  of  occurrence  on  the  ordinate.  Thus,  if  "the"  is 
the  most  frequent  word,  and  if  it  occurs  one  thousand  times,  then 
the  point  so  determined  is  (1,  1000).  Such  graphs,  made  from  a 
wide  variety  of  sources,  are  well-approximated  by   the   equation 

Jr  =C, 
where 

/  =  frequency, 
r  =  rank,  and 
C  =  Constant. 


Information  Processing  Theory  39 

If  the  curve  is  plotted  on  log-log  coordinates,  this  rectangular 
hyperbola  becomes  a  straight  line  with  slope  of  minus  one.  The 
above  equation  is  equivalent  to 

nf  =  K, 
where 

/  =  frequency  of  occurrence, 
n  =  number  of  words  of  that  frequency,  and 
K  =  Constant. 

This  form  is  the  more  usual  representation  of  a  frequency  dis- 
tribution. 

The  first  explanation  of  this  regularity  which  I  wish  to  discuss 
is  due  to  Zipf  (8),  and  exemplifies  what  I  will  call  a  mentalistic 
theory,  although  I  shall  try  to  avoid  defense  of  that  term.  The 
basis  of  Zipf's  explanation  is  the  Principle  of  Least  Effort,  which 
itself  requires  explanation.  People,  says  Zipf,  behave  so  as  to 
minimize  effort,  and  this  strategy  underlies  behavior  of  all  forms. 
He  is  at  pains  to  emphasize  that  effort  includes  not  only  actual 
work  but  mental  effort  as  well,  including  the  mental  effort  to 
decide  which  path  involves  the  least  effort.  And  here  is  where  the 
trouble  begins.  Since  a  person  is  unable  to  predict  the  future 
exactly,  he  must  make  guesses.  The  Principle  then  becomes  the 
statement  that  a  human  will  behave  so  as  to  "minimize  the  average 
rate  of  probable  work."  At  this  point  it  is  clear  that  no  problems 
are  solved  by  the  Principle  because,  in  order  to  make  a  prediction, 
we  must  determine  the  subject's  view  of  the  world  and  understand 
his  decision  process,  which  of  course  was  the  problem  with  which 
we  began. 

Since  this  is  an  important  point,  let  me  state  it  somewhat 
differently.  Although  Zipf  provides  elaborate  discussion  of  what 
he  means  by  effort,  he  never  gets  around  to  telling  us  how  it  is 
to  be  measured,  nor  does  he  ever  rigorously  state  what  is  meant 
by  this  Principle  of  Least  Effort.  Since  the  length  of  time  over  which 
this  undefined  effort  is  to  be  averaged  is  also  unspecified,  it  is 
clear  that  we  can  adjust  the  foresightedness  of  our  subjects  to 
obtain  whatever  results  we  desire.  Since  the  word  "probable"  is 
thrown  in,  our  only  recourse  is  to  the  subjective  probabilities 
of  the  subjects  in  order  to  apply  the   principle;   and  subjective 


40  Information  Storage  and  Neural  Control 

experience  is  by  definition  private.  The  Principle  is  seen  to  be  an 
elastic  phrase  which  can  be  distorted  to  fit  whatever  data  are 
presented.  Zipf  has  committed  an  error  whicli  psychologists  have 
been  attempting  to  eliminate  for  fifty  years.  This  is  why  I  have 
used  the  term  "mentalistic"  to  describe  Zipfs  theory. 

But  let  us  take  one  quick  look  at  how  the  principle  is  applied 
to  the  word-frequency  results.  Zipf  argues  on  teleological  grounds. 
From  the  viewpoint  of  the  speaker's  purpose,  communication 
would  require  least  effort  if  one  word  could  be  used  to  convey 
every  meaning,  for  then  no  decision  would  be  necessary.  From 
the  auditor's  viewpoint,  every  meaning  should  have  a  separate 
code,  for  then  his  effort  would  be  least.  Thus,  we  have  the  Foice  of 
Unification  in  opposition  to  the  Force  of  Diversification,  tending  to 
create  a  vocabulary  balance.  I  wish  to  quote  the  following  passage 
from  Zipfs  book  as  an  example  of  the  application  of  the  Principle: 

"We  obviously  do  not  yet  know  that  there  is  in  fact  such  a 
thing  as  vocabulary  balance  between  our  hypothetical  Forces  of 
Unification  and  Diversification,  since  we  do  not  yet  know  that 
man  invariably  economizes  with  the  expenditure  of  his  effort;  for 
that,  after  all,  is  what  we  are  trying  to  prove.  Nevertheless — and 
we  shall  enumerate  for  the  sake  of  clarity — if  1)  we  assume  ex- 
plicitly that  man  does  invariably  economize  with  his  effort,  and 
if  2)  the  logic  of  our  preceding  analysis  of  a  vocabulary  balance 
between  the  two  Forces  is  sound,  then  3)  we  can  test  the  validity 
of  our  explicit  assumption  of  an  economy  of  effort  by  appealing 
directly  to  the  objective  facts  of  samples  of  actual  speech  that 
have  served  satisfactorily  in  communication.  Insofar  as  4)  we  may 
find  therein  evidence  of  a  vocabulary  balance  of  some  sort  in 
respect  of  our  two  Forces,  then  5)  we  shall  find  ipso  facto  a  con- 
firmation of  our  assumption  of  1)  an  economy  of  effort."  (9). 

The  argument  which  has  been  presented  is  scientifically  sound: 
deduce  something  from  theoretical  assumptions;  if  the  deduction 
is  empirically  verified,  the  theory  has  found  support.  Our  only 
quarrel  is  with  the  weakness  of  the  prediction.  All  that  has  been 
predicted  is  that  passages  of  actual  speech  or  writing  will  not  be 
repetitions  of  the  same  single  word,  nor  will  all  words  be  different. 
The  absence  of  detailed  specification  of  the  constructs  of  the  theory 
leads  only  to  predictions  which  are  trivial,  and  yet  the  superficial 


Information  Processing  Theory  41 

rigor  of  the  statement  of  the  Principle  gives  the  impression  that 
something  has  been  said — something  which  sounds  very  reason- 
able and  powerful. 

The  weakness  of  such  explanations  would  seem  not  to  deserve 
such  extensive  comment,  and  yet  the  problem  has  come  up  so 
frequently,  particularly  in  psychology,  that  apparently  it  does  not 
hurt  to  point  out  such  fallacies.  Not  all  cases  are  so  obvious, 
unfortunately,  and  many  theories  which  appear  rigorous  to  the 
most  competent  of  scientists  are  found  at  times  to  fail  on  this 
same  count.  I  shall  return  to  this  point  later. 

The  amazing  regularity  found  in  the  word-frequency  data, 
regularity  which  seems  to  be  so  hard  to  come  by  in  the  field  of 
human  behavior,  deserves  more  serious  attempts  at  explanation. 
Fortunately,  other  workers  have  attacked  the  same  problem;  and, 
fortunately,  for  our  purposes  of  today,  one  such  approach  illus- 
trates a  stochastic  theory  and  another  illustrates  an  application 
of  information  theoretic  concepts. 

The  stochastic  model  is  due  to  Simon  (10).  The  approach  is 
to  postulate  probabilistic  decision  rules  and  from  these  to  derive 
the  statistical  properties  of  a  device  which  follows  the  rules.  The 
challenge  is  to  postulate  rules  which  will  yield  the  statistical 
properties  of  the  observations,  in  this  instance  the  frequency 
distribution  of  words.  It  should  be  pointed  out  that  the  weaker, 
i.e.,  the  more  general,  are  the  underlying  postulates,  the  better 
is  the  theory.  Thus,  as  in  any  theory,  we  wish  to  account  for  as 
much  as  possible  with  as  little  as  necessary. 

Simon's  basic  model  rests  on  only  two  assumptions.  From  these 
he  is  able  to  derive  a  frequency  distribution  known  as  the  Yule 
distribution,  which  has  all  of  the  properties  required  for  fitting 
the  word-frec|uency  data.  As  a  matter  of  fact,  slight  variations 
on  the  assumptions  yield  slight  differences  in  the  resulting  dis- 
tribution. These  various  forms  of  the  theory  can  be  plausibly 
associated  with  various  real  world  situations,  and  the  theory  thus 
accounts  for  several  phenomena,  such  as  the  distribution  of  authors 
by  number  of  professional  papers  published,  the  distribution  of 
incomes,  and  the  distribution  of  biological  species  by  genera. 
Furthermore,  the  steady  state  statistical  properties  are  fairly  insen- 
sitive to  minor  changes  in  the  assumptions. 


42  Information  Storage  and  Neural  Control 

With  regard  to  word  production,  Simon  argues  that  an  author 
selects  words  not  only  by  association  with  other  words  he  has 
already  put  down  in  this  passage  but  by  imitation  of  the  language 
as  well.  In  other  words,  the  next  word  which  an  author  will  choose 
is  determined  by  the  frequency  distribution  of  his  present  effort, 
i.e.,  the  subject  under  discussion,  and  by  the  statistical  properties 
of  the  language  he  is  using.  We  thus  need  to  postulate  two  "birth 
processes."  Further,  we  will  consider  a  passage  of  fixed  length, 
as  is  reasonable  if  we  consider  that  our  sample  under  analysis  is 
but  a  segment  selected  from  the  author's  total  word  production. 
Thus,  we  must  postulate  a  "death  process"  which  specifies  which 
words  are  dropped  from  the  sample  at  one  end  of  the  passage  as 
we  add  words  at  the  other  end. 

Let  (3  be  the  proportion  of  words  added  by  imitation,  and  let 
f*{i)  be  the  relative  frequency  of  words  which  have  occurred 
exactly  /  times  each.  The  following  assumptions  shall  prove  to 
be  sufficient. 

Birth  Processes 

i.  The  probability  of  adding  a  word,  already  having  occurred 

i  times,  by  association  is  {\-^)i  f*{i)- 
ii.  The  probability  of  adding  a  word,  already  having  occurred 
i  times,  by  imitation  is  ^{i-c)  /*(i). 

Death  Process 

iii.  If  a  word  of  frequency  i  is  dropped,  all  instances  of  that 
particular  word  are  dropped;  this  occurs  with  proba- 
bility /*(z). 

Note  that  in  both  birth  processes,  the  probability  of  adding  a 
word  is  proportional  to  the  total  number  of  occurrences  of  that 
word  and  all  other  words  used  equally  often.  The  assumption 
that  a  word  will  be  chosen  with  probability  proportional  to  its 
own  frequency  is  a  special  case  of  our  assumptions  i.  and  ii.; 
hence,  the  assumptions  used  are  more  general.  The  factors  in- 
volving jS  are  self  explanatory.  The  constant,  c,  appearing  in 
assumption  ii.  may  be  made  plausible  by  the  following  considera- 
tions. In  any  given  passage,  not  all  of  the  words  in  the  language 
will  have  been  used  yet.   We  wish  to  allow  the  possibility  that 


Information  Processing  Theory  43 

a  word  never  before  used  in  this  sample  will  be  produced.  Thus, 
we  wish  to  attenuate  slightly  the  probabilities  attached  to  the 
association  process. 

The  death  process,  though  intuitively  less  satisfactory,  would 
be  true  if  all  occurrences  of  a  given  word  were  closely  grouped  as 
if  associated  only  with  the  topic  of  the  moment.  Thus,  if  a  page  is 
removed  from  the  sample,  it  is  likely  that  all  occurrences  of  every 
word  on  it  will  be  removed. 

Since  we  are  discussing  a  sample  of  constant  size,  we  may 
write  an  equation  to  indicate  that  total  births  are  balanced  by 
total  deaths.  But,  furthermore,  we  are  concerned  with  the  so-called 
steady  state  of  this  stochastic  mechanism.  This  is  the  situation 
which  persists  when  the  sample  size  is  large  enough  that  the 
statistical  distribution  remains  invariant.  Thus,  the  number  of 
words  dropped  from  the  category  witli  relative  frequency  f*{i) 
must  be  just  balanced  by  the  number  of  words  entering  that 
category,  which  is  the  number  of  births  in  the  category  with 
relative  frequency  /*(/-l).  These  requirements  enable  us  to  write 
the  following  equation 

births  in  (z  — 1)  minus  births  in  (z)  minus  deaths  in  (0  =  0 

(I-/3C-1)  j*{i-\)  -  (i-^c)  f*{i)  -  f*(i)  =  0 

which  may  be  rewritten  as 

/•«=(:5^)/-o-i) 

thereby  recursively  defining  the  desired  quantity.  This  function 
has  the  required  properties. 

In  the  example  of  the  stochastic  theory,  then,  the  assumptions 
are  probabilistic  decision  rules  and  the  deductions  are  made 
analytically. 

The  third  description  of  language  production  is  due  to  Mandel- 
brot (11)  and  exemplifies  the  application  of  information  theoretic 
concepts.  Superficially,  this  approach  is  similar  to  that  of  Zipf, 
for  Mandelbrot  derives  the  equation  for  the  standard  curve  by 
minimizing  the  cost  of  coding  the  speaker's  ideas  into  words, 
subject  to  the  constraint  of  a  fixed  amount  of  infoimation  trans- 
mitted per  word-  However,  Mandelbrot  is  quite  specific  as  to 
what  he  means  by  both  information  and  cost. 


44  Information  Storage  and  Neural  Control 

Zipf  argued  that,  while  the  speaker's  effort  (cost)  would  be 
least  if  only  one  word  were  used,  this  situation  does  not  persist 
because  the  listener's  decoding"  efforts  would  be  too  great.  Infor- 
mation theory  allows  us  to  put  this  notion  on  a  sounder  base. 
You  have  seen  that  a  message  which  is  always  sent  can  convey 
no  information,  and  that  the  larger  the  vocabulary,  or  set  of 
alternatives,  from  which  messages  are  selected,  the  greater  is  the 
information  which  they  convey.  On  the  other  hand,  the  process 
of  deciding  which  message  is  next  to  be  sent  is  also  more  difficult 
when  the  set  of  messages  is  larger.  Mandelbrot  has  proposed  that 
the  balance  of  these  two  factors  may  be  conceived  as  the  basis  for 
word  statistics,  and  in  this  we  see  the  similarities  with  Zipf.  How- 
ever, Mandelbrot  has  employed  a  specific  definition  of  infoimation, 
and  has  rigorously  defined  the  probleni.  Let  us  examine  the  main 
features  of  his  derivation  for  a  problem  which  is  formally  identical 
with  the  one  stated  above:  Given  a  fixed  average  cost  per  word, 
what  will  be  the  frequency  distribution  of  the  words  to  give  inaxi- 
mum  information  per  word? 

Let  Cr  be  the  cost  of  the  r-th  most  frequent  word,  which  occurs 
with  probability  p^.  Average  cost  per  word,  C,  is  then 

C   =  X  VrCr. 
r 

Also, 

r 

must  hold.  The  problem  is  then  to  maximize 

H    =     -IZ  Vr  log  Pr 

T 

subject  to  the  above  conditions. 
When  this  is  done,  it  is  found  that 

Vr  =  AAI 

where  A,  B,  and  M  are  constants  which  have  interpretations  in 
terms  of  the  coding  process.  One  further  step  is  needed  in  order 
to  complete  the  derivation,  and  this  involves  relating  C,-  and  rank,  r. 
Mandelbrot  has  managed  to  show  that  if  words  are  coded 
"optimally,"  the  resulting  word  statistics  will  be  correct.  Suppose 
that  words  are  coded  from'  some  elementary  units.  These  units 


Information  Processing  Theory  45 

are  undefined,  being  like,  perhaps,  phonemes,  but  not  necessarily 
so  identified.  It  is  only  necessary  to  assume  that  words  are  com- 
posed of  these  units  and  that  the  cost  of  a  word  is  equal  to  the 
sum  of  the  costs  of  the  units.  To  illustrate,  take  the  special  case 
where  each  unit  has  the  same  cost  attached  to  its  use.  The  least 
expensive  words  are  tlien  the  ones  composed  of  single  units.  If 
there  are  M  units,  there  are  M  such  minimum  cost  words,  M~ 
double  unit  words  (second  in  cost),  and  so  on. 

The  rank  of  a  word  will  be  determined  by  the  number  of  words 
which  can  be  coded  with  Cr  or  fewer  symbols.  For  example,  if 
words  are  coded  as  binary  sequences,  M  =  2,  and  there  are 
fourteen  codes  of  three  or  fewer  digits  (0,  1,  00,  01,  10,  11,  000, 
001,  010,  Oil,  100,  101,  110,  ill).  Thus  a  word  of  cost  3  will  have 
rank  Hand  )•{?>)  =  14. 


In  general, 


c  c 

r{Cr)  =i:ii/^=i:ii/^- 1 


Cj. 

+  1 

1  - 

-  M 

1  - 

-  M 

M 

(71/- 

1) 

1  -  71/ 
1  -  M 


so  that 


M  -  1 


71/^'  -  1  =  /•  (71/  -  l)iW-' 


Now, 


Cr   =    Cr  log,/  71/ 

=    log,;  [(71/'''-    1)    +    1] 

=  log,/  [(7I/'''-  -  1)  +  .^7  (71/  -  l)-\/M  {M  -  !)-'] 

,         r  +  7l/(J/-l)-' 
=  log,/ 


71/  (71/  -  ir' 

=  log,/  (/•  +  M  {M  -  I)-')  -  log,/  .1/  +  log,/  (.1/  -  1) 
which  is  of  the  form 

Cr  =  log,/  (/•  +  m)  +  / 
where  m  and  jn  are  factors  independent  of  r.  Mandelbrot  shows 
that  the  general  form  of  the  expression  for  CV  \s>  the  same  no  matter 


46  Information  Storage  and  Neural  Control 

what  coding  rules  are  assumed,  provided  that  the  coding  results 
in  a  ranking  of  words  by  cost. 

Substituting  this  last  equation  in  our  first  result  yields 
Pr  =  P  ijr  -\-  my  , 
which  reduces  to  the  standard  equation  when  w  =  0  and  B  =   —  1 . 
The  additional  parameters  allow  closer  fit;  but  since  each  has  a 
"physical  interpretation,"  we  are  not  really  cheating. 

My  purpose  is  not  to  compare  the  adequacy  of  these  three 
particular  theories — this  has  been  argued  elsewhere:  Simon  (12, 
13,  14),  Mandelbrot  (15,  16),  Rapoport  (17) — but  to  contrast 
the  theoretical  style.  The  procedure  of  Mandelbrot,  then,  is  to 
start  from  ceitain  assumptions  and  to  deduce  the  resulting  prop- 
erties. In  his  case,  the  assumptions  were  stated  in  information 
theoretic  terms  and  the  deductions  were  analytic. 

My  final  example  illustrates  the  information  processing  ap- 
proach. Unfortunately,  as  I  rhentioned  earlier,  it  does  not  deal 
with  the  same  data,  although  it  is  concerned  with  verbal  pro- 
duction. Hence,  we  may  contrast  the  underlying  notions  even  if 
we  cannot  compare  theoretical  validity. 

This  description  is  due  to  Yngve  (18)  who  has  attempted  to 
explain  some  of  the  salient  features  of  English  grammar.  As  a 
starting  point,  Yngve  has  pointed  out  that  English  often  provides 
several  grammatically  correct  and  semantically  equivalent  ways 
of  saying  the  same  thing,  and  that  some  of  these  ways  are  quite 
complicated.  On  the  other  hand,  the  grammars  of  formal  mathe- 
matical notations,  such  as  that  of  algebra,  impose  severe  limits 
on  the  number  of  forms  permitted,  and  yet  these  restrictions  do 
not  hamper  expressive  power  nor  limit  "sentence"  length.  Let 
us  consider  just  two  examples.  In  English  the  standard  form  of 
modification  places  modifiers  before  that  which  is  modified.  Thus, 
we  have  such  phrases  as  "the  big,  happy  man."  But  we  may  also 
reverse  this  order —  which  logically  should  be  completely  ade- 
quate— in  such  phrases  as  "a  man  as  tall  as  a  circus  giant."  Why 
do  we  not  avoid  such  discontinuous  constituents  (some  modifiers 
in  front,  some  behind)  and  use  the  more  consistent  form  "an  as 
tall  as  a  circus  giant  man"  or  "an  as  a  circus  giant  tall  man"? 

Secondly,  note  that  English  provides  both  active  and  passive 
voices:  "Johnny  gave  the  ball  to  Billy"  and  "The  ball  was  given 


Information  Processing  Theory  Al 

to  Billy  by  Johnny."  Surely,  since  both  are  equivalent,  we  are 
just  complicating  things  by  allowing  two  grammatical  forms.  In 
algebra  we  do  not  have  a  symbol  \  as  in  "B\A,"  which  means 
the  same  as  "A/B",  but  we  may  say  in  English  "B  divided  into 
A"  as  well  as  "A  divided  by  B." 

To  explain  these  and  many  other  aspects  of  English  grammar, 
Yngve  postulates  a  mechanism  for  sentence  production.  Assume 
that  the  brain  has  a  large  memory  in  which  are  stored  rules  such 
as  S  =  NP  +  VP;  NP  =  T  +  N;  VP  =  V  +  A;  T  =  the;  T  =  a; 
A  =  away;  V  =  went;  V  =  ran;  N  =  man.  Such  rules  define  a 
grammar  in  that  they  can  generate  sentences  if  used  in  the  fol- 
lowing fashion: 

S  =  NP  +  VP 

-T+N+V+A 
=  the  man  went  away 
or  S  =  a  man  ran  away 

By  selecting  various  rules,  we  may  generate  various  sentences,  all 
grammatical. 

However,  in  order  to  generate  sentences  in  the  prescribed  left- 
to-right  fashion,  it  is  necessary  that  we  complete  the  expansion 
of  the  left-most  phrases  while  "keeping"  our  place,"  i.e.,  remem- 
bering the  higher  order  rules  which  are  guiding"  the  sentence 
production.  If  we  had  a  scratch  pad  on  which  to  keep  our  place, 
its  contents  at  various  staoes  mio'ht  look  like  this: 


Verbalized 

On  scratch  pad 

S 

NP  VP 

TN  VP 

the  N  VP 

the 

N  VP 

the 

man  VP 

the  man 

VP 

the  man 

V  A 

the  man 

went  A 

the  man  went 

A 

the  man  went 

away 

man  went  away 

48  Information  Storage  and  Neural  Control 

CUearly,  the  type  of  grammar  rules  will  determine  both  how 
much  we  have  written  down  at  any  time  and  the  maximum 
capacity  required  of  the  scratch  pad.  Since  rules  may  be  used 
recursively,  i.e.,  we  permit  rules  such  as  S  =  S  +  and  +  S,  we 
might  generate  grammatical  sentences  which  exceed  the  capacity 
of  any  given  scratch  pad.  This  is  not  such  a  danger  in  algebraic 
notation,  which  is  not  generally  used  as  a  spoken  language  except 
for  short  expressions,  but  it  could  be  critical  in  spoken  English. 
Since  human  span  of  attention  is  quite  limited — and  we  have 
some  pretty  consistent  evidence  as  to  what  this  limit  is — English 
has  evolved  rules  of  grammar  which  spare  our  mental  scratch  pads. 
For  example,  we  can  see  that  elaborate  phrases  which  occur  at 
the  beginning  of  a  sentence  must  be  expanded  while  keeping  in 
mind  the  structure  of  that  which  is  to  follow.  Grammarians  ad- 
monish us  not  to  use  such  "top-heavy"  sentences.  It  is  not  sur- 
prising to  find  that  we  have  been  provided  with  alternate  ways  of 
modifying  nouns,  and  that  these  ways  allow  us  to  postpone  some 
of  the  modifiers  until  we  have  gotten  rid  of  the  object  of  modi- 
fication. Discontinuous  constituents  are  such  mechanisms. 

The  same  argument  accounts  for  the  existence  of  the  passive 
voice  when  the  active  is  just  as  accurate.  If  the  subject  of  a  sentence 
is  greatly  elaborated,  we  can  postpone  it  until  later  by  making  it 
the  predicate  of  a  sentence  in  the  passive  voice.  Note  how  the 
following  sentence,  used  as  an  exaniple  by  Yngve  and  taken  from 
a  U.  S.  patent,  organizes  the  information  so  that  one  need  not 
expand  the  middle  while  keeping  in  mind  other  features:  "The 
said  rocker  lever  is  operated  by  means  of  a  pair  of  opposed  fingers 
which  extend  from  a  pitman  that  is  oscillated  by  nieans  of  a  crank 
stud  which  extends  eccentrically  from  a  shaft  that  is  rotatably 
mounted  in  a  bracket  and  has  a  worm  gear  thereon  that  is  driven 
by  a  worm  pinion  which  is  mounted  upon  the  drive  shaft  of  the 
motor."  The  same  sentence  can  be  expressed  in  the  active  voice, 
but  this  requires  a  memory  which  is  beyond  the  capabilities  of 
most  of  us.  The  sentence  is  ungrammatical  for  that  reason,  accord- 
ing to  Yngve's  model:  "A  pair  of  opposed  fingers  (that  extend 
from  a  pitman  (which  a  crank  stud  (that  extends  eccentrically 
from  a  shaft  (which  is  rotatably  mounted  in  a  bracket  and  which 
a  worm  gear  (that  a  worm  pinion  (which  is  mounted  upon  the 


Information  Processing  Theory  49 

drive  shaft  that  the  motor  has)  drives)  is  on))  osciUates))  operate 
the  said  rocker  arm." 

Tiie  preceding  examples,  tiien,  represent  four  approaches  to  the 
same  general  type  of  observation.  I  have  called  them  mentalistic, 
statistical,  information  theoretical,  and  information  processing 
theoretical.  The  latter  consists  of  postulating"  some  sort  of  mechan- 
istic decision  procedure;  the  operation  of  the  mechanism  is  then 
examined  and  compared  with  human  behavior.  Assumptions  are 
stated  as  processes;  the  method  of  deduction  is  not  analytic.  For 
processes  more  complex  than  that  in  the  Yngve  example,  the  de- 
duction often  takes  the  foim  of  specifying  the  processes  for  a  digital 
computer,  the  running  of  which  then  provides  the  predictions. 

There  are  yet  several  points  of  this  discussion  which  deserve 
more  elaboration.  First,  why  go  to  all  the  trouble  and  expense  to 
build  and  instruct  this  device  when  we  might  do  better  to  hire  a 
mathematician,  whose  services  are  certainly  cheaper,  to  solve  the 
problem  analytically?  This  is  certainly  a  good  suggestion,  and 
many  people  who  have  resorted  to  simulation  might  better  have 
resorted  to  mathematics.  But  the  systems  which  are  of  major 
interest  to  the  psychologist  and  biologist  have  the  property  of 
being  complex.  Mathematics,  although  it  has  earned  its  place  of 
respect  in  science,  is  not  a  completely  developed  discipline.  The 
task  of  writing  equations  for  the  human  system  is  far  too  difficult. 
Some  attempts  have  been  made  to  describe  mathematically  cer- 
tain learning  processes,  for  example.  Bush  and  Mostellar  (19), 
Estes  (20);  but  it  has  been  necessary  to  limit  the  complexity  of 
the  equations  in  the  interest  of  getting  them  solved.  Learning 
processes  have  pretty  well  resisted  linear  descriptions.  It  is,  how- 
ever, possible  to  define  in  computer  terms  systems  which  cannot 
be  defined  in  normal  mathematical  notation;  and  if  the  system 
can  be  defined  as  a  computer  program,  a  computer  can  simulate 
the  behavior  of  the  system.  It  is  important  to  realize  that  writing 
a  program  is  analogous  to  writing  an  equation,  and  running  the 
program  is  analogous  to  solving  the  equation.  It  is  then  clear 
what  I  meant  when  I  said  that  the  program  is  a  theory:  it  is  a 
theory  in  the  same  sense  that  a  mathematical  equation  is  a  theory — 
it  makes  some  well-defined  assumptions  and  makes  some  predic- 
tions which  are  rigorously  deduced  from  these  assumptions. 


50  Information  Storage  and  Neural  Control 

With  this  analogy  in  mind,  it  is  easier  to  elaborate  on  the  other 
points.  One  may  argue  that  having  discovered  one  set  of  opera- 
tions which  accounts  for  the  behavior  of  a  computer  system  does 
not  assure  us  that  the  same  set  of  operations  is  involved  in  the 
human  system.  This  is  a  truism  which  also  applies  to  mathe- 
matical theorizing;  that  is  to  say,  more  than  one  equation  can 
fit  the  same  set  of  data.  Ultimately,  we  must  live  with  this  prob- 
lem, for  if  a  theory  accounts  for  all  data  within  its  domain,  then 
it  is  as  good  as  a  theory  can  be  even  though  there  is  no  assurance 
that  its  underlying  assumptions  have  any  basis  in  reality.  Such 
considerations  have  forced  philosophers  of  science  to  conclude  that 
reality  has  no  meaning;  we  can  only  ask  if  the  assumptions  work, 
not  if  they  are  real.  The  job  of  the  scientist  is  that  of  the  inven- 
tor who  creates  descriptions,  not  of  the  explorer  who  discovers 
reality. 

Even  leaving  this  ultimate  state  aside,  it  is  important  to  con- 
tinue on  this  same  point,  but  at  a  more  practical  level.  If  we  have 
a  program  which  accounts  for  a  small  segment  of  human  behavior, 
how  have  we  progressed?  Seldom  are  we  satisfied  with  a  theory 
of  small  segments  of  behavior.  Let  us  expand  our  program  until 
it  is  more  encompassing.  If  this  can  be  done  by  making  use  of 
some  of  the  same  postulated  operations,  we  achieve  the  parsimony 
which  we  seek.  Let  us  look  at  programs  written  by  other  people 
to  describe  other  things.  If  they  consist  of  markedly  similar  por- 
tions, then  we  again  have  made  progress.  Eventually,  when  a 
certain  process  or  feature  has  turned  up  frequently  enough  as  an 
asset,  we  may  forget  our  philosophy  and  begin  to  look  within  the 
human  system  to  see  if  we  cannot  find  independent  evidence  for 
the  existence  of  some  such  process.  We  have  thus  generated  two 
types  of  hypotheses:  those  which  make  predictions  about  similar 
types  of  behavior,  and  those  which  give  us  clues  about  the  com- 
position of  the  organism.  I  shall  return  to  some  examples  of  the 
latter  at  the  conclusion  of  this  paper. 

The  third  point  on  which  I  wish  to  elaborate  is  a  matter  of 
practical  research  strategy.  The  process  of  simulation  provides  an 
important  fringe  benefit  which  becomes  apparent  only  after  trial. 

It  has  long  been  a  feature  of  psychological  theorizing  that 
would-be   theories  suffer  from  chronic   vagueness.   The   result  is 


Information  Processing  Theory  51 

a  theory  which  can  be  stretched  to  fit  anything.  The  genesis  of 
this  difhcuhy  lies  in  the  fact  that  the  theorist  knows  what  he  is 
saying  and  so  does  his  audience.  Hence,  it  is  often  possible  to  put 
together  assumptions  which,  logically,  will  not  fit,  or  to  make 
deductions  which,  logically,  do  not  follow.  These  unfortunate 
juxtapositionings  may  go  unnoticed  by  an  intelligent  theorist  and 
his  informed  listeners,  who  can  readily  and  unwittingly  supply 
the  missing  pieces,  ignore  the  excesses,  and  beg  the  answer  which 
they  know  is  there  even  if  it  is  not.  The  computer,  though,  is  a 
very  stupid  audience.  From  one  point  of  view,  it  may  prove  more 
valuable  now  while  it  is  stupid  than  later  when  it  is  not;  for  today 
it  will  not  tolerate  vagueness.  When  a  theorist  with  an  idea  sits 
down  to  convey  his  idea  to  a  inachine  he  almost  invariably  finds 
that  he  must  first  sharpen  it  up.  And  when  the  machine  attempts 
to  simulate  the  idea,  the  theorist  almost  invariably  finds  it  will 
not  do  what  it  is  supposed  to  do. 

These  lengthy  elaborations  on  a  fairly  concise  statement  point 
up  the  similarities  between  the  process  of  computer  simulation 
and  the  other  techniques  of  theory  construction.  The  computer  has 
not  answered  the  many  problems  which  were  formulated  by  these 
other  techniques.  The  computer  will  not  make  scientists  out  of 
programmers.  It  is  just  another  way  of  theorizing  which  has 
certain  special  advantages,  certain  special  disadvantages,  and  the 
same  old  problems. 

1  have  attempted  to  show  how  process  models  may  be  stated 
and  why  computer  simulation  is  often  an  appropriate  means  for 
their  analysis.  It  is  quite  legitimate  to  ask  what  such  efforts  to 
date  have  implied  about  information  storage  and  neural  control 
or,  to  be  more  classical,  neurophysiology.  When  computer  sci- 
entists discover  processes  which  appear  to  be  useful  building 
blocks  for  explaining  human  behavior  or  for  constructing  artificial 
intelligences,  it  is  natural  to  ask  if  actual  mechanisms  for  per- 
forming these  processes  can  be  found  within  the  central  nervous 
system.  The  observations  of  the  reflex  led  Sherrington  to  inquire 
as  to  its  basis,  with  a  great  deal  of  benefit  to  science.  Pavlov  ex- 
amined the  conditioned  reflex  and  based  his  psychology  on  it. 
The  discovery  of  more  complex  processes  could  likewise  direct 
efforts  in  neurophysiological  research. 


52  Information  Storage  and  Neural  Control 

Of  course,  such  procedures  are  dangerous,  and  I  hesitate  to 
make  any  very  strong  suggestions.  The  danger  lies  in  the  fact  that 
a  theory  which  embodies  an  hypothesized  mechanism,  Hke  any 
otlier  theory  involving  an  assumption,  can  only  prove  the  suf- 
ficiency of  the  hypothesis,  not  its  necessity.  Anyone  who  accepts 
directions  from  a  psychologist  runs  the  risk  of  getting"  lost.  None- 
theless, I  will  indicate  a  few  possibilities  based  on  mechanisms 
which  have  been  found  useful  in  psychological  and  computer 
theory. 

One  observation  which  has  proved  highly  important  to  psy- 
chological process  models  has  appeared  in  the  preceding  discussion 
of  Yngve's  hypothesis.  I  refer  to  the  concept  of  a  limited  scratch 
pad,  or  immediate  memory,  as  it  is  called.  It  has  often  been  recog- 
nized that  permanent  memory  can  persist  even  after  severe  dis- 
turbance of  the  ongoing  cerebral  activity,  such  as  that  brought 
about  by  freezing  or  electroshock.  Since  any  form  of  persistent 
trace  must  undoubtedly  require  periods  of  time,  at  least  on  the 
order  of  seconds,  for  establishment,  then  some  temporary  form  of 
storage,  basically  different  from  the  permanent  form,  must  be 
utilized  to  maintain  the  information  until  it  can  be  permanently 
stored.  Miller  (21)  has  shown  that  the  capacity  of  this  immediate 
memory,  as  inferred  from  a  variety  of  psychological  studies,  is 
remarkably  constant.  This  capacity  is  not  measured  in  bits  of 
information,  however,  but  in  terms  of  the  number  of  symbols 
which  can  be  temporarily  remembered;  i.e.,  a  subject  may  retain 
about  seven  binary  digits,  about  seven  decimal  digits,  or  about 
seven  monosyllabic  adjectives,  all  of  which  differ  in  amount  of 
information  as  defined  by  Shannon.  Thus,  the  hunian  is  capable 
of  conceptually  complex  activity  largely  because  he  is  capable  of 
dealing  with  informationally  rich  symbols,  and  he  is  provided 
with  a  capacity  which  is  largely  independent  of  the  richness  of 
his  thoughts. 

By  measuring  a  subject's  success  at  discriminating  various 
numbers  of  stimuli  which  differ  along  one  diniension,  one  finds 
that  the  capacity  of  the  human  communication  channel  is  rela- 
tively constant  at  about  seven  discriminations.  If  one  then  gives 
the  subject  the  task  of  discriminating  stimuli  which  vary  on  two 
dimensions,   one  discovers  that  the  subject,   although  unable  to 


Informatioti  Processing  Theory  53 

distinguish  forty-nine  categories,  can  do  better  than  in  the  one- 
dimensional  case.  For  example,  a  subject  who  can  discriminate 
wiiich  of  ten  positions  a  point  occupies  on  a  line  cannot  place  the 
point  in  one  of  one  hundred  cells  of  a  square,  but  can  manage 
only  twenty-five.  This  is  just  what  would  be  predicted  if  ten  cells 
of  immediate  memory  were  divided  into  two  groups  of  five.  In 
other  words,  the  compound  discrimination  reduces  the  accuracy 
of  discrimination  for  each  dimension,  but  still  allows  independent 
examination  of  each. 

The  question  arises  as  to  the  underlying  neurological  structure. 
Is  there  a  single  set  of  pathways  which  performs  this  function  for 
all  inputs  including  internal  inputs?  It  seems  unlikely,  though  not 
impossible,  that  such  a  set  of  pathways  is  localized  in  one  geo- 
graphic position  in  the  brain;  but  even  if  it  is  diffusely  distributed, 
as  are  other  memory  functions,  one  may  still  ask  if  one  set  serves 
in  common.  Little  work  of  the  kind  summarized  by  Miller  has 
been  done  on  cross-modality  studies,  but  one  wonders  if  there  is 
a  "final  common  path"  for  all  sense  modalities. 

Cllosely  related  to  the  notion  of  informationally  ricii  symbols 
is  the  concept  of  a  hierarchically  organized  memory.  It  is  fairly 
clear  from  both  logical  and  psychological  considerations  that 
nriemory  organization  is  such  that  one  trace  can  evoke  a  number 
of  others,  each  of  which  can  in  turn  evoke  a  number  of  others, 
and  so  forth;  i.e.,  one  trace  is  associated  with  several  others,  and 
any  one  of  them  can  be  elicited  without  eliciting  the  otiiers.  Such 
structures  liave  largely  been  ignored  in  classical  stimulus-response 
models,  where  the  theories  have  been  concerned  with  the  forming 
of  a  single  association  between  two  traces.  Neurophysiological 
theories,  perhaps  reflecting  the  concern  of  the  psychologist,  have 
concentrated  on  exploring  the  method  of  single  associations.  Some 
meaningful  questions  might  be  asked  as  to  the  adecjuacy  of  linear 
neurological  models  for  explaining  hierarchical  structures. 

One  such  question  is  related  to  the  concept  of  set,  which  has 
been  found  extremely  useful,  if  not  necessary,  in  psychological 
theories,  and  which  has  turned  up  under  a  variety  of  names  with 
only  minor  variations  in  meaning.  It  is  recognized  that  a  subject 
can  be  "set,"  by  instructions  or  by  other  experimental  manipula- 
tions, so  as  to  give  responses  of  a  certain  class,  to  perform  operations 


54  hiformation  Storage  and  Neural  Control 

more  quickly,  or  to  overlook  completely  otherwise  obvious  solution 
paths  in  a  problem  situation.  If  one  were  to  instruct  a  computer 
so  that  it  had  this  capability,  it  would  be  required  that  the  set 
information,  given  before  the  critical  task,  provide  information 
(or  set  switches)  at  a  number  of  different  places  in  the  piogram. 
This  is  generally  accomplished  by  setting  a  "flag,"  which  is  tested 
by  various  subroutines,  or  by  setting  several  flags,  one  in  each 
subroutine.  The  result  is  a  memory  structure  which  might  be 
called  diffusely  localized.  This  type  of  signal  must  be  extremely 
flexible,  and  must  be  controlled  by  the  executive  program;  i.e., 
it  must  be  at  a  higher  level  in  the  process  structure.  To  my  knowl- 
edge, no  information  exists  concerning  the  cerebral  mechanism 
which  could  explain  such  a  phenomenon,  nor  has  anyone  worried 
much  about  it.  Although  perhaps  other  mechanisms  are  con- 
ceivable, it  seems  necessary  that  communication  channels  of  some 
sort  must  exist  between  the  higher  control  centers  and  several 
lower  centers,  or  that  the  nerve  nets  which  define  processes  must 
be  constructed  so  that  they  can  be  rapidly,  but  temporarily  altered 
by  some  signal  in  a  higher  control  center. 

Finally,  I  wish  to  point  out  a  feature  underlying  all  of  the 
computerized  brain  models  which  deal  with  the  learning  or  growth 
of  connections  between  neurons.  Such  models  have  been  proposed 
as  the  basis  for  such  complex  functions  as  pattern  recognition 
(Rosenblatt,  22);  yet  each  rests  on  fairly  simple  and  standard 
assumptions  of  the  sort  discussed  above  in  connection  with  Hebb's 
growth  hypothesis:  "If  neuron  B  fires  immediately  after  neuron  A, 
the  probability  increases  that  A  will  fire  B."  Although  such  a 
process  is  quite  feasible,  no  direct  physiological  evidence  defines 
its  mechanism,  so  the  assumption  remains  a  psychological  one. 
It  is  almost  certain  to  be  correct,  and  yet  perhaps  we  should  not 
give  up  the  search  for  alternate  mechanisms — if  not  to  replace 
this  notion,  then  to  complement  it.  For  example,  the  firing  of 
neuron  A  followed  by  the  firing  of  neuron  B  might  increase  the 
efficiency  of  all  other  connections  at  the  A-B  synapse  as  well.  Or 
perhaps  the  A-B  "growth"  takes  place  only  if  B  subsequently  fires 
C,  which  bears  some  relation  to  A.  The  neurological  mechanisms 
underlying  these  suggestions  are  not  so  plausible  as  those  of  the 
Hebb  hypothesis,  but  if  they  are  true  they  might  have  a  profound 


Information  Processing  Theory  55 

effect  on  the  behavior  of  a  highly  interconnected  net.  Here  is 
where  computer  simulations  might  be  used  to  explore  new  pos- 
sibilities. By  studying  the  organizing  effects  of  such  additional 
mechanisms,  which  are  just  as  easily  programmed,  we  might 
reinitiate  some  originality  into  essentially  similar  models. 

The  fact  that  psychologists  and  biologists  are  beginning  to 
think  in  terms  of  processes  in  addition  to  stimulus-response  associa- 
tions and  equations  provides  a  more  obvious  link  between  their 
work  and  that  of  the  physiologist.  It  is  the  promise  of  this  new 
link  which  has  revitalized  discussions  of  cross-fertilization  resulting 
in  conferences  with  titles  like  this  one.  The  value  of  these  new 
conceptions  remains  to  be  seen,  but  it  is  probably  safe  to  assume 
that  anything  which  brings  our  disciplines  closer  together  can  do 
no  harm. 

REFERENCES 

1.  Turing,  A.   M.:  On  computable  numbers,  with  an  application  to 

the  Entscheindungs-problem.  Proc.  London  Math.  Sac,  series  2,  42: 
230-265,  1937. 

2.  McCulloch,  W.  S.,  and  Pitts,  W.:  A  logical  calculus  of  the  ideas 

immanent  in  nervous  activity.  Bull.  Math.  Biophysics,  5.- 11 5-1 33, 
1943. 

3.  McCulloch,    W.    S.:    Agathe  Tyche, — of  nervous   nets — the  lucky 

reckoners.  Proc.  Syrnp.  on  Mechanization  of  Thought  Processes,  Ted- 
dington,  England,  f959. 

4.  McCulloch,   W.   S.:    The  reliability  of  biological   systems,  in  Self- 

Organizing  Systems,  Interdisciplinary  Conference  on  Self-Organizing 
Systems,  ed.  by  Yovits,  M.  C.  and  Cameron,  S.,  New  York, 
Pergamon  Press,  1960. 

5.  von  Neumann,  J.:  Probabilistic  logics  and  the  synthesis  of  relial^le 

organisms  from  unreliable  components,  in  Automata  Studies,  Shan- 
non, C.  E.  and  McCarthy,  J.,  Princeton,  Princeton  University 
Press,  1956. 

6.  Hebb,  D.  O.:  The  Organization  of  Behavior;  a  Neuropsychological  Theory. 

New  York,  Wiley  &  Sons,  1949. 

7.  Rochester,  N.,  Holland,  J.  H.,  Haibt,  L.  H.,  Duda,  \V.  L.:  Tests 

of  a  cell  assembly  theory  of  the  action  of  the  brain,  using  a  large 
digital  computer.  IRE  Trans,  on  Information  Theory,  IT-2;80-93, 
Sept.,  1956. 


56  Information  Storage  and  Neural  Control 

8.  Zipf,  G.  K.:  Human  Behavior  and  the  Principle  oj  Least  Effort.  Cambridge, 

Addison-Wesley,  1949. 

9.  Zipf,  G.  K.:  Ibid.,  p.  22. 

10.  Simon,  H.  A.:  On  a  class  of  skew  distribution  functions.  Biometrika, 

42.-425-440,  1955. 

11.  Mandelbrot,  B.:  An  informational  theory  of  the  statistical  structure 

of  language,     in   Information    Theory,    by  Jackson,    W.,    London, 
Butterworths,   1953. 

12.  Simon,  H.  A.:  Some  further  notes  on  a  class  of  skew  distribution 

functions.  Information  and  Control,  J.-80-88,  1960. 

13.  Simon,  H.  A.:  Reply  to  "final  note"  by  Benoit  Mandelbrot.  Infor- 

mation and  Control,  4;21 7-223,   1961. 

14.  Simon,  H.  A.:  Reply  to  Dr.  Mandelbrot's  post  scriptum.  Information 

and  Control,  4.- 305-308,  1961. 

15.  Mandelbrot,  B.:  A  note  on  a  class  of  skew  distribution  functions. 

Analysis  and  critique  of  a  paper  by  H.   Simon.   Information  and 
Control,  2:90-99,  1959. 

16.  Mandelbrot,  B.:  Final  note  on  a  class  of  skew  distribution  functions: 

analysis  and  critique  of  a  model  due  to  H.  A.  Simon.  Information 
and  Control,  4.- 198-21 6,  1961. 

17.  Rapoport,  A.:  Comment:  the  stochastic  and  the  'teleological'  ration- 

ales of  certain  distributions  and  the  so-called  principle  of  least 
effort.  Behav.  Sci.,  2;147-161,  1957. 

18.  Yngve,  V.:  A  model  and  an  hypothesis  for  language  structure.  Proc. 

Am.  Phil.  Soc,  704:444-466,  1960. 

19.  Bush,  R.  R.,  and  Mostellar,  F.:  Stochastic  Models  of  Learning,  New 

York,  Wiley  &  Sons,  1955. 

20.  Estes,    W.    K.:    Toward    a    statistical    theory    of   learning.    Psychol. 

Rev.,  57:94-107,  1950. 

21.  Miller,  G.  A.:  The  magical  number  seven,  plus  or  minus  two:  some 

limits  on  our  capacity  for  processing  information.   Psychol.  Rev., 
<5J:81-97,  1956. 

22.  Rosenblatt,  F.:  The  perceptron:  a  probabilistic  model  for  informa- 

tion storage  and  organization  in  the  brain.  Psychol.  Rev.,  (55:386- 
408,  1958. 


PART  II— INFORMATION  IN  BIOLOGICAL  SYSTEMS 

Moderator:   Heather  D.  Mayor,  Ph.D. 


CHAPTER 
IV 

GENETIC  CONTROL  OF  PROTEIN  SYNTHESIS 

Harrison  Echols,  Ph.D. 

INTRODUCTION 

kjOME  ten  years  ago  the  work  of  Beadle,  Tatum,  and  Horo- 
witz (1)  led  to  the  famous  "one  gene-one  enzyme"  hypothesis, 
which  asserted  that  gene  control  over  cell  metabolism  is  exerted 
through  genetic  determination  of  the  structural  specificity  of 
enzymes.  I  would  like  to  discuss  our  present  knowledge  and 
beliefs  concerning  genetic  control  of  protein  synthesis  by  starting 
with  this  concept  of  the  "structural  gene"  and  inquiring  into  the 
chemical  nature  of  the  gene  and  into  the  process  by  which  the 
gene  controls  protein  specificity.  Finally,  I  shall  briefly  consider 
the  concept  of  "regulatory  genes"  concerned  with  controlling  the 
rate  of  action  of  the  structural  genes. 

CHEMICAL  IDENTIFICATION  OF  GENES 

It  is  now  generally  accepted  that  deoxyribonucleic  acid  (DNA) 
stores  the  genetic  information  of  the  cell.  The  evidence  for  this 
comes  chiefly  from  work  with  bacteria  and  bacterial  viruses,  and 
is  based  primarily  on  three  types  of  genetic  transfer  experiments: 
transformation,  virus  infection,  and  bacterial  conjugation  (2).  In 
transformation  experiments  purified  DNA  extracted  from  one 
bacterial  population  has  been  shown  to  carry  genetic  information 
to  another  bacterial  population.  For  example,  DNA  from  a  strain 
of  Bacillus  sub  til  is  which  possesses  the  ability  to  synthesize  the 
amino  acid  tryptophan  can  confer  this  biosynthetic  ability  on  a 

59 


60  Information  Storage  and  Neural  Control 

strain  of  B.  siibtilis  which  previously  could  not  synthesize  tryp- 
tophan. 

Evidence  that  DNA  is  the  genetic  material  in  a  DNA-protein 
virus  comes  from  studies  of  the  infection  of  Escherichia  coli  with 
bacteriophage  T2.  Virtually  all  of  the  DNA  of  the  virus  enters 
the  infected  bacterium,  and  virtually  none  of  the  associated 
protein  enters.  Finally,  in  bacterial  conjugation,  DNA  is  trans- 
ferred from  a  donor  to  a  recipient  strain  of  E.  coli.  The  amount 
of  DNA  transferred  is  proportional  to  the  number  of  genes  trans- 
ferred, again  suggesting  that  the  DNA  carries  the  genetic  in- 
formation. 

There  is,  then,  excellent  evidence  that  DNA  is  the  genetic 
storage  material  in  bacteria  and  some  viruses  (there  are  ribonucleic 
acid  (RNA)  containing  viruses  in  which  the  RNA  has  been  shown 
to  be  the  genetic  material).  The  generalization  to  higher  organ- 
isms of  this  picture  of  DNA  as  the  storehouse  of  genetic  information 
rests  largely  upon  the  observations  that  the  DNA  content  per  cell 
nucleus  is  proportional  to  chromosome  number;  haploid  sperm 
cells,  for  example,  have  one-half  the  DNA  of  diploid  somatic 
cells  (3).  Further,  the  chromosomal  DNA  is  quite  stable  meta- 
bolically  as  befits  a  genetic  storage  unit.  At  present,  however, 
much  of  our  belief  in  the  idea  that  genes  are  universally  DNA 
comes  from  a  feeling  that  nature  ought  to  be  universal  about 
such  things  as  the  storage  and  transfer  of  genetic  information, 
so  that  what  holds  true  for  bacteria  should  hold  true  for  man. 

If  we  accept  DNA  as  the  genetic  material,  we  can  then  ask  how 
such  a  molecule  stores  genetic  information.  The  simplest  hypothesis 
concerning  this  point  follows  from  a  consideration  of  the  chemical 
structure  of  DNA.  DNA  is  a  polymer  of  deoxyribonucleotides 
linked  together  by  phosphate  bridges  between  deoxysugars  to 
give  a  sugar-phosphate  "backbone"  with  purine  and  pyrimidine 
side  groups  (Fig.  la).  The  only  topographic  feature  of  this  covalent 
"primary"  structure  which  forms  a  likely  candidate  for  informa- 
tion storage  is  the  base  sequence  of  the  purines  and  pyrimidines. 

A  consideration  of  the  probable  three-dimensional  structure  of 
DNA  tends  to  reinforce  this  view.  The  Watson-Crick  model  (4) 
for  DNA  structure  proposes  that  the  molecule  consists  of  two 
chains  forming  a  double  helix  with  hydrogen  bond  pairing  between 


Genetic  Control  of  Protein  Synthesis 


61 


NH..f     l,Ot°' 


A 

..  T 
..  T 

A 

XL... 

..  C 
A 

T  .. 
A 

..  T 
A 

r  ._ 

-.JL 

G- 

C  .. 

[a] 


[b] 


Fig.  1.  The  Structure  of  DNA.  (a)  Part  of  a  polynucleotide  chain  showing  the 
sugar-phosphate  backbone  with  purine  (adenine)  and  pyrimidine  (thymine) 
side  groups,  (b)  Schematic  representation  of  base  pairing  between  the  two  chains. 
The  sugar-phosphate  chains  are  represented  by  the  parallel  vertical  lines  and 
the  bases  by  horizontal  lines,  (c)  The  double-helix.  Base  pairs  are  represented 

by  horizontal  lines. 

the  bases  adenine  (A)  and  thymine  (T)  and  between  guanine  (G) 
and  cytosine  (C)  (Figs,  lb  and  Ic).  A  and  T  are  called  comple- 
mentary bases  because  of  this  pairing  phenomenon,  and,  similarly, 
G  and  C  are  complementary.  This  model  is  now  supported  by 
evidence  from  a  variety  of  chemical  and  physical  experiments. 
Since  the  double  helix  model  reveals  no  new  irregularities  in 
topography,  one  feels  reasonably  confident  that  the  mode  of 
storage  of  genetic  information  in  DNA  is  in  the  linear  sequence 
of  the  four  bases  A,  G,  T,  C  along  the  DNA  chain.  The  linear 
aspect  of  the  information  storage  mechanism  is  supported  by 
genetic  studies  which  indicate  linearity  of  the  fine-structure  genetic 
map  (5),  the  order  of  mutations  within  a  genetic  region  controlling 
a  sinsie  metabolic  function. 


GENES  AS  DETERMINANTS  OF  PROTEIN  STRUCTURE 
(The  Coding  Problem) 

We  have  sketched  briefly  the  evidence  that  genes  are  DNA  and 
that  the  "genetic  code"'  consists  chemically  of  the  base  sequence 


62 


Information  Storage  and  Neural  Control 


of  the  DNA.  Let  us  now  discuss  how  a  gene  imparts  catalytic 
specificity  to  an  enzyme.  Enzymes  consist  of  a  linear  chain  of  amino 
acids  (the  primary  structure),  coiled  in  part  into  an  a-helix  (the 
secondary  structure),  and  folded  into  a  compact  and  specific 
three-dimensional  structure   (the  tertiary  structure)    (Fig.   2). 


I 
N 

0=C 
H-C- 

H-N 
I 

c= 

CH3-C- 

H 
I 


CH,-Q 

=0 
H 


[c3 


''  lb] 


ft] 


Fig.  2.  The  Structure  of  Protein,  (a)  Part  of  a  polypeptide  chain  showing  the 
peptide  bonded  backbone  with  side  gi'oups  characteristic  of  individual  amino 
acids  (here  alanine  and  phenylalanine),  (b)  Schematic  representation  of  the 
a-helix  showing  the  hydrogen  bonds  required  to  maintain  it.  (c)  The  folded 
polypeptide  chain  in  myoglobin  providing  the  specific  three  dimensional  struc- 
ture of  tlie  protein  (as  determined  by  the  x-ray  crystallographic  work  of  Kendrew 

and  collaborators)  (19). 


The  working  hypothesis  for  the  past  few  years  concerning  gene 
control  over  protein  specificity,  usually  called  the  sequence  hy- 
pothesis (6),  states  that  the  base  sequence  of  the  DNA  specifies 
the  primary  structure  of  the  protein — the  sequence  of  amino  acids. 
The  original  argument  was  based  primarily  on  two  points:  first, 
the  base  sequence  of  DNA  is  linear,  and  the  only  corresponding" 
linear  object  in  the  protein  is  the  amino  acid  sequence;  second, 
since  proteins  diff'er  widely  in  amino  acid  composition,  it  was 
difficult  to  see  how  such  differences  could  arise  other  than  by 
genetic  specificity.  The  argument  is  now  much  stronger.  A  number 


Genetic  Co?itrol  of  Protein  Synthesis  63 

of  substitutions  of  one  amino  acid  for  another  have  been  found  in 
the  abnormal  hemoglobins  (7)  which  are  presumed  to  be  products 
of  a  mutationally  altered  globin  gene.  In  addition,  mutant  bac- 
terial strains  producing  an  altered  alkaline  phosphatase  (8)  and 
an  altered  tryptophan  synthetase  (9)  have  been  shown  to  have 
substituted  one  amino  acid  for  another.  Similarly,  a  number  of 
substitutions  have  been  described  in  tlie  tobacco  mosaic  virus  "coat 
protein"   (10). 

From  the  sequence  hypothesis,  it  is  a  short  step  to  the  usual 
statement  of  the  "genetic  coding  problem":  how  the  sequence  of 
four  bases  in  DNA  specifies  the  twenty  amino  acids  commonly 
occurring  in  protein.  The  first  step  toward  "solving"  the  coding 
problem  is  really  to  show  that  the  problem  as  stated  exists— to 
demonstrate  that  the  base  sequence  of  DNA  does  specify  the  amino 
acid  sequence  of  the  protein.  On  the  protein  side,  evidence  that 
mutations  can  cause  amino  acid  substitutions  has  been  men- 
tioned. On  the  DNA  side,  the  determination  of  nucleotide  sequence 
is  not  possible  at  present,  but  a  prediction  (or  corollary)  to  the 
sequence  hypothesis  has  been  used  to  arrive  at  an  experimentally 
feasible  system.  This  prediction  states  that  the  order  and  relative 
position  of  point  mutations  within  the  structural  gene  for  a  par- 
ticular protein,  presumably  reflecting  base  alterations,  should 
correspond  to  the  order  and  relative  position  of  amino  acid  sub- 
stitutions in  proteins  produced  by  these  mutated  genes. 

Work  on  the  bacterial  enzymes  alkaline  phosphatase  (8)  and 
tryptophan  synthetase  (9)  has  shown  that  two  mutations  linked 
genetically  affect  amino  acids  in  the  same  region  of  the  respective 
proteins,  so  that  we  can  feel  some  confidence  that  the  sequence 
hypothesis  is  correct.  Can  one  determine  which  bases  code  which 
amino  acids  by  this  combined  genetic  and  protein  chemical 
approach?  The  answer  is  probably  yes,  provided  that  mutagens 
specific  for  a  single  base  can  be  developed  and  used;  but  the 
number  of  amino  acid  substitutions  which  must  be  accumulated 
is  almost  prohibitively  large.  Recently,  a  much  more  direct 
approach  to  working  out  the  nature  of  the  genetic  code  and 
probably  its  explicit  solution  has  appeared  somewhat  unexpectedly 
on  the  scene.  This  approach  indicates  that  the  future  of  the  work 
with  mutationally  altered  proteins  probably  lies  in  the  realm  of 


64  Information  Storage  and  Neural  Control 

protein  chemistry — in  the  effect  of  amino  acid  substitutions  on 
protein  structure  and  specificity — and  in  the  confirmation  of  the 
correctness  of  the  biochemical  approach  which  we  shall  now 
consider,  rather  than  in  the  determination  of  the  code  for  each 
amino  acid. 

THE  MECHANISM  OF  PROTEIN  SYNTHESIS  AND  THE 
BIOCHEMICAL  APPROACH  TO  THE  GENETIC  CODE 

The  attempt  to  understand  the  intei  mediate  steps  by  which 
genetic  information  is  transferred  into  specific  protein  structure 
obviously  poses  a  very  interesting  biological  problem.  As  recently 
as  a  year  ago,  however,  no  one  would  have  predicted  that  a  crude 
cell-free  extract  of  E.  coli  could  be  forced,  even  in  principle,  to 
yield  precise  information  about  the  genetic  code.  The  discovery 
which  revolutionized  the  coding  search  and  opened  what  might 
be  called  the  biochemical  approach  to  the  genetic  code  was  the 
finding  of  Nirenberg  and  Matthaei  (11)  that  one  could  trick  the 
E.  coli  extract  into  making  a  most  unnatural  protein — the  polyamino 
acid  polyphenylalanine — by  adding  a  most  unnatural  piece  of 
genetic  material — the  polyribonucleic  acid  of  uridylic  acid  (poly  U). 

To  explain  the  significance  of  this  experiment,  it  is  necessary 
first  to  describe  briefly  present  ideas  on  the  mechanism  of  protein 
synthesis.  There  is  believed  to  be  a  flow  of  information  from  DNA 
through  RNA  to  protein  involving  three  classes  of  RNA:  ribosomal 
RNA,  transfer  RNA,  and  "messenger"  RNA.  Chemically,  all  of 
these  RNA's  are  polymers  with  a  sugar-phosphate  backbone  like 
DNA,  but  with  ribose  sugar  instead  of  deoxyribose,  and  with  the 
base  uracil  (U)  instead  of  thymine  (T). 

Ribosomal  RNA  exists  in  the  cell  in  cytoplasmic  ribonucleo- 
protein  particles  (ribosomes),  which  are  generally  considered  to 
be  the  cellular  sites  of  protein  synthesis  (12).  Messenger  RNA  is 
assumed  to  carry  the  genetic  information  detailing  the  specific 
amino  acid  sequence  of  the  protein  from  the  DNA  to  the  ribosome. 
Presumably  the  messenger  RNA  binds  to  the  non-specific  ribosome 
(probably  to  ribosomal  RNA)  and  serves  as  the  information 
bearing  "template"  for  protein  synthesis  (13).  Transfer  RNA's 
bind  amino  acids  specifically  (with  the  aid  of  enzymes).  They  are 


Genetic  Cotitrol  of  Protein  Synthesis 


65 


thought  to  carry  amino  acids  to  the  ribosome-messenger  complex 
and  to  act  as  an  "adapter"  to  position  amino  acids  in  the  proper 
place  for  their  polymerization  into  specific  proteins  (12).  The 
transfer  RNA  presumably  "recognizes"  the  code  for  a  particular 
amino  acid  in  the  messenger  RNA  in  order  to  provide  specific 
positioning  of  the  amino  acid. 

The  evidence  that  transfer  RNA  is  an  intermediate  in  protein 
synthesis  is  very  good,  at  least  in  in  vitro  systems,  and  there  is 
strong  experimental  support  for  the  idea  that  ribosomes  are  the 
site  of  protein  synthesis  from  both  in  vivo  and  in  vitro  studies  (12). 
The  question  of  whether  there  is  a  distinct  messenger  RNA  loosely 
attached  to  nonspecific  ribosomes,  or  whether  the  genetically 
specific  RNA  is  built  into  ribosomes  as  they  are  synthesized,  giving 
specific  ribosomes,  is  still  a  matter  of  some  controversy.  In  the 
case  of  virus  infected  E.  co/i,  there  is  strong  evidence  favoring  the 
loosely  bound  messenger  view  (14).  At  present  the  model  described 
(and  shown  schematically  in  Figure  3)  is  the  most  adequate  to  ex- 
plain existing  experimental  results. 


.    DNA 


i 


poly 


U 


T-RNH-AA 

\ 


Protem  T-RNVAA 
/  ^ 


Fig.  3.  Schematic  representation  of  the  normal  protein  synthesizing  system  (on 
the  left)  and  the  synthetic  system  (on  the  right).  DNA  has  been  removed  from 
the  synthetic  system  by  the  enzyme  deo.xyribonuclease  and  the  synthetic  mes- 
senger poly  U  replaces  the  normal  messenger  RNA. 


One  can  imagine  a  simple  base-pairing  mechanism  by  which 
all  of  this  can  occur.  The  messenger  RNA  may  be  synthesized 
with  a  DNA  primer  by  a  base-pairing,  enzyme-catalyzed  process 
which  produces  a  "complementary"  copy  or  translation  of  the 


66 


Information  Storage  and  Neural  Control 


DNA  in  which  each  base  in  the  RNA  is  the  complement  to  eacli 
base  in  the  DNA.  For  example,  the  sequence  ATGC  in  DNA 
would  be  translated  into  UACG  in  the  RNA  because  U,  replacing 
T  in  RNA  but  having  similar  base  pairing  properties  will  form  a 
hydrogen-bonded  base  pair  with  A,  A  with  T,  C  with  G,  and 
G  with  C.  An  enzyme  has  been  found  which  appears  to  catalyze 
this  process.  Messenger  RNA  may  bind  to  ribosomal  RNA  by 
means  of  rather  general  regions  of  base  complementarity.  Finally, 
transfer  RNA  need  only  have  a  base  sequence  complementary 
to  the  messenger  RNA  base  code  for  its  particular  amino  acid 
to  fulfill  its  function,  since  pairing  of  the  complementary  bases 
will  correctly  position  the  amino  acid.  If  the  DNA  sequence  AAA 
codes  the  amino  acid  phenylalanine,  then  the  messenger  RNA 
will  have  the  complementary  sequence  UUU  and  the  transfer 
RNA  for  phenylalanine  a  sequence  AAA.  The  UUU  sequence 
in  the  messenger  RNA  will  pair  with  the  AAA  sequence  in  the 
transfer  RNA  to  provide  for  the  insertion  of  phenylalanine  into 
its  genetically  determined  site  in  the  protein  (Fig.  4).  The  gene 
DNA  and  its  messenger  RNA  are  equivalent  in  informational 
content,  since  one  is  a  direct  translation  of  the  other. 


C 


.((>Ala 


lAirrAiriATA 

t     •     I     I     •     I 

iu  iu  iu  iu  iu  i  m-pna 

RNA  Flare 


Ribosomal  5ur?cxce 

Fig.  4.  Hypothetical  base  pairing  scheme  for  protein  synthesis.  Poly  U  is  shown 
in  its  role  of  messenger.  The  poly  U  chain  binds  loosely  to  a  segment  of  ribosomal 
RNA  flaring  out  from  the  "protein  surface"  of  the  ribosome.  Transfer  RNA  for 
phenylalanine  is  presumed  to  contain  an  AAA  sequence  complementary  to  the 
UUU  of  the  poly  U  and  therefore  "positions"  a  sequence  of  phenylalanines  for 
polymerization  into  polyphenylalanine. 


Genetic  Control  of  Protein  Synthesis  67 

The  implication  of  the  Nirenberg  experiment  is  that  polyuridylic 
acid  is  the  messenger  RNA  for  polyphenylalanine  and  that  a 
sequence  of  U  is  the  messenger  RNA  code  for  plienylalanine  (or  a 
sequence  of  A  is  the  DNA  code).  The  "synthetic"  and  "normal" 
systems  are  compared  in  Figure  3.  Since  there  exists  an  enzyme, 
discovered  by  Ochoa  and  Grunberg-Manago  (15),  which  will 
catalyze  the  random  synthesis  of  ribonucleotides  into  a  polymer, 
there  is  now  a  very  powerful  tool  available  for  investigating  the 
genetic  code.  For  example,  a  mixed  polymer  of  A  and  U  provides 
for  sequences  of  AAA,  AAU,  AUA,  UAA,  AUU,  UAU,  UUA, 
and  UUU,  choosing  only  triplets  for  purposes  of  illustration, 
(It  should  be  noted  tiiat  at  least  three  bases  per  amino  acid  are 
required  if  four  bases  are  to  specify  twenty  amino  acids.)  If  poly 
AU  is  added  as  a  synthetic  messenger,  then  amino  acids  coded  by 
the  above  triplets  will  be  incorporated  into  a  polypeptide  chain. 
Even  if  some  of  the  triplets  are  "nonsense"  in  that  they  do  not 
specify  an  amino  acid,  by  using  a  large  excess  of  U  some  poly- 
peptide formation  can  be  assured  by  providing  a  polyphenyl- 
alanine 'handle"  so  that  those  triplets  which  spell  an  amino  acid 
will  not  be  lost. 

This  approach  has  been  pursued  very  successfully  by  the  Ochoa 
and  Nirenberg  groups  to  describe  the  most  probable  code  letter 
for  fourteen  of  the  twenty  amino  acids  (16,  17).  To  carry  out  the 
synthetic  messenger  experiment,  the  coli  extract  is  first  treated 
("preincubated")  to  remove  existing  messenger  RNA.  Existing 
DNA  is  removed  by  the  enzyme  deoxyribonuclease  so  that  new 
messenger  RNA  cannot  be  synthesized.  Then  synthetic  messenger 
RNA  is  added,  and  the  amount  of  C^^  amino  acid  incorporated 
into  protein-like  material  (insoluble  in  trichloroacetic  acid)  is  deter- 
mined by  radioactivity  measurements.  Any  significant  incorpora- 
tion of  a  CI'^  amino  acid,  using  the  UA  polymer  as  the  messenger 
RNA,  implies  that  the  code  for  that  amino  acid  consists  of  some 
combination  of  A  and  U  or  of  a  sequence  of  A.  One  can  then 
hope  to  separate  a  2U1A  from  a  1U2A  or  a  3A  code  by  deter- 
mining the  ratio  of  the  observed  incorporation  of  a  given  amino 
acid  to  that  of  phenylalanine  and  comparing  this  ratio  with  that 
expected  for  the  calculated  number  of  3U,  2U1A,  1U2A,  and 
3x\  sequences  (using  a  polymer  with  U  in  large  excess  so  that  the 


68  Information  Storage  and  Neural  Control 

numbers  will  be  quite  different,  and  assuming  that  3U  is  the  code 
for  phenylalanine). 

The  way  to  complete  the  determination  of  the  genetic  code  by 
discovering  the  actual  sequence  of  bases  is  also  clear  in  principle 
using  the  biochemical  approach.  It  should  be  possible  to  add 
small,  known  ribonucleotide  sequences  to  poly  U  enzymatically 
and  to  use  these  messengers  to  produce  polyphenylalanine  plus 
the  amino  acids  coded  by  these  sequences  (if  any).  Unless  there  are 
some  large  surprises  lurking  around  the  corner,  the  genetic  code 
for  E.  coll  may  well  be  officially  solved  within  the  next  three 
years  or  so.  There  remains  the  question  of  whether  the  coli  code 
is  common  to  all  organisms,  although  most  of  the  limited  infor- 
mation available  argues  for  universality.  Even  if  the  code  is 
different  in  higher  organisms,  the  techniques  evolved  for  the  coli 
system  should  be  generally  applicable.  All  that  is  needed  is  a 
crude,  cell-free,  protein-synthesizing  system  plus  the  proper  syn- 
thetic messenger  to  trick  the  system. 

CONTROL  OF  THE  RATE  OF  PROTEIN  SYNTHESIS 

(The  Regulatory  Problem) 

The  process  by  which  genetic  information  is  converted  into 
protein  specificity  is  rapidly  becoming  spelled  out,  and  the  com- 
plete unraveling  of  the  exact  nature  of  the  genetic  code  providing 
this  specificity  of  protein  structure  is  on  the  horizon.  However, 
the  genetic  control  necessary  to  provide  for  the  adaptive  skill  of 
the  microorganism  and  for  the  much  more  complicated  growth 
pattern  of  the  differentiated  organism  cannot  be  accounted  for 
simply  by  the  ability  of  genes  to  control  protein  structure.  The 
structural  gene,  structural  messenger,  and  ribosome  constitute  a 
protein  factory,  always  working  at  the  same  rate  for  all  proteins. 
It  seems  obvious  that  there  must  exist  regulatory  genes  involved 
in  turning  on  and  off  the  structural  genes  and  in  varying  the 
enzyme  complement  of  the  cell. 

Recent  work  with  bacteria  has  shown  the  existence  of  genes 
which  serve  to  control  the  rate  of  protein  synthesis  in  response  to 
changes  in  external  conditions.  Normally,  the  production  of  the 
lactose-hydrolyzing    enzyme    ^-galactosidase    by    the    bacterium 


Genetic  Control  of  Protein  Synthesis  69 

E.  coli  can  vary  roughly  a  thousand-fold  up  to  a  maximum  of 
some  6  per  cent  of  the  cellular  protein  if  a  jS-galactoside,  an 
"inducer,"  is  present  in  the  growth  medium.  Mutants  have  been 
isolated  which  have  lost  this  control  of  the  rate  of  enzyme  syn- 
thesis. Jacob  and  Monod  (13)  have  divided  these  "constitutive" 
mutants  into  two  genetically  and  functionally  distinct  classes 
designated  i^  and  o^  By  studying  the  dominance  properties  of 
these  mutations  in  partially  diploid  strains  carrying  both  i+  and 
i~  and  both  0+  and  o*"  (both  inducible  and  constitutive  genetic 
structures)  Jacob  and  Monod  have  developed  a  model  of  the 
control  process  (Fig.  5).  This  model  proposes  that  a  "repressor" 
material  is  made  under  the  control  of  the  i  gene,  which  they  call 
a  regulator  gene,  and  that  this  repressor  binds  to  a  site  near  the 
structural  gene,  the  O  or  operator  gene,  preventing  formation  of 
the  structural  messenger. 

Requlator  Operator   jtructural 

Gene  bene  Genes 


Repression  or  Tnduction       Proteins 

Fig.  5.  The  model  proposed  by  Jacob  and  Monod  for  the  mechanism  controlling 
the  rate  of  action  of  the  structural  gene.  A  repressor  material  is  made  under  the 
control  of  the  DNA  of  the  regulator  gene.  This  repressor  material  acts  (after 
possible  metabolite  activation)  by  binding  to  a  DNA  site  adjacent  to  the  struc- 
tural gene  or  genes  subject  to  rate  control  by  the  repressor  and  preventing  the 
formation  of  structural  messengers. 

In  this  model  the  regulation  is  negative;  genes  are  noimally 
functional  and  are  turned  off  by  a  repressor.  Similar  analysis  of 
the   system   controlling   alkaline    phosphatase   synthesis    (18)    has 


70  Information  Storage  and  Neural  Control 

indicated  that  a  gene  involved  in  the  regulation  of  this  enzyme 
can  hav^e  a  positive  effect,  i.e.,  can  be  involved  in  turning  on  a 
gene  to  its  full  capacity.  Therefore,  the  generality  of  the  model 
proposed  for  the  /3-galactosidase  system  is  not  at  present  estab- 
lished; but  the  essential  feature  of  the  model — the  proposed 
existence  of  specific  gene  products  which  exert  a  controlling 
influence  over  the  rate  of  synthesis  of  the  structural  messenger 
RNA — is  very  appealing.  It  serves  as  a  valuable  guide  to  future 
experimental  efforts  aimed  at  trying  to  understand  the  control 
process  at  a  chemical  level. 

CONCLUDING  COMMENTS 

We  have  considered:  1)  the  chemical  nature  of  the  gene; 
2)  the  "sequence  hypothesis"  which  serves  as  the  basis  for  our 
definition  of  the  genetic  coding  problem;  3)  the  evidence  sup- 
porting the  sequence  hypothesis  from  combined  genetic  and 
chemical  studies;  4)  the  recent  rather  dramatic  progress  of  the 
biochemical  approach;  and,  finally,  5)  the  problem  of  regulation. 
We  cannot  at  present  unequivocally  separate  fact  from  fancy. 
However,  the  evidence  now  extant  certainly  favors  our  main 
conclusions:  1)  that  the  genetic  information  of  an  organism  is 
contained  in  the  base  sequence  of  its  DNA;  2)  that  the  base 
sequence  of  the  DNA  of  "structural  genes"  specifies  the  amino 
acid  sequence  of  proteins;  3)  that  an  RNA  "messenger"  carries 
the  genetic  information  from  the  structural  gene  to  the  ribosome 
for  protein  synthesis;  and,  finally,  4)  that  the  base  sequence  of 
the  DNA  of  certain  "regulator  genes"  specifies  a  material  which 
exerts  a  controlling  influence  over  the  rate  of  protein  synthesis. 
It  should  be  emphasized,  however,  that  most  of  the  evidence  for 
these  conclusions  comes  from  work  with  microorganisms  and  that 
the  generalization  to  higher  organisms  is  chiefly  an  act  of  faith. 

REFERENCES 

1.  Horowitz,  N.   H.:   Biochemical  genetics  of  neurospora,  Advances  in 

Genetics,  J.' 33,  1950. 

2.  Levinthal,  C:  Coding  aspects  of  protein  synthesis.  Revs.  Mod.  Physics, 

J7.-249,  1959. 


Genetic  Control  of  Protein  Synthesis  71 

3.  Ris,  H.:   The  Chemical  Basis  of  Heredity,  ed.  by  McElroy  and  Glass. 

Baltimore,  Johns  Hopkins  Press,  1957. 

4.  \Vatson,  J.  D.,  and  Crick,  F.  H.  C:  The  structure  of  DNA.  Cold 

Spr.  Harb.  Symp.  Qiiant.  Biol,  7S.T23,  1953. 

5.  Benzer,  S.:  On  the  topology  of  the  genetic  fine  structure.  Proc.  Nat. 

Acad.  Sci.,   Wash.,  45:1607,  1959. 

6.  Crick,  F.  H.   C:  On  protein  synthesis.  Symp.  Sac.  Expt.  Biol.,    12: 

138,   1958. 

7.  Ingram,  V.  M.:  Hemoglobin  and  Its  Abnormalities,  Springfield,  Thomas, 

1961. 

8.  Rothman,  F.:  Cold  Spr.  Harb.  Symp.  Qjiant.  Biol,  26.-1961,  in  press. 

9.  Yanofsky,  C,  Helinski,  R.,  Mahng,  B.:  Ibid. 

10.  Wittman,  H.  G.:  Comparison  of  the  tryptic  peptides  of  chemically 

induced  and  spontaneous  mutants  of  tobacco  mosaic  virus.  Virology^ 
12:609,  1960. 

11.  Nirenberg,  M.,  and  Matthaei,  J.  H.:  The  dependence  of  cell-free 

protein  synthesis  in  E.  coli  upon  naturally  occurring  or  synthetic 
polyribonucleotides.   Proc.  Nat.  Acad.  Sci.,    Wash.,  47.-1588,   1961. 

12.  Berg,  P.:  Specificity  in  protein  synthesis.  Ann.  Rev.  Biochem.,  30:29 o, 

1961. 

13.  Jacob,  F.,   and  Monod.  J.:   Genetic  regulatory  mechanism  in  the 

synthesis  of  proteins.  J.  Mol.  Biol,  3.-318,  1961. 

14.  Brenner,  S.,  Jacob,  F.,  and  Meselson,  M.:  An  unstable  intermediate 

carrying  information  from  genes  to  ribosomes  for  protem  synthesis. 
Nature,  190:576,  1961. 

15.  Grunberg-Manago,   M.,   and  Ochoa,   S.:   Enzymatic  synthesis  and 

breakdown  of  polynucleotides;  polynucleotide  phosphorylase.  J. 
Am.  Chem.  Soc,  77.-3165,  1955. 

16.  Lengyel,   P.,  Speyer,  J.   F.,  Basilio,   C,  and  Ochoa,  S.:   Synthetic 

polynucleotides  and  the  amino  acid  code,  IV.  Proc.  Nat.  Acad. 
.Sci.',   Wash.,  48:282,  1962. 

17.  Martin,  R.  G.,  Matthaei,  J.  H.,  Jones,  O.  W.,  and  Nirenberg,  M.  W.: 

Ribonucleotide  composition  of  the  generic  code.  Biochem.  and 
Biophys.  Research  Comm.,  6.-410,  1962. 

18.  Garen,  A.,  and  Echols,  H.:  Properties  of  two  regulating  genes  for 

alkahne  phosphatase.  J.  Bad.,  83:291,  1962. 

19.  Kendrew,  J.  C,  Dickerson,  R.  E.,  Strandberg,  B.  E.,  Hart,  R.  G., 

Davies,  D.  R.,  Phillips,  D.  C,  and  Shore,  V.  C:  Structure  of 
Myoglobin.  A  three-dimensional  Fourier  synthesis  at  2  A  reso- 
lution. Nature,  185:422,  1960. 


72  Information  Storage  and  Neural  Control 

DISCUSSION  OF  CHAPTER  IV 

Mike  McGlothlen  (Houston,  Texas) :  What  about  suppressor 
genes  where  you  have  a  mutation  of  the  structural  gene  and  then 
a  counter-mutation  of  the  type  that  causes  the  still  mutated 
structural  gene  to  produce  normal  enzymes? 

Harrison  Echols  (Madison,  Wisconsin):  The  theory  which  is 
now  usually  advanced  to  explain  these  suppressor  mutations  is 
that  a  suppressor  is  a  mutation  which  has  affected  the  translation 
mechanism;  i.e.,  it  has  perhaps  affected  the  ability  of  the  soluble 
RNA  to  bind  the  correct  amino  acid.  The  soluble  RNA  then  makes 
mistakes  which  partially  rectify  the  mutational  mistake.  For 
example,  suppose  that  the  original  change  in  the  protein  was  a 
substitution  of  the  amino  acid  alanine  for  glycine  and  that  the 
suppressor  mutation  is  such  that  some  of  the  time,  in  protein 
synthesis,  glycine  is  put  back  in  place  of  alanine.  In  this  case, 
you  would  now  get  a  reduced  level  of  the  original  premutation 
type  of  protein.  Certain  suppressors  may  also  involve  a  change  in 
the  concentration  of  some  cell  constituents,  which  leads  to  the 
activation  of  a  mutationally  altered  protein. 

McGlothlen:  Would  you  care  to  say  anything  about  the  origin 
of  the  secondary  and  tertiary  structure  of  proteins?  Presumably, 
the  sequence  of  amino  acids  is  controlled  by  sequences  in  DNA, 
but  what  about  the  folding,  etc.,  that  produces  the  active  un- 
denatured  form  of  an  enzyme? 

Echols:  We  think  that  this  comes  about  purely  from  a  deter- 
mination of  the  primary  structure.  The  secondary  structure  is  a 
matter  of  solution  thermodynamics.  A  repeating  chain  of  amino 
acids  forms  an  alpha  helix  if  the  solvent  is  not  too  hard  on  hydrogen 
bonds.  To  get  the  specific  three-dimensional  structure  is  a  tougher 
problem.  However,  we  can  imagine  that  as  the  newly  synthesized 
protein  comes  off  the  ribosome,  there  are  regions  of  the  protein 
which  are  capable  of  bonding  and  are  in  very  close  proximity  to 
each  other.  There  is  actually  some  evidence  that  the  primary 
structure  does  determine  the  three-dimensional  structure  of  the 
protein  ribonuclease.  This  derives  from  experiments  in  which  the 
protein  is  unfolded  and  then  caused  to  fold  again.  One  can  break 
all  four  of  the  disulfide  bonds  by  reduction  to  SH,  and  unwind 
the  protein  into  a  completely  random  coil.   One  would  expect, 


Genetic  Control  of  Protein  Synthesis  73 

just  by  considering  random  re-formation  of  four  disulfide  bonds, 
tliat  105  possible  alternative  forms  of  the  protein  should  exist. 
However,  what  is  found  is  that  oxidation  to  bring  back  the  disulfide 
bonds  produces  something  like  90  per  cent  enzymatically  active 
protein.  So  even  though  the  system  is  far  from  physiological,  with 
this  protein,  at  least,  one  appears  to  get  the  total  three-dimensional 
structure  purely  from  the  primary  structure. 

Arthur  Shapiro  (New  York,  New  York):  A  protein  is,  of 
course,  made  up  primarily  of  amino  acid  chains;  but  most  pro- 
teins, particularly  the  specific  ones— enzyme  proteins —  do  contain 
polysaccharides  and  do  contain  very  specific,  important,  and 
critical  terminal  groups.  Is  it  the  general  notion  now  that  the 
DNA  chain  contains  the  information  that  determines  these  non- 
amino  acid  fractions  in  the  protein  molecule  as  well,  or  is  this 
supposed  to  be  a  different  kind  of  thing?  If  so,  is  there  any  clue 
as  to  what? 

Echols;  In  general,  as  proteins  and  enzymes  have  been  purified 
more  carefully  and  more  successfully,  they  have  been  found  in 
most  cases  to  contain  nothing  except  amino  acids.  I  would  feel 
that  the  appearance  of  a  sugar  group  or  some  other  moiety  attached 
to  a  protein  would  either  be  a  nonspecific  accident  or  the  result 
of  a  specific  site  built  into  the  amino-acid-determined  structure 
of  the  protein.  In  other  words,  I  think  the  complete  specificity 
of  the  protein  comes  about  because  the  DNA  specifies  the  amino 
acid  sequence. 

Frank  Morrell  (Palo  Alto,  California):  We  know  of  some 
agents  that  can  alter  DNA,  such  as  x-ray,  etc.  Could  you  elaborate 
on  the  sorts  of  agents  that  can  selectively  alter  base  sequence  in 
RNA? 

Echols:  One  which  is  widely  used  is  nitrous  acid.  This  removes 
the  amino  groups  from  bases,  producing  a  change  of  the  base 
cytosine  into  the  base  uracil  in  RNA. 

Morrell:  What  is  the  consequence  of  alteration  of  the  base 
sequence  in  RNA  without  simultaneous  alteration  of  DNA? 

Echols:  Mutagenic  agents  which  aff'ect  RNA  but  not  DNA  are 
not  known.  Thus,  in  treating  a  bacterium  with  a  mutagenic 
chemical,  both  the  DNA  and  the  RNA  are  involved.  To  my 
knowledge,  no  one  has  been  able  to  purify  a  messenger  RNA. 


74  Information  Storage  and  Neural  Control 

Until  this  is  done,  it  will  not  be  possible  to  study  the  effects  of 
specific  agents  on  the  RNA  and  on  the  resulting  proteins.  The 
only  kinds  of  messenger  RNA's  which  are  available  are  the  syn- 
thetic ones. 

James  E.  Darnell,  Jr.  (Cambridge,  Massachusetts):  The  viral 
RNA's  were  the  first  to  be  treated  with  deaminating  agents. 
Schuster's  work  with  tobacco  mosaic  virus  (TMV),  which  showed 
that  deamination  of  cytosine  resulted  in  mutation,  confirmed  the 
fact  that  the  chemical  change  is  preserved  in  the  progeny  particles. 
The  deamination,  which  results  in  mutation  of  TMV  particles 
and  change  in  the  protein  code,  is  assumed  to  be  preserved  from  the 
initial  change  in  the  RNA  of  the  virus.  If  one  considers  viral 
RNA's  as  messengers,  which  they  are,  then  this  type  of  messenger, 
at  least,  can  be  treated  with  mutagens,  and  the  residual  damage, 
if  you  will,  can  be  preserved. 

Echols:  I  suppose  I  am  being  unfair  to  the  viral  RNA's,  although 
it  has  not  been  clearly  established  that  the  viral  RNA  functions 
directly  as  a  messenger. 

Heather  D.  Mayor  (Houston,  Texas):  There  are  clear  in- 
dications from  viral  RNA  that  the  seat  of  genetic  information 
can  be  RNA  as  well  as  DNA.  I  think  there  is  quite  good  evidence 
that  an  RNA  virus  can  act  as  its  own  messenger.  I  should  like  to 
ask  Dr.  Echols  if  he  has  any  information  from  mutations  on  closely 
positioned  bases  in  DNA.  Is  there  any  evidence  that  the  code  is 
indeed  a  triplet  sequence  rather  than,  say,  a  multiple  of  six?  In 
fact,  are  there  any  data  indicating  that  six  bases  could  represent 
the  fundamental  unit  of  the  code? 

Echols:  I  think  that  there  is  no  compelling  evidence,  even 
taking  Crick's  work  into  consideration,  defining  the  size  of  the 
coding  unit.  However,  the  work  of  Nirenberg  and  Ochoa  cer- 
tainly suggests  that  the  number  of  bases  which  code  an  amino 
acid  cannot  be  an  exceedingly  large  number.  If  a  large  number 
were  required,  the  only  thing  which  should  promote  incorporation 
of  most  amino  acids  would  be  a  polymer  containing  all  four  bases. 

Mayor:  If  you  have  four  bases  and  a  triplet  code,  you  could 
possibly  get  codes  for  sixty-four  different  amino  acids  instead  of 
the  twenty  we  know  to  be  involved.  Do  you  think  that  dif- 
ferent combinations  of  bases  may  code  the  same  amino  acids? 


Genetic  Control  of  Protein  Synthesis  75 

Echols:  The  work  which  Nirenberg  and  Ochoa  have  done 
suggests  that  there  may  be  a  degeneracy  in  the  code;  i.e.,  there 
may  be  more  than  one  coding  unit  which  specifies  a  given  amino 
acid.  They  only  use  polymers  containing  uracil,  but  they  have 
already  worked  up  to  nineteen  or  twenty  amino  acids.  This 
suggests  that  there  is  either  a  degeneracy  in  the  code  or  a  rather 
surprising  selection  for  "sense"  codes  containing  uracil.  Also, 
incorporation  of  at  least  one  amino  acid  is  promoted  by  polyiners 
containing  different  coding  units. 


CHAPTER 
V 

CODING  BY  PURINE  AND 

PYRIMIDINE  MOIETIES  IN  ANIMALS, 

PLANTS,  AND  BACTERIA* 

Saul  Kit,  Ph.D. 


T 


INTRODUCTION 


HE  transfer  of  genetic  information  in  biological  systems  may 
be  represented  by  the  following  schematic  diagram: 


DNA 

DNA 

Filial  cells 


Parental  cell 


Figure  1 


Implicit  in  this  schematic  diagram  are  three  concepts:  1)  that 
genetic  information  is  coded  in  deoxyribonucleic  acids  (DNA); 
2)  that  the  transfer  of  genetic  information  from  parental  to  filial 
cells  involves  the  replication  of  the  DNA  and  distribution  of 
"equal"  amounts  to  the  daughter  cells;  and  3)  that  the  expression 
of  genetic  potential  within  a  cell  involves  the  transcription  of 
information    from    DNA    to    "Informational"    ribonucleic    acids 


*This  investigation  was  aided  in  part  by  grants  from  the  American  Cancer  Society, 
The  Leukemia  Society,  Inc.,  and  The  National  Cancer  Institute  (CY-4064,  CY-4238). 

76 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria 


11 


(RNA)  and  the  translation  of  the  Informational-RNA  for  specific 
protein  synthesis. 

The  key  to  the  understanding  of  tliese  concepts  is  tlie  model  for 
the  structure  of  DNA  proposed  in  1953  by  Watson  and  Crick  (77). 
Watson  and  Crick  suggested  that  DNA  consists  of  two  helical 
polynucleotide  chains  of  opposite  polarity  which  are  twined  round 
one  another  (Fig.  2).  Each  chain  is  built  from  four  mononucleo- 
tide units  which  are  joined  together  by  3',  5'  phosphodiester 
bonds.  Each  of  the  four  mononucleotide  units  consists  of  either  a 


Fig.  2.  Simplified  model  of  die  DNA  double  helix  showing  hydrogen  bonding 

and  the  DNA  strands  of  opposite  polarity.  P  =  orthophosphate;  S  =  deoxyribose; 

A  =  adenine;  T  =  thymine;  G  =  guanine;  G  =  cytosine. 


78 


V/ 


Information  Storage  and  Neural  Control 
OH  K  .OH 


HjC O— PO3H2 


HjC O PO3H2 


Ribose    Phosphate  Deoxyribose   Phosphate 

Fig.  3.  Chemical  formulas  for  tlie  pentose  phosphates  found  in  DNA  and  RNA. 


N^C — NH. 


N=C— OH 


N — C — NH 


^CH 


N— C— N 


>" 


N: 


■NH, 


HC       C — N^  HjN-C     C— N^  0=C 


CH 


H— N CH 


Adenine  (A) 


Guanine  (G) 


Cytosine  (C) 


N==C — OH 
0=C         CH 
H N CH 


N:=:C OH 


0=C       C CH, 


H N — CH 


N: 


-NH. 


o=c 


■CH, 


H— N CH 


Uracil  (U) 


Thymine  (T)  Methyl   Cytosine  (MC) 


Fig.  4.  Chemical  formulas  for  the  purine  and  pyrimidine  bases  found   in  DNA 

and  RNA. 


Pyrimidine  Moieties  i?i  Animals,  Plants,  and  Bacteria  79 

purine  or  a  pyrimidine  base  connected  in  nucleoside  linkage  to 
deoxyribose-5 '-phosphate  (Fig.  3).  The  purine  and  pyrimidine 
bases  are  adenine  (A),  guanine  (G),  cytosine  (C),  and  thymine  (T) 
(Fig.  4).  In  addition,  methyl  cytosine  (MC)  may  partly  replace 
cytosine  in  the  DNA  of  certain  plant  and  animal  cells  and  glu- 
cosylated  hydroxymethylcytosine  may  replace  cytosine  in  the 
DNA  of  the  T-even  bacteriophages.  The  four  bases  A,  G,  C,  and 
T  are  the  symbols  of  the  genetic  alphabet,  just  as  dot  and  dash 
are  the  symbols  of  the  Morse  code.  Triplets  of  bases,  such  as  TTT 
or  TGC,  may  be  the  letters  of  the  genetic  alphabet  and  each 
tri]3let  may  specify  a  particular  amino  acid  of  a  protein  chain. 
Tims,  the  sequence  of  triplets  along  a  polynucleotide  chain  would 
determine  the  amino  acid  sequence  of  a  protein. 

The  two  DNA  chains  are  held  together  by  hydrogen  bonds 
between  the  bases,  each  base  being  joined  to  a  companion  base 
on  the  other  chain  (Fig.  5).  The  pairing"  of  the  bases  is  specific, 


HO 


OH 


o  /^  HO-P  =  0 


0=d-OH      X      r         -\j  ^  ^ 


VvO^   0    — ,v 


HO  H 


o 


T       A  H   OH 


OH 

I  HO 


0»d-OH 


O x2 


E  I 


_x cf     V°V 

HO  H  I 


-a  M 


0    >>/ 


G  C  H  OH 

Fig.  5.  Hydrogen  bonding  between  deoxyadenylic  acid  and  thymidylic  acid  and 
between  deoxyguanylic  acid  and  deoxycytidylic  acid. 


80  Information  Storage  and  Neural  Control 

adenine  (A)  going  with  thymine  (T)  and  guanine  (G)  going  with 
cytosine  (C).  The  phosphate  groups  of  the  DNA  chains  are 
accessible  to  hydrogen  or  hydroxyl  ions  and  to  dyes  and  are, 
therefore,  on  the  outside,  whereas  the  bases  occur  opposite  one 
another  on  the  inside.  From  x-ray  diffraction  studies,  it  has  been 
deduced  that  there  is  a  succession  of  flat  nucleotides  spaced 
3.36  A  apart  and  standing  out  perpendicular  to  the  fiber  axis. 
The  structure  is  relatively  rigid  and  serves  as  a  template  for  either 
its  own  replication  or  for  the  replication  of  "Informational"  RNA. 
Plausible  mechanisms  for  DNA  replication  and  for  spontaneous 
mutations  were  embodied  in  the  proposals  of  Watson  and  Crick 
(77).  These  mechanisms  are  strongly  supported  by  a  large  number 
of  experiments.  According  to  the  proposal  for  DNA  replication, 
a  twin  stranded  DNA  molecule  partially  unwinds,  and  each  base 
attracts  a  complementary  free  nucleotide  already  available  for 
polymerization  within  the  cell.  These  free  nucleotides,  whose 
phosphate  groups  already  possess  the  free  energy  necessary  for 
polyesterification,  then  link  up  with  one  another,  after  being  held 
in  place  by  the  parental  template  chains,  to  form  a  new  poly- 
nucleotide molecule  of  the  required  nucleotide  sequence.  Thus, 
each  DNA  strand  serves  as  a  template  for  the  synthesis  of  a  com- 
plementary strand.  The  replication  process  can  be  schematically 
represented  as  follows: 

-A-C-T-G-->-A-C-T-G- 

-A-c-T-G-/*  :    :    :    : 

'•••  -T-G-A-C- 

-T-G-A-C-\ 

-T-G-A-C-->-T-G-A-C- 
Parental  DNA  '.'.'. 

Duplex  DNA  Chains  .... 

After  Unwinding        -A-C-T-G- 

New  DNA 
Duplexes 

It  is  a  corollary  of  the  Watson-Crick  hypothesis  that  a  change  of 
one  or  a  few  nucleotides  in  the  DNA  sequence  will  be  mutagenic. 
Mechanisms   for   spontaneous   mutation    and    for   experimentally 


Pyrimidine  Aioieties  in  Animals,  Plants,  and  Bacteria  81 

induced  mutations  have  been  suggested  on  the  basis  of  this  con- 
cept. Watson  and  Crick  (77)  pointed  out  that  the  specificity  in 
DNA  structure  (adenine  pairing  with  thymine  and  guanine  with 
cytosine)  resuhs  from  the  assumption  that  each  of  the  bases 
possesses  one  tautomeric  form  which  is  very  much  more  stable 
than  any  of  the  otlier  possibilities.  The  fact  that  the  compound  is 
tautomeric,  however,  means  that  the  hydrogen  atoms  can  occasion- 
ally change  their  location.  Thus,  a  spontaneous  mutation  might 
be  caused  by  a  base  occurring  very  occasionally  in  one  of  the  less 
likely  tautomeric  forms  at  the  moment  when  tlie  complementary 
chain  is  being  formed.  For  example,  whereas  adenine  will  nor- 
mally pair  with  thymine,  if  there  is  a  tautomeric  shift  of  one  of 
its  hydrogen  atoms,  it  can  pair  with  cytosine.  The  next  time 
pairing  occurs,  the  adenine  (having  resumed  its  more  usual 
tautomeric  form)  will  pair  with  thymine,  but  the  cytosine  will 
pair  with  guanine,  and  so  a  change  in  the  sequence  of  bases  will 
have  occurred. 

Mutations  may  also  be  induced  by  chemical  agents.  Let  us 
consider  two  categories  of  chemically  induced  mutations:  1)  those 
which  result  from  the  conversion  of  one  nucleotide  base  in  a 
DNA  chain  co  another  nucleotide  base  (transition),  and  2)  those 
which  result  from  the  deletion  of  a  base  from  the  chain. 

The  conversion  of  cytosine  to  thymine  may  be  effected  by 
adding  nitrous  acid  to  cells  (58).  Cytosine  is  deaminated  by 
nitrous  acid  to  uracil.  When  DNA  which  has  been  deaminated 
in  this  way  undergoes  replication,  the  uracil  will  attract  adenine 
during  the  complementary  base  pairing.  The  next  time  pairing 
occurs  (second  cycle  of  replication)  the  adenine  which  had  paired 
with  uracil  will  now  pair  with  thymine.  Hence,  following  two 
cycles  of  DNA  replication,  the  original  cytosine-guanine  base  pair 
will  have  been  converted  to  a  thymine-adenine  base  pair. 

A  mutagenic  change  from  thymine  to  cytosine  may  be  induced 
in  cells  by  the  use  of  the  thymidine  analog,  bromodeoxyuridine  (21). 

A  nucleotide  base  change  in  the  DNA  chain  expresses  itself 
as  an  amino  acid  change  in  the  protein  chain  whose  synthesis  is 
controlled  by  the  altered  DNA. 

A  third  mutagenic  agent,  nitrogen  mustard,  may  alkylate  some 
of  the  guanine  groups  of  the  DNA  chain  at  the  N(7;  position. 


82  Informatio7i  Storage  and  Neural  Control 

Two  stages  of  degradation  follow:  7-alkylguanine  splits  off,  and 
a  slow  fission  of  the  sugar  phosphate  chain  follows  (41).  If  in  the 
process  of  replication  of  the  DNA  which  has  been  exposed  to 
nitrogen  mustard,  the  guanine  of  the  template  is  skipped,  the 
resulting  sequence  of  nucleotide  bases  in  the  daughter  chain  will 
be  altered.  Another  mutagen  which  apparently  acts  by  deleting 
a  base  from  the  DNA  chain  is  proflavin,  an  acridine  derivative, 
A  series  of  T-4  bacteriophage  mutants  induced  by  proflavin  have 
been  studied  by  Crick  et  al.  (13).  The  mutants  in  almost  all  cases 
manifest  a  complete  inactivation  of  the  function  of  the  eene 

Equations  1  through  3  depict  schematically  some  current  ideas 
about  genetic  coding  (13): 

(TRIPLET) 

I  '   I  II  1  I 1  I 1  I 1   I 1 

[1]    TGCTGCTGGTGCTGGTGGTGA--- 

I  -  '  — '-I— I— 1  —  1— I  — l-_l_l_l_l_l_l_l_l_|_|__L  I 
Starting  ' 

point  Normal  DNA  Ghain 


I  I  I  1  I 1 1 1  r 


"1  r 


[2]    TGGTGTTGGTGGTGGTGGTGA   - 

I — I — I — I — l^-l I I I I I ^1 L_l l_l__l_L_l._[ 

Starting:  T 


^iD 


point  Gonversion  of  G  to  T  by  Nitrous  Acid 

(Gene  may  still  be  functional,  amino  acid  in 
protein  chain  is  changed). 


[3]TGGTGTGGTGGTGGTGGTGA      

I — ' — I— I— I— 1^1 I I I L_l I I l_I_^l_^l__I_| 

Starting  T 

point  G  deleted  from  Ghain  by  Proflavin  or  Nitrooen 

Mustard 

(Gene  inactivated). 

It  is  assumed  that  a  triplet  of  three  nucleotide  bases  (for  example, 
TGG)  codes  each  particular  amino  acid  in  the  protein  chain.  It  is 
further  assumed  that  the  DNA  chain  is  translated  from  a  fixed 
starting  point  and  that  the  genetic  code  is  nonover  lap  ping.  If  one  base 
in  the  chain  is  altered  (for  example,  the  change  of  G  in  the  second 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  83 

triplet  of  Equation  1  to  T  in  the  second  triplet  of  Equation  2), 
only  one  amino  acid  in  the  resulting  protein  chain  will  be  altered; 
that  is,  the  amino  acid  coded  by  TGT  replaces  the  amino  acid 
coded  by  TGC.  The  protein  with  the  changed  amino  acid  may 
still  be  functional  or  partly  functional. 

If  a  nucleotide  base  is  deleted  following  nitrogen  mustard  or 
proflavin  treatment,  the  gene  is  inactivated.  In  Equation  3,  it  is 
seen  that  G  has  been  deleted  from  the  second  triplet  of  Ecjuation  1. 
With  a  nonoverlapping  code,  the  second  triplet  now  becomes 
TCT,  the  third  triplet  becomes  GCT,  etc.  In  other  words,  all 
triplets  from  TCT  on  are  changed.  Thus,  all  amino  acids  in  the 
protein  chain  which  are  coded  by  the  second  triplet  to  the  last 
triplet  are  changed,  and  the  new  protein  chain  cannot  function. 
As  a  result,  the  gene  controlling  the  synthesis  of  that  protein  has 
been  inactivated. 

For  a  more  detailed  discussion  of  mutation  mechanisms  at  the 
chemical  level,  the  reader  is  referred  to  the  papers  of  Freese  (21), 
Lawley  (41)  and  Crick  et  al.  (13). 

It  is  apparent  that  a  knowledge  of  the  number  of  DNA  molecules 
in  a  given  cell  and  of  the  entire  nucleotide  sequence  of  each  mole- 
cule, along  with  the  code  by  which  DNA  and  RNA  sequences  are 
translated  to  the  amino  acid  sequences  of  proteins,  would  suffice  as  a 
"blue  print"  for  describing  any  organism.  Such  a  total  description 
is,  of  course,  not  available  to  us  as  yet.  We  do,  however,  know  cer- 
tain characteristics  of  the  DNA  of  many  species  of  bacteria,  plants, 
animals,  and  viruses.  The  amount  of  DNA  per  cell  is  known  in  many 
instances.  This,  in  a  sense,  tells  us  how  "thick"  each  "genetic  book 
of  instructions"  is.  We  also  have  knowledge  of  the  average  nucleo- 
tide composition  of  the  DNA  of  different  species,  that  is,  of  how  fre- 
quently the  "alphabet  symbols"  are  repeated  in  each  book.  There  is 
also  some  knowledge  of  the  range  of  composition  within  a  particular 
cell.  These  topics  will  be  discussed  in  the  second  section  of  this 
paper. 

In  the  third  section  of  this  paper,  I  shall  consider  the  composition 
of  RNA  molecules  and  the  proposed  mechanisms  for  transcribing 
information  from  DNA  to  RNxA.  Finally,  I  shall  briefly  touch  on 
the  translation  of  the  DNA-RNA  code  to  amino  acid  sequences 
of  proteins. 


84 


Information  Storage  and  Neural  Control 


THE  DEOXYRIBONUCLEIC  ACIDS  (DNA) 
Amount  of  DNA  per  Cell 

The  amount  of  DNA  per  cell  varies  greatly  from  the  simplest 
to  the  most  complex  organisms.  Some  representative  values 
are  shown  in  Table  I.  Mammals,  reptiles  and  amphibians  often 
contain  about  six  picograms  of  DNA  per  cell;  but  many  fish 
and  birds  contain  only  about  a  third  of  this  amount.  Bacteria  and 
fungi  cells  have  about  1/100  the  amount  of  DNA  found  in  the 
higher  animals.  Larger  viruses  such  as  vaccinia  and  T-even  bac- 
teriophage have  about  1/10,000  the  amount  of  DNA  per  particle 
and  the  smallest  viruses,  such  as  bacteriophage  0X174  and  the 
Shope  papilloma  virus  have  one  millionth  the  amount  of  DNA 
per  particle  found  in  the  cells  of  higher  animals. 

TABLE  I 

DNA  Content  Per  Cell  of  Various  Species 
(All  Values  Expressed  as  Picograms  DNA  Per  Cell  or  Virus  Particle) 


2.6x10-6 

Reference 

<^X174  phage 

(65,  66) 

T-2  phage 

2x10-" 

(66) 

Rabbit  papilloma  virus  (Shope) 

6.6xl0-« 

(56) 

Vaccinia  virus 

3  X  10-^ 

(56) 

E.  coli  B  (log  phase) 

0.0137 

(26) 

Clostridium 

0.0245 

(75) 

Yeast  (diploid) 

0.05 

(50) 

Neurospora 

0.017 

(47) 

Fish 

Shark 

5.46 

(75) 

Sturgeon 

3.2 

(75) 

Carp 

3.49 

(75) 

Perch 

1.9 

(75) 

Catfish 

1.89 

(75) 

Barracuda 

1.37 

(75) 

Amphibians 

Frog 

15.0 

(75) 

Toad 

7.33 

(75) 

Reptiles 

Green  turtle 

5.27 

(75) 

Wood  turtle 

4.92 

(75) 

Alligator 

4.98 

(75) 

Birds 

Domestic  fowl 

2.34 

(75) 

Guinea  hen 

2.27 

(75) 

Mammals 

Man 

6.8 

(75) 

Rabbit 

6.5 

(75) 

Rat 

5.7 

(75) 

Mouse 

5.0 

(75) 

Pyrimidnie  Moieties  in  Animals,  Plants,  and  Bacteria  85 

Since  the  molecular  weight  of  Shope  papilloma  virus  is  about 
4x  10"  (78)  and  the  molecular  weight  of  an  average  nucleotide 
base  pair  is  about  600,  it  is  obvious  that  Shope  papilloma  virus 
has  a  total  of  about  4  x  10V6  x  10",  or  6600  nucleotide  base  pairs 
in  its  DNA.  Vaccinia  virus  has  roughly  300,000  base  pairs;  bac- 
teria have  roughly  20x10*^  base  pairs,  and  mammalian  cells  a 
total  of  about  7x10^  base  pairs.  If  there  are  no  restrictions  as  to 
the  proportion  in  which  base  pairs  occur  or  as  to  the  sequence  in 
which  they  occur,  the  number  of  different  DNA  molecules  possible 
is  4",  where  n  is  the  number  of  base  pairs.  Thus  it  is  clear  that 
DNA  provides  an  adequate  basis  for  gene  specificity. 

The  increase  in  the  relative  amount  of  DNA  from  the  lowest  to 
the  highest  forms  of  life  reflects  the  need  for  an  increasing  number 
of  genetic  units  for  embryogenesis  and  differentiation  and  for 
various  regulatory  mechanisms. 

How  Large  are  DNA  Molecules? 

The  molecular  weight  of  Shope  papilloma  virus  is  about  4x10'' 
(78).  The  weight  average  molecular  weights  of  most  of  the  DNA 
preparations  which  have  been  studied  are  about  5-14  x  10^  How- 
ever, the  molecular  weight  of  DNA  may  actually  be  much  greater 
than  this.  Very  high  molecular  weight  DNA  has  been  isolated 
from  the  T-even  bacteriophages  and  it  is  possible  that  the  entire 
genome  of  the  T-even  bacteriophages  consists  of  one  long  DNA 
chain  having  a  molecular  weight  of  90x10''  to  150x10"  (15). 
There  is  reason  to  believe  that  very  long  DNA  molecules  are 
partially  fragmented  to  smaller  pieces  when  they  are  isolated  from 
tissues  and  viruses. 

Average  Composition  of  DNA 

The  average  nucleotide  base  composition  of  DNA  molecules 
may  be  measured  by  hydrolyzing  the  DNA  and  then  measuring 
the  nucleotide  bases  after  they  have  been  resolved  by  paper 
chromatography,  ion  exchange  chromatography,  or  paper  elec- 
trophoresis. To  the  extent  that  a  mixture  of  DNA  molecules  can 
be  partially  separated,  the  distribution  of  base  compositions  among 
the  molecules  of  the  mixture  can  also  be  estimated. 

Two  other  important  methods  are  available  for  measuring  the 
molar  nucleotide  composition  of  DNA:  1)  equilibrium  sedimenta- 


86 


Injormation  Storage  and  Neural  Control 


< 

Q 

H 

< 

Z 

o 

.1^ 


o 


* 

o 

R 

+ 

^^ 

s^ 

^;^ 

^ 

<^ 

ic: 

G 

-s^ 

sr- 

«u 

-w 

Ml 

W 

it 

l=/l 

o 

-^ 

e- 

^ 

q 

CN]f\l-i^->^TfTtul\OVOr--COCOOOOOCNCNCNCNCN^-iT--0]01M010!tN]MT+Tj-LriiOv-iO 


r-'T-.tococ<-ir<-i':t-LnLnsoi^r-r^[~-oooooooooooo^-^^^^^T-Hroco^^ij-)Ln 

C^C7^C^C^C^C^C^C^C^C^C^C^C^C^C^C^C^C^C^OOOOOOOOOOOOOOC^ 


< 

o 

So 

Oh  ro 

Q 

Q 

O 

Group 

9484 
RD 

75  a 

Group 

35 
3478 

<  -*  O]  J 

<!,  p:,  HH  ^  ^  c\  ~~^<N  T^ 


a  a  o 


h  2  «  S 


S  6 

3    3 

'-3  '-v 


P    3 

^  '-a 


ETC  oj  '^  .2 

M    IJ    >    C    in 

;i3  bp  >.  oj  C 
"3  "^  ^  'b  3 

P  =^  S  3  ^ 
fs  o  2  «  jS 


en    3 

3X! 


.2-2  S 

O    rt 


.cti 


>-        Hj  ^ 
P  3    O 

^r,  cj    c^   ^^ 

p  03    CL  en  -O 

0)  (/]    „    3    t^ 


■O    C;    3    «    3 
•-<    tn    C    —    -^  ' 

y  2i  3' 


'   _   tH   to:  3   brj  tn 
>  3  n  3  o  >  -s 

■^i='~—  oj^o—  oo-^n 

.,     .„_     m  —  .^c^sS^rt^iJ'-" 

UU22OM0HpqOi-If0c/3fiHKpqEEc^ciI| 


o  'S 


fl    tn  .22 

a  en  r3 


I  °-c  o 

O      lU      Sh 
I   _Q    ^O     ^ 


i-.^-i-rio'-'i:—  5 
o  o  o  o  x.y  o  y  ^ 


3    3-5 


^28=  10^:3^^33  = 


&Hh-lh-l 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria 


87 


JS 
ho 

S 


-0 


^°l^ 
>-i^ 

w 


;coQ, 


>t^ 


CA  m  o  o 


CN  iri  -;  . 

3-  u-iH 


3-^ 


■c 

u 
br 

rt 

OJ 

C 

0 

__, 

,_, 

,_, 

5J 

5-' 

C 

a: 

C 

0 

0 
0 

0 

0 

0 

0 

rt 

02 

d 

a7i 

6 

c« 

m 

« 

« 

"C 

.5 

-n 

2 

0 

03 

tj 

"o 

"u 

3 

«j 

'-H 

lU 

0 

<u 

C 

i-l 

u 

vi 

ryi 

i-< 

r/l 

^ 

XI  _a 

V 

0 

i; 

U 

u 

u 

0 

0-c 

Z  0  iz;  >  2;  w  w  w  a 

'H  o 


w;^ 


O    >.  w   nj   3 

^  ^  c  c  •- 

six;   n!   03    « 
Di  c/3  c/5  C/3  W 


u  e  S  '-' 

S  S  C  X 

crt  O  a; 

ij ,.  e  w  s 


XI  i:  isi 


;     3   3   ^   3 
0   3  C   i^  h 


!  Ji  (« 


p  73  «  -M  3  u 

J5    O    U    l)  s^   3 

M  w  03  o  «j  r; 

.-y)  oa  CL,  <  Q  « 


3   3 

11 

O    2j 

.2,0 


o 


"3   n!   C 


3  3  3 

r  r-  ^  ■ 

S  S  c« 

D  _j  ^j 


T3    3 


0 


2  t;  ii  -T.  « 


iOc^p: 


£ 

X 

73 

5J 

G 

n 

0    0 

-1 

3 

> 

a 

U 

F 

s  a 

X3 
0 
N 

£ 

0 
■a 
0 

X  0 
0  -a 
0  3 

0  0 
a  H. 

C 

G 

J3 

X 

u 

u    u 

C^Qpi 

^ 

Oh  C^  t^  c)? 

JJ  X 

o 


ss 


-^  X 


^  r 


88  Information  Storage  and  Neural  Control 

tion  in  CsCl  density  gradients;  and  2)  the  estimation  of  the  melting" 
temperature  (T,„)  of  DNA  by  a  study  of  the  change  in  DNA 
absorption  as  a  function  of  temperature  (16).  The  latter  two 
methods  have  the  advantages  that  less  material  is  required  to 
analyze  the  DNA  and  that  an  estimate  of  the  variation  from  the 
mean  of  nucleotide  base  composition  can  be  made.  Of  the  three 
methods,  density  gradient  centrifugation  has  the  highest  accuracy, 
requires  the  least  DNA  per  experiment,  and  permits  the  detection 
of  DNA  molecules  of  unusual  base  composition  even  where  the 
latter  comprise  less  than  5  per  cent  of  the  total  DNA.  A  further 
discussion  of  density  gradient  centrifugation  will  be  presented  later. 

Since  DNA  is  double  stranded*  and  the  guanine  and  adenine 
of  each  chain  are  paired,  respectively,  with  the  cytosine  and 
thymine  of  the  complementary  chains,  the  total  purine  bases 
(A+G)  are  equal  to  the  total  pyrimidine  bases  (C+T)  and  the 
total  6-amino  bases  (C+A)  are  equal  to  the  total  6-keto  bases 
(G+T).  However,  the  ratios  of  guanine  plus  cytosine  to  adenine 
plus  thymine  (G+C)/A+T)  are  not  the  same  in  different  or- 
ganisms and  provide  a  parameter  by  which  organisms  can  be 
characterized. 

The  molar  (G  +  C)  content  of  the  DNA  of  seventy-two  different 
bacterial  species,  thirteen  species  of  higher  plants,  ten  species  of 
algae,  four  species  of  fungi,  two  species  of  protozoa,  sixteen  species 
of  invertebrates,  twenty-three  species  of  animals,  twelve  bacterial 
viruses,  six  animal  viruses  and  rickettsiae,  and  twelve  insect 
viruses  have  been  measured  and  are  shown  in  Tables  II  through  IX. 

The  molar  per  cent  (G+C)  varies  from  26.5  per  cent  in  the 
protozoan,  Tetrahymena,  to  73  molar  per  cent  (G  +  C)  for  the 
bacterium,  Mycobacterium  phlei.  Of  the  170  species  listed  in  Tables 
II  through  IX,  112  have  DNA  molecules  whose  average  molar 
(G  +  C)  contents  are  40  to  60  per  cent.  This  is  not  surprising.  It  is 
probable  that  Kutagenic  transitions  of  adenine-thymine  to  qua- 
nine-cystosine  base  pairs  in  one  part  of  a  DNA  molecule  are 
compensated  for  by  transitions  from  quanine-cystosine  to  adenine- 
thymine  base  pairs  in  other  parts  of  the  molecule  so  that  the  molar 
per  cent  (G  +  C)  remains,  on  the  average,  close  to  50  per  cent. 


*Exceptions  to  this  statement  are  the  DNA  of   the  bacteriophages,    0X174  and 
SI 3,  which  are  single  stranded   (65,  66). 


Pvrimidine  Moieties  in  Animals,  Plants,  and  Bacteria 


89 


X 

^_, 

C5 

K 

ffi 

td 

« 

o 

«J' 

oi 

H 

> 

^ 


QJ 


^   ^   r-<    r^   00   ■r-   ■^ 


ON   CM   00   CO   ^-1 
Tf  lO  CO  lO  NO 


\ocMcor^r^t-~-oocNNO      -^-^ 
ssoooooocM-^r^oooo 

COT)--*'<t'<^'^'<^'*'*"*'*"* 


in  t^  o  o  00 


rt  b:  o  5  &.  c  a^s  «j  h 


O     >,J2     >     !- 


2  '^-' 

^    > 


S"^!^    ^?    « 


>  JS      .  —    « 


I  S  "rt  -S  J3    3    (^    u   03 
i<C/3CU&hUO-i<C/:3 


riaticum 

ipiens 

•denschel 

1    ■? 

m                ?3 

3     «-^ 

33 

•s  0-^1 

"2  H  S 

c^ 

fci    M  'S    3 

n3    u   ^ 
C    4^    0 

03 

Ddictyon  i 
trodesmu 
ella  ellips 
desmus  q 

0  u-S 
■a  2  ^ 

g    rt    cfl    « 

O 

^--2g 

S  -C  j::  J= 

1    >, 

§:  K  <  O  c/3 

"SoiOH 

SO 

D  ii 


P3 


-  .s 

3    >- 
U    2 

3  6 
u  S 


cq       Ci 


90 


Information  Storage  and  Neural  Control 


<o 


c 

"^ 

w 

^ 

tV) 

^^ 

+ 

■g 

ei 

>i 

f^" 

w 

e) 

•^ 

1 

■&^ 


CM  CS  CM  Tl- 


so  r^ 


■b  ,    '^ 

all 
S  c  " 

^  n!    «    rt 
"^    O 

Q  ^fo 


C«    C 


,1 

3   w  M  >ii     o   3   «  M  V 


>  ^ 


oooo-*oooo-<too-^oooo 


CNvOCM^cOcOOOMO-* 


o 


O  Ln  M  fO  un  r^ 
o  o  o  o  o  o 
r-  r~-  r^  t~~  r-~  r- 


8 -a 


f  o  ,-^*   • 

S"  >  DM 

■'-'  "^  w  +-< 

r-  •«  S-i    4J    5" 

a  nj  ij  -^  i! 

^  dJ  ■•-'     o    *^ 


^   O    i-   oj   «   2 


3 

3 
C 

3 

3 

c 

-0 

^^ 

p 

>< 

S 

3 

Echinoc 
Psamme 
Oak  mo 
Paraccn 
Arbacia 
Silk  woi 
Arbacia 
House  f 

ii  S 


S   ii   3; 


S  6 


« j> JQU 


|l3(jCCoCCu 


t^    B3 


Pyriffiidine  Moieties  in  Animals,  Plants,  and  Bacteria 


91 


o  txi  o:^  o:) 


m  00  00  00  ro 


fe§&§^°fe? 

&?fe5 

^ 

£5&-5 

Tj-  (N]  (N]  ri 

rO  ■" 

CO 

TT  -"T 

Tj-  tT  ^  ^ 

^  Tl- 

-:!■ 

■^  Tt- 

&5&°fe?£?&?£^°^^°&§fe?       ^6?^fe?65^^^ 


(^1  o 

o  o 


O  O 

o  o 


o  o 


r^  r^  Tj-  Lo 
■^  m  Tj-  Tj- 

o  o  o  o 
o  o  o  o 

o  o  o  o 


00OOC->OOO'>-'Oi-' 
CsOOCnOCsOOOO 


s  s 


V  v         JJ 


gi 


00 


„  c 


'H.'a 


3   3 

E  B 


CJ 


^  c 

be- 

t'  ^ 

c 

-?  -^  ^ 

N    U    U 

<^    a 

gl^ 

5^1H 

«    w 

^  si  VI 

QUO 

QJ 

G 

G;-D  bo 


2 

•—         tti           re  B       n 

u           -— -          j3  re       nj 

5 

'-'            ^             .,  -C       ;3 
■-             u        ^  «  c        P, 

^ 

^ 

a 

man 

Lass  1 

bull 

calf 

monk 

dog 

rabbi 

Chine 

goldc 

rat 

O        O 


m 

^ 

o 

o 

d 

I--  CO 

^  o 

CN  o  o  cv  1 

so  i~~  r^  so 

'~-'— 

-^^ 

U 

eg 

J^  3 

CW^ 

M      r^ 

c 

^  4^ 

V 

aj  re 

"a 

en 

-S3 

bB* 

C 

C  '^ 

IH 

3^=; 

^ 

-J 

i-T 

.S  u 

dj 

re  JJ 

> 

u    •> 

JD  ;s 

bp 

'a 

re 

« 

V 

M 

c 

3 

bO 

O 

B 

o 


92 


Information  Storage  and  Neural  Control 


TABLE  VII 
The  Densities  and  Base  Compositions  of  Bacteriophage  DNA 


Density  Gradi 

'ent 

Centrifu 

■gat  ion 

Chemical  Analysis 

Phage 

Density  gcm-^ 

% 

,  G+C 

Reference 

%G+C 

Reference 

T2  + 

1.700 

35 

(66) 

T4  + 

1.698 

34.4 

(66) 

T6  + 

1.707 

34.2 

(66) 

T5 

1.702 

43 

(60) 

39 

(66) 

0X174  +  +  + 

1.72 

43 

(65) 

44 

(66) 

Salmonella  Al  +  + 

43.4 

(66) 

Phage  alpha 

1.704 

44 

(11) 

42.5 

(11) 

Tl 

1.705 

46 

(60) 

48 

(66) 

T7 

1.710 

51 

(60) 

48 

(66) 

T3 

1,712 

53 

(60) 

49.6 

(66) 

Xvir 

1.710 

51 

(60) 

50 

(66) 

P22 

50 

(66) 

+T2,  T4,  T6  contain  hydroxymethylcytosine 
+  + Preliminary  or  tentative  data 
+  +  +  Single  stranded  DNA 

TABLE  VIII 
DNA  Base  Composition  of  Animal  Viruses  and  Rickettsia 


Density  Gradient  Centrifugation 


Chemical  Analysis 


Virus 

Density  gcm-^ 
1.714 

%G+C 
50 

Reference 
(78) 

%  G+C 

Reference 

Shope  papilloma 

49 

(78) 

Vaccinia 

1.698 

39 

(38) 

40.6 

(10) 

Fowl  pox 

38 

(52) 

Rickettsia  burnetti 

1.704 

45 

(60) 

45 

(10) 

Rickettsia  prowazeki 

30.8 

(10) 

Rickettsia  rickettsii 

37.5 

(10) 

TABLE  IX 
Base  Composition  of  DNA  of  Insect  Viruses  (3) 


Host  Species 

%G+C 

Polyhedral  viruses 

Porthetria  dispar 

58.7 

Lymantria  monacha 

51.5 

Clioristoneura  fumiferana 

51.2 

Ptychopoda  seriata 

47.6 

Malacosoma  americanum 

42.7 

Malacosoma  disstria 

42.2 

Bombyx  mori 

42.7 

Colias  philodice  eurytheme 

42.5 

Neodiprion  sertifer 

37.3 

Tipula  poluclosa  Merg.  (T.I.V.)  * 

31.5 

Capsular  viruses 

Cacoecia  muriana 

37.6 

Choristoneura  fumiferana 

34.8 

"Reference  (74). 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  93 

The  variation  in  composition  of  DNA  molecules  among  dif- 
ferent species  of  microorganisms  is  very  great.  C/ostricIium perfringens 
has  only  32  molar  per  cent  (G  +  C)  while  at  the  other  extreme 
of  the  distribution,  Alycobacterium  phlei  has  73  molar  per  cent 
(G  +  C)  (Table  II).  Fungi  vary  from  36  to  54  per  cent  (G  +  C) 
in  the  four  species  investigated  and  the  two  protozoan  strains  so 
far  measured  {Tetrahymena  and  Euglena)  contain  26.5  and  47  molar 
per  cent  (G+C),  respectively  (Table  IV).  The  range  of  average 
DNA  values  among  algae  is  also  rather  great — 36.9  molar  per 
cent  (G+C)  for  diatomic  algae  to  64  molar  per  cent  (G+C)  for 
the  green  alga,  Chlamydomonas  reinhardi  (Table  III).  The  dis- 
tribution of  average  values  among  different  species  of  higher 
plants  (Table  III)  and  invertebrates  (Table  V)  is  much  narrower: 
Plant  DNA  composition  varies  from  35  molar  per  cent  (G+C) 
for  tobacco  leaves  to  48.4  molar  per  cent  for  Triticum  vulgare,  and 
ainong  invertebrate  species  the  values  vary  from  34.9  molar  per 
cent  for  the  echinoderm,  Echinocardium  cordatum  to  44  molar  per 
cent  (G  +  C)  for  the  crab,  Carcinus  maenas. 

The  range  of  values  is  very  narrow  indeed  for  the  average 
composition  of  the  twenty-three  vertebrate  DNA  species  so  far 
examined.  The  values  range  from  40  to  44  molar  per  cent  (G  +  C) 
(Table  VI).  The  range  for  DNA  aniinal  viruses  appears  to  be 
greater  than  that  for  the  host  animal  species:  38  molar  per  cent 
(G+C)  for  fowl  pox  virus  and  50  molar  per  cent  (G+C)  for  the 
Shope  papilloma  virus  of  rabbits  (Table  VIII).  Various  insect 
viruses  manifest  in  their  DNA  average  (G  +  C)  contents  of  31.5  to 
58.7  per  cent  (Table  IX),  a  range  which  is  also  somewhat  broader 
than  the  compositions  of  the  few  insect  DNA's  so  far  studied. 

The  T-even  bacteriophage  DNA's  have  about  35  molar  per 
cent  (G  +  C),  a  value  outside  the  range  of  the  bacterial  host  in 
which  the  viruses  grow  {E.  coli,  51  per  cent  (G+C)  (Table  VII). 
The  DNA  from  a  number  of  other  bacteriophages  (T-1,  T-3,  T-7) 
and  the  lysogenic  phages  (X,  P22,  salmonella  Al)  have  average 
molar  (G+C)  contents  which  are  very  similar  to  those  of  the 
bacterial  host  cells  {E.  coli.  Shigella,  Salmonella). 

Base  Composition  of  DNA  Strands 

Since  a  DNA  molecule  consists  of  two  complementary  nucleotide 
strands  of  opposite  polarity,  the  question  arises  as  to  whether  the 


94 


Information  Storage  and  Neural  Control 


C3H      LUNG 


C3H      BRAIN 

Fig.  6.  Ultraviolet  photograph  of  the  banding  of  mouse  DNA  and  Streptomyces 
viridochromogenes  DNA  in  a  CsCl  density  gradient.  The  photograph  was  taken 
after  centrifugation  for  twenty-four  hours  at  25°  at  44,770  rev.  per  minute.  The 
Streptomyces  band  (p25°  =[l.729  gcm-^)  appears  at  the  left.  Two  bands  with 
mean  densities  of  1.701  and  1.690  gcm-^  were  obtained  with  mouse  DNA.  The 
narrow  band  at  the  right  is  due  to  the  meniscus  of  the  solution. 


two  Strands  have  similar  or  grossly  dissimilar  average  nucleotide 
compositions.  Only  fragmentary  data  are  presently  available. 
There  are  indications  that  the  DNA  strands  of  animal  cells  differ 
slightly  in  thymine,  and  hence,  in  adenine  content  (14).  The 
DNA  strands  of  the   bacteriophage,   alpha,   have   been  separated 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria 


95 


and  shown  to  differ  in  density  (11).  The  density  differences  could 
reflect  differences  in  the  base  compositions  of  tlie  strands.  Tliere 
are  alternative  explanations,  however.  For  example,  one  strand 
might  contain  more  glucose  residues  attached  to  hydroxymethyl- 
cytosine  than  tlie  other  strand. 

Equilibrium  Sedimentation  of  DNA  in  CsCl  Density  Gradients 

Equilibrium  sedimentation  of  DNA  in  CsCl  density  gradients 
provides  an  elegant  method  for  characterizing  DNA  with  respect 
to  average  nucleotide  composition  and  heterogeneity  of  composi- 
tion (46).  Following  the  centrifugation  of  a  DNA  solution  for 
twenty-four  hours  in  a  CsCl  density  gradient,  the  DNA  tends  to 
form  a  band  at  a  position  in  the  cell  corresponding  to  its  effective 
buoyant  density.  A  photograph,  taken  with  ultraviolet  light,  of  the 
banding  of  bacterial  and  mouse  DNA  is  shown  in  Figure  6,  and 
a  microphotodensitometer  tracing  of  the  photograph  is  presented 
in  Figure  7.  The  white  (ultraviolet  light  absorbing  areas)  in  the 


24  Hours,  25°C  (CsCI-p=  1.7208) 


24  Hours,  25°C  (Cs  CI- p=  1.7165) 


Fig.  7.  Microdensitometer  tracing  of  photograph  of  the  banding  of  mouse  DNA 
and  Streptomyces  DNA  after  density  gradient  centrifugation.  See  Figure  6. 


96  Information  Storage  and  Neural  Control 

center  part  of  the  photograph  correspond  to  Streptomyces  virido- 
chromogenes  DNA  and  to  two  mouse  DNA  bands,  respectively. 
The  effective  mean  buoyant  density  of  the  Streptomyces  virido- 
chromogenes  DNA  band  is  1.729  gcm~^  The  corresponding  values 
for  the  principal  and  minor  mouse  DNA  bands  are  1.701  and 
1.690  gcm~^  (Table  II  and  Table  VI).  The  mean  densities  of  the 
bands  can  be  measured  with  an  accuracy  of    ±0.001   gcm~^ 

It  has  been  shown  by  Rolfe  and  Meselson  (54)  and  by  Schild- 
kraut  et  al.  (60)  that  the  mean  effective  buoyant  densities  of  double 
stranded  DNA  bands  vary  linearly  with  the  molar  per  cent  (G+C) 
content  of  the  DNA.  For  example,  Streptomyces  viridochromogenes 
DNA  (density  =  1.729  gcm~^)  contains  73  molar  per  cent  (G+C), 
Escherichia  coli  DNA  (density  =  1.710  gcm"^)  contains  51  molar 
per  cent  (G+C),  and  mouse  DNA  (density  =  1.701  gcm~^)  con- 
tains 42  molar  per  cent  (G+C).  Thus,  if  the  density  of  DNA  is 
measured  by  equilibrium  sedimentation  in  CsCl,  the  molar  per 
cent  (G  +  C)  can  be  calculated.  TABLES  II  to  VIII  show  the 
densities  of  DNA  preparations  from  various  sources  and  the 
agreement  between  molar  per  cent  (G  +  C)  as  calculated  from 
the  densities  of  the  bands  and  as  determined  directly  by  chemical 
analyses. 

The  standard  deviations  of  the  DNA  bands  expressed  in  density 
units  (a  )  depend  upon  at  least  two  factors:  1)  the  molecular 
size  of  the  DNA,  and  2)  the  heterogeneity  of  DNA  composition 
within  the  sample.  This  follows  from  the  following  considerations. 
The  centrifugal  field  tends  to  drive  the  DNA  into  a  region  where 
the  sum  of  the  forces  acting  on  a  given  molecule  is  zero.  This 
concentrating  tendency  is  opposed  by  Brownian  motion,  with  the 
result  that  at  equilibrium,  the  macromolecules  are  distributed  with 
respect  to  concentration  in  a  band  of  width  inversely  related  to 
their  molecular  weight  (46). 

When  a  DNA  population  consists  of  molecules  which  differ 
considerably  in  density  and  in  molar  per  cent  (G+C),  discrete 
and  nonoverlapping  DNA  bands  may  be  formed.  For  example, 
two  discrete  DNA  bands  are  formed  with  mouse  DNA  (Figs.  6,  7) 
and  there  is  no  overlapping  between  the  mouse  DNA  (p  =  1.701 
gcm"^)  and  the  Streptomyces  (p  =  1.729  gcm"^)  DNA  bands.  On 
the  other  hand,  if  a  DNA  preparation  consists  of  a  heterogeneous 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  97 

population  of  molecules  differing  only  slightly  in  density  and  in 
molar  per  cent  (G  +  C),  the  DNA  bands  will  overlap,  and  there 
will  be  an  increase  in  the  overall  standard  deviation  of  the  band. 
So  long  as  the  total  variance  of  a  band  and  the  number  average 
molecular  weight  of  a  DNA  preparation  are  known,  it  is  possible 
to  calculate  the  contribution  which  heterogeneity  of  composition 
makes  to  the  band  width  (16,  36,  54).  An  independent  estimate 
of  the  heterogeneity  of  composition  of  a  DNA  preparation  may 
also  be  made  from  the  DNA  thermal  denaturation  curves.  These 
independent  estimates  agree  satisfactorily. 

Nonoverlapping  Bands 

Density  gradient  centrifugation  experiments  have  permitted  a 
number  of  interesting  conclusions  concerning  DNA.  First,  as 
already  mentioned,  the  DNA  obtained  from  many  bacterial 
species  form  discrete  bands  which  do  not  overlap.  These  observa- 
tions indicate  that  the  respective  organisms  have  no  DNA  mole- 
cules with  common  density,  and,  by  inference,  that  they  have  no 
DNA  molecules  with  common  nucleotide  base  composition  (16). 
Since  the  metabolism  and  replication  of  bacteria  do  have  much  in 
common,  it  is  generally  thought  that  many  of  their  proteins  should 
be  identical  or  very  similar.  Yet  all  current  discussions  of  the 
way  in  which  DNA  can  control  the  sequence  of  amino  acids  in 
proteins  require  a  direct  correlation  between  the  composition  of 
the  DNA  and  of  the  protein.  There  are  a  number  of  ways  in 
which  this  dilemma  can  be  resolved.  The  simplest  approach  is  to 
assume  that  some  amino  acids  are  coded  by  more  than  one  nucleo- 
tide triplet.  For  example,  each  of  the  triplets  UCC  and  UAC 
might  specify  the  amino  acid  threonine.  Thus,  the  dependence  of 
the  amino  acid  composition  of  proteins  on  the  DNA  nucleotide 
composition  would  not  be  exacting  and  DNA  molecules  having 
different  compositions  could  code  very  similar  proteins.  A  genetic 
code  in  which  two  or  more  nucleotide  triplets  are  used  to  code  a 
particular  amino  acid  is  called  a  '"degenerate"  code  (13). 

There  are  a  number  of  arguments  which  can  be  advanced  in 
favor  of  this  hypothesis.  By  studying  nitrous  acid  induced  mutants 
of  bacteriophage,  Tessman  (73)  demonstrated  that  each  of  the 
complementary  DNA  strands  is  functional.  The  experiments  "* of 


98  Information  Storage  and  Neural  Control 

Chamberlin  and  Berg  (7)  suggest  that  genetic  information  can 
be  transcribed  from  either  of  the  two  DNA  strands  to  infor- 
mational-RNA.  Thus,  when  single  stranded  0X174  DNA  was 
used  as  a  template  for  RNA  polymerase,  an  RNA  strand  having 
a  composition  complementary  to  the  0X174  DNA  was  formed. 
However,  when  single  stranded  0X174  DNA  was  used  as  a  tem- 
plate for  DNA  polymerase,  double  stranded  DNA  was  formed. 
The  double  stranded  0X174  DNA  could  then  be  used  as  a  primer 
for  RNA  polymerase  and,  in  this  case,  RNA  was  formed  having 
a  composition  identical  to  that  of  double  stranded  0X174  DNA. 
If  both  of  the  complementary  DNA  strands  are  ultimately  trans- 
lated from  the  same  fixed  starting  point,  the  foregoing  experi- 
ments would  suggest  that  complementary  nucleotide  triplets  code 
the  same  amino  acid.  On  the  other  hand,  if  the  complementary 
DNA  strands  are  translated  from  opposite  starting  points,  the 
results  would  point  to  one  of  two  possibilities:  1)  that  two  dif- 
ferent nucleotide  triplets  code  the  same  amino  acid,  or  2)  that  the 
complementary  DNA  strands  are  identical  even  though  they  have 
opposite  polarities.  The  second  of  these  possibilities  would  further 
require  that  one  half  of  a  given  DNA  strand  be  complementary  to 
the  opposite  half.  The  latter  restrictions  do  not  seem  to  apply 
to  the  0X174  DNA  studied  by  Chamberlin  and  Berg  (7,  65). 
Davern's  experiments  (14)  also  suggest  that  these  restrictions  are 
unlikely  as  a  general  proposition.  Hence,  some  form  of  a  "degen- 
erate" genetic  code  seems  to  be  the  most  appealing  hypothesis  at 
this  time. 

Unimodal  and  Bimodal  Distributions 

The  DNA  from  almost  every  species  so  far  examined  forms  one 
discrete  band  after  density  gradient  centrifugation.  Streptomyces 
viridochrornogenes  DNA  (Figs.  6,  7)  illustrates  this  unimodal  dis- 
tribution. Several  examples  have  now  been  found  in  which  DNA 
manifests  a  bimodal  distribution.  Mouse  DNA  illustrates  the  bimodal 
DNA  distribution  (36).  Mouse  DNA  manifests  a  major  component 
having  a  density  of  1.701  gcm~^  and  a  second  minor  component, 
comprising  about  8  per  cent  of  the  total  DNA,  having  a  density 
of  1.690  gcm~^  (Figs.  6,  7).  A  bimodal  DNA  distribution  is  also 
observed  with  guinea  pig  DNA.  In  this  case,  the  major  component 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  99 

is  the  lighter  one  (1.697  gcm~^)  and  the  minor  component  is  the 
heavier  one  (1.703  gcm"^^)   (36). 

Since  the  density  of  DNA  is  linearly  related  to  the  molar  per 
cent  (G+C),  it  is  possible  that  the  mouse  and  guinea  pig  minor 
components  consist  of  a  population  of  DNA  molecules  differing 
in  molar  per  cent  (G+C:).  However,  alternate  explanations  for 
the  minor  component  may  be  offered.  This  point  will  be  resolved 
when  the  guinea  pig  and  mouse  minor  DNA  components  are 
isolated  in  pure  forin. 

An  interesting  bimocial  distribution  has  been  found  by  Sueoka 
(72)  in  crab  testes  DNA  (Table  V).  The  major  DNA  component 
has  a  density  of  approximately  1.705  gcm"^  However,  a  very 
light  minor  component  also  occurs  which  is  indistinguishable  from 
a  double  stranded  polynucleotide  in  which  only  two  of  the  four 
nucleotide  bases  are  present.  The  bases  involved  are  adenine  and 
thymine  and  the  polymer  is  called  the  (deoxy-A-T)  polynucleotide. 
The  function  of  this  unusual  polynucleotide,  deoxy-A-T,  is 
unknown. 

Heterogeneity  of  Composition  of  DNA 

Since  the  DNx^  of  any  species  is  quite  heterogeneous,  it  is  of 
interest  to  compute  an  upper  bound  on  the  standard  deviation 
{aac)  of  the  distribution  of  the  guanine-cytosine  base  pairs  over 
the  population  of  DNA  molecules.  The  upper  bound  of  {ctgc)  is 
given  as: 

[4]  O-GC  max    =    10  O- density 

where  o- density  is  the  standard  deviation  of  the  DNA  distribution 
in  the  CsCl  density  gradient  (54).  It  should  be  pointed  out  that 
the  actual  value  of  (tqc  lies  considerably  below  the  calculated 
upper  bound  because  thermal  motion  of  the  DNA  molecules  con- 
tributes significantly  to  band  width. 

The  DNA's  of  nine  bacterial  species  form  bands  in  the  density 
gradient  with  o- density  in  no  case  greater  than  0.003  gcm~^  The 
corresponding  upper  bound  on  the  standard  deviation,  (tqc,  of 
the  molecular  content  of  guanine  plus  cytosine  is  therefore  in  no 
case  greater  than  0.03.  It  is  remarkable  that  the  standard  deviation 
of  guanine-cytosine  content  within  the  molecular  population  of 


100  Information  Storage  and  Neural  Control 

any  one  bacterial  species  covers  less  than  one  tenth  of  the  range 
over  which  the  mean  guanine-cytosine  content  varies  among  the 
various  species. 

Doty  and  co-workers  (1^)  have  sliown  that  the  total  variance 
of  the  DNxA.  bands  equals  the  sum  of  the  variance  due  to  molecular 
weight  {(tmw'^)  and  that  due  to  density  heterogeneity  (or density)- 

[5]  C'r    =    (J''mW   +   0-"  density 

Since  the  variance  due  to  molecular  weight  can  be  estimated, 
<^^deiisity  can  be  calculated.  The  latter  value  can  be  expressed  in 
units  of  molar  per  cent  (G+C);  that  is,  in  terms  of  ctqc. 

For  bacterial  DNA,  Doty,  et  al.  (16)  have  shown  that  age  is 
actually  equal  to  about  ±1.7  molar  per  cent  (G  +  C). 

The  values  for  animal  tissues  are  considerably  greater  (Table  VI). 
The  standard  deviation  in  units  of  density  (o-density)  ranges  from 
0.0037  to  0.0047  gcm"~^  The  latter  value  is  equivalent  to  a  standard 
deviation  corresponding  to  the  molecular  content  of  guanine  plus 
cytosine  of  about  3  molar  per  cent  (G  +  C). 

DNA  obtained  from  various  adult  tissues  of  mice  and  from  mouse 
tumors  do  not  differ  significantly  in  their  effective  buoyant  den- 
sities or  in  the  standard  deviations  of  the  density  gradient  bands. 
Although  all  differentiated  tissues  of  a  given  organism  are  pre- 
sumed to  have  identical  genomes,  there  is  evidence  that  normal 
tissues  and  tumors  differ  genetically.  The  fact  that  no  significant 
differences  between  the  DNA  of  normal  and  of  malignant  tissues 
have  been  found  does  not  necessarily  contradict  the  latter  concept. 
Instead,  it  reflects  the  fact  that  existing  techniques  are  insufficiently 
sensitive  to  detect  such  differences.  It  is  quite  possible  that  thou- 
sands of  point  mutations  exist  in  the  genomes  of  cancer  cells. 
These  could  not  be  detected  by  the  relatively  gross  physical 
methods  so  far  employed. 

Although  significant  differences  between  the  DNA  of  adult 
animal  tissues  and  the  DNA  of  tumors  have  not  been  found,  it 
has  been  possible  to  recognize  specific  differences  between  the 
DNA  of  various  species  of  higher  animals  (37).  As  shown  in  Table 
VI,  frog,  turtle,  and  alligator  DNA  are  slightly  heavier  than  other 
vertebrate  DNA's.  Chinese  hamster  and  frog  DNA  have  relatively 
low  standard  deviations  for  the  DNA  bands.  Also,  mouse  and 
guinea  pig  DNA  manifest  bimodal  distributions. 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  101 

Another  point  of  interest  is  the  fact  that  the  small  heterogeneity 
of  base  composition  among  the  DNA  molecules  of  an  organism 
seems  to  be  true  for  smaller  regions  within  molecules  (72).  This 
indicates  that  the  intramolecular  distribution  of  (G+C)  and 
(A+T)  pairs  is  fairly  unifoim,  although  in  short  regions  (for 
example,  tri-  or  tetranucleotides)  nonrandomness  has  been  demon- 
strated. 

The  Formation  of  Hybrid  DNA  Molecules  and  Their  Use  in 
Studies  of  DNA  Homologies 

DNA  molecules  having  the  same  average  molar  nucleotide  com- 
position do  not  necessarily  have  the  same  nucleotide  sequence  along 
the  DNA  chains.  Since  techniques  for  measuring  the  nucleotide 
sequence  are  not  currently  available,  possible  sequence  homologies 
between  DNA  molecules  from  different  sources  must  be  investi- 
gated by  indirect  methods.  There  have  been  two  general  approaches 
to  this  problem  so  far:  1)  an  analysis  of  the  distribution  of  oligo- 
nucleotides in  partial  hydrolysates  of  DNA;  and  2)  a  study  of 
DNA  hybrids. 

Burton  (6)  has  measured  the  distribution  of  short  chain  oligo- 
nucleotides in  acid  degradation  products  of  DNA.  Differences 
could  be  detected  in  the  distribution  of  dinucleotides  and  tri- 
nucleotides of  four  animal  and  four  bacterial  species. 

The  formation  of  hybrid  DNA  molecules  has  been  investigated 
by  Schildkraut  and  co-workers  (61).  These  studies  depend  upon 
the  fact  that  each  DNA  molecule  consists  of  two  complementary 
strands  which  can  be  separated  in  solution.  Strand  separation  can 
be  accomplished  by  heating  the  DNA  to  a  temperature  which 
will  "melt"  the  hydrogen  bonds  which  hold  together  the  double 
stranded  helix.  One  of  the  DNA  preparations  to  be  tested  is 
labeled  with  N^^  C''\  or  deuterium,  so  that  it  will  form  a  heavy 
band  when  it  is  centrifuged  in  a  density  gradient.  The  second 
DNA  preparation  is  of  normal  density.  The  heavy  and  the  light 
DNA  molecules  are  mixed  and  the  strands  are  separated  by 
heating.  Wlien  DNA  is  slowly  cooled,  the  complementary  strands 
attract  each  other  and  the  hydrogen  bonds  are  reformed  {renatura- 
tion).  Thus,  the  DNA  duplexes  are  reconstituted. 

Let  us  consider  the  situation  when  two  DNA  molecules  from 
the  same  species  are  renatured,  but  where  one  molecule  is  labeled 


102 


Information  Storage  and  Neural  Control 


with  heavy  isotopes  and  the  second  is  of  normal  density.  Upon 
renaturation,  hybrid  molecules  of  intermediate  density  will  be 
formed: 


[6] 


Normal  Density 
DNA 


II, ,        II 


Heavy   Density 
DNA  (n'^,    c'^) 


Upon    Renaturation  Gives 


Heated  Gives 


Single  Strands 


Hybrid   DNA  of 
Intermediate   Density 


The  normal  density  DNA,  "heavy"  density  DNA,  and  "inter- 
mediate" density  DNA  can  be  resolved  as  discrete  bands  by 
density  gradient  centrifugation.  Hybrid  formation,  then,  is  de- 
tected by  the  presence  of  a  new  DNA  band  of  intermediate  density. 
It  should  be  emphasized  that  hybrid  formation  can  take  place 
only  when  long  regions  of  the  nucleotide  sequences  of  DNA 
molecules  are  identical  or  very  nearly  so.  Molecules  having  the 
same  average  (G+C)  content  but  differing  in  the  sequence  of 
G,  C,  T,  and  A  along  the  polynucleotide  chain  will  not  form 
hybrids. 

The  possibility  that  renaturation  and  hybrid  formation  might 
take  place  between  the  DNA  of  bacterial  strains  with  close  taxo- 
nomic,  physiological,  and  genetic  relationships  was  investigated 
by  Schildkraut  et  al.  (61).  Hybrid  formation  was  readily  demon- 
strated between  the  DNA  of  E.  coli  and  of  six  other  E.  coli  strains. 

Interspecies  hybridization  of  DNA  was  also  demonstrated  in 
certain  instances  for  bacteria  having  the  same  nucleotide  content 
of  (G  +  C).  Thus,  the  DNA  from  B.  subtilis  and  B.  natto  formed 
hybrids,  and  in  addition,  the  DNA  from  E.  coli  K-12  formed 
hybrids  with  those  from  E.  coli  B.  and  Shigella  clysenterioe.  Sig- 
nificantly, these  are  instances  where  genetic  exchange  has  been 


Pyrimidine  Moieties  in  Aiiimals,  Plants,  and  Bacteria  103 

demonstrated  by  conjugation  or  transduction.  No  hybrid  formation 
was  detected  between  the  DNA  oi  E.  coli  K-12  and  that  oi  Salmonella 
typhimurium.  The  latter  bacteria  mate  but  transduction  from  one 
to  the  other  occurs  only  to  a  very  limited  extent,  if  at  all. 

Aside  from  the  taxonomic  importance  of  this  technique,  it  offers 
a  rational  approach  to  the  study  of  genetic  compatibility  where 
genetic  exchanges  have  not  been  demonstrated.  In  addition,  the 
technique  has  found  application  in  connection  with  the  problem 
of  information  transfer  between  DNA  and  RNA.  The  latter 
experiments  will  be  discussed  in  the  next  section  of  this  paper. 

THE  RIBONUCLEIC  ACIDS  (RNA) 

General  Characteristics 

The  ribonucleic  acids  (RNA),  which  mediate  the  transfer  of 
genetic  information  between  DNA  and  proteins  (Fig.  1),  differ 
chemically  from  DNA  in  several  ways  (35):  1)  The  sugar  com- 
ponent of  RNA  is  ribose,  instead  of  deoxyribose  (Fig.  2);  2)  Uracil 
(U),  instead  of  thymine,  is  the  6-keto-pyrimidine  base  in  RNA 
(Fig.  3);  3)  RNA  is  a  single  stranded,  flexible  polynucleotide  coil 
unlike  DNA  which  is  rather  stiff  and  double-stranded;  and  4)  Most 
RNA  molecules  are  much  shorter  in  length  than  DNA.  Also, 
RNA  is  less  stable  in  alkaline  solutions  than  is  DNA. 

Four  classes  of  RNA  are  known:  l)transfer-RNA,  2)  ribosomal- 
RNA,   3)   messenger  or  informational-RNA,   and  4)   virus-RNA. 

Transfer-RNA 

Transfer-RNA  (T-RNA  or  S-RNA)  consists  of  a  family  of 
molecules  which  function  in  the  activation  of  amino  acids  and  in 
the  transfer  of  the  activated  amino  acids  to  the  ribosomal  tem- 
plates so  that  they  can  be  linked  together  to  form  proteins.  Prob- 
ably, a  different  and  characteristic  transfer-RNA  molecule  is 
required  for  each  of  the  twenty  amino  acids.  Yeast  T-RNA 
specific  for  the  activation  of  the  amino  acid,  valine,  has  recently 
been  obtained  by  Stephenson  and  Zamecnik  (70)  in  highly 
purified  form  (65-80  per  cent).  Holley  et  al.,  (31)  have  partially 
purified  the  alanine,  valine,  and  tyrosine  T-RNA  of  yeast  and 
have  studied  the  oligonucleotide  content  of  ribonuclease  digests. 


104  Information  Storage  and  Neural  Control 

T-RNA  molecules  have  a  molecular  weight  of  about  25  to 
30,000  (80  to  100  nucleotide  chain  length).  The  sedimentation 
constant  of  T-RNA  is  about  4S.  Three  additional  characteristics 
are  of  interest:  1)  It  has  been  shown  that  guanine  mononucleotide 
terminates  one  end  of  the  T-RNA  chain,  2)  cytidylic  acid- 
cytidylic  acid-adenylic  acid  is  the  trinucleotide  which  terminates 
the  other  end  of  the  T-RNA  chain,  and  3)  each  T-RNA  chain 
contains  an  unusual  mononucleotide,  pseudouridylic  acid  (PsU). 
The  function  of  pseudouridylic  acid  is  unknown.  An  amino  acid 
can  be  attached  to  the  adenylic  acid  end  of  the  chain  as  shown 
schematically  in  Equations  7  through  9  (30): 

[7]    Amino  acid  +  ATP  — ^  Adenyl  ~  Amino  acid  +  Pyro- 
phosphate 

[8]    Adenyl  «  Amino  acid  +  G- — PsU— -C-C-A  -^ 

(T-RNA) 

Adenylic  acid  +  G PsU C-C-A  «  Amino  acid 

(T-RNA  with  activated  amino  acid) 

[9]    T-RNA  «  Amino  acid  +  Ribosomes  -^  T-RNA  + 
Ribosomes  (Amino  Acid) 

Transfer-RNA  comprises  approximately  10  per  cent  of  the  total 
cellular  RNA.  The  mononucleotide  content  of  the  T-RNA  from 
a  number  of  different  sources  has  been  determined  (Table  X). 
T-RNA  molecules  from  all  sources  have  a  high  content  of  guanine 
and  cytosine  (53-61  molar  per  cent  (G  +  C)). 

TABLE  X 

%  (G  +  C)  Content  of  Transfer — RNA 


Species 

(%G+C) 
59 

%Psoudoun 

<dylic 

Rejerence 

Rabbit  appendix] 

(63) 

Rat  liver 

58 

3.95 

(51) 

E.  coli' 

59.6 

1.22 

(51) 

E.  coli^ 

61.0 

2.1 

(17) 

Saccharomyces  cerevisiae 

53.1 

3.05 

(51) 

Yeast  alanine-T-RNA 

61.3 

3.7 

(31) 

Yeast  valine-T-RNA 

56.5 

4.7 

(31) 

Yeast  tyrosine-T-RNA 

56.5 

4.7 

(31) 

D.  pneumoniae 

52.6 

(82) 

Micrococcus  lysodeikticus 

56.8 

(82) 

Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria 


105 


Ribosomal-RNA  and  Total  Cellular  RNA 

Ribosomal  RNA  comprises  the  bulk  of  the  RNA  of  cells  (approxi- 
mately 80  per  cent).  The  ribosomes,  particles  consisting  of  about 
half  protein  and  half  RNA,  are  the  organelles  where  amino  acids 
are  assembled  into  protein  chains.  Ribosomal-RNA  seems  to  con- 
sist of  two  components  having  sedimentation  constants  of  about 


TABLE  XI 

Guanine  and  Cytosine  Content  of  Ribonucleic 
Acids  from  Various  Organisms 


Bacteria 

%G+C 
50.0 

Reference 

Proteus  vulgaris 

(2) 

Pasteurella  tularensis 

50.8 

(1) 

Escherichia  coh 

54.8 

(1) 

Azotobacter  vinelandii 

55.8 

(1) 

Sarcina  lutea 

56.9 

(1) 

Mycobacterium  tuberculosis  BCG 

59.3 

(1) 

Algae 

Ghaetocerus  decipiens 

52.1 

(1) 

Rhabdonema  adriaticum 

54.8 

(1) 

Scenedesmus  quadricauda 

56.1 

(1) 

Hydrodictyon  reticulatum 

56.4 

(1) 

Thalassiosira  nordenscheldii 

56.5 

(1) 

Hig/ier  Plants 

Equisetum  sp.  (spores) 

52.9 

(1) 

Zea  mays  (germs) 

54.8 

(1) 

Phaseolus  vulgaris  (sprout) 

55.2 

(1) 

Pea 

55 

(69) 

Tobacco  leaves 

56.9 

(45) 

Fungi 

Saccharomyces  cerevisiae 

50 

(2) 

Penicillium  stolonigerum 

50.6 

(2) 

Aspergillus  niger 

56.2 

(2) 

Protozoa 

Euglena  gracilis 

54.8 

(2) 

Tetrahymena  pyriformis 

39.1 

(2) 

Higher  Animals 

Rat  liver 

64.5,  60 

(33,  69) 

Rabbit  appendix 

62 

(63) 

Mouse  tissues 

64.0 

(34) 

Ghicken  tissues 

59.7 

(39) 

Duck  liver 

56.9 

(45) 

Calf  liver 

65.0 

(44) 

Sheep  liver 

71.2 

(44) 

Pig  liver 

64.1 

(44) 

Gat  brain 

58 

(44) 

Starfish  eggs 

58 

(44) 

Rana  catesbiana  eggs  (frog) 

68 

(20) 

Invertebrates 

Spider  oocytes  (cytoplasm) 

52.1 

(18) 

(Tegenaria  domestica) 

106  Information  Storage  and  Neural  Control 

25S  and  16S  (see  reference  35).  The  molecular  weights  of  these 
components  are  of  the  order  of  magnitude  of  500,000  and  1,000,000. 
It  was  formerly  believed  that  ribosomal-RNA  functioned  as  a 
template  for  protein  synthesis  but  recent  experiments  have  cast 
doubt  on  this  concept  (4,  27).  Possibly,  ribosomal-RNA  is  inert 
with  respect  to  genetic  coding.  Ribosomal-RNA  is  only  slowly 
synthesized  in  the  cell  and  the  mechanism  by  which  it  is  synthesized 
is  unknown.  There  is  no  obvious  relationship  between  the  com- 
position of  either  ribosomal-RNA  or  T-RNA  and  that  of  the 
DNA  of  a  given  organism. 

Since  ribosomal-RNA  comprises  the  bulk  of  the  RNA  of  a  cell, 
the  nucleotide  composition  of  ribosomal-RNA  is  similar  to  that 
of  the  total  cellular  RNA.  The  mononucleotide  compositions  of 
the  total  RNA  of  bacteria,  algae,  higher  plants,  fungi,  protozoa, 
and  higher  animals  are  shown  in  TABLE  XI  and  those  of  the 
ribosomal-RNA  of  several  organisms  in  Table  XII. 

TABLE  XII 

%  G  +  G  OF  RiBOSOMAL  RNA 


Species  and  Fraction 

46.7 
46.6 

Reference 

Saccharomjces  cerevisiae 
Large  granules 
Small  granules 

(51) 
(51) 

E.  coli  B 

Large  granules 
Small  granules 

54.1 
53.5 

(51) 
(51) 

Rat  liver 

Microsomes 

60.9 

(51) 

Rabbit  appendix 

Aqueous  phenol  extract 

62.5 

(63) 

Mouse  Tissues 

Microsomes — aqueous  phenol  extract 

66 

(34) 

The  molar  per  cent  (G+C)  of  the  total  RNA  and  the  ribosomal- 
RNA  of  organisms  generally  increases  as  the  molar  per  cent 
(G+C)  of  the  DNA  increases.  However,  the  correlation  between 
total  RNA  and  DNx^  composition  is  rather  weak  (1)  and  the 
change  in  RNA  composition  from  one  species  to  another  is  much 
less  than  that  of  the  DNA.  Belozersky  and  Spirin  (2)  have  com- 
piled the  RNA  nucleotide  composition  of  fifty-five  strains  of 
bacteria.  Some  representative  values  are  shown  in  Table  XI.  The 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  107 

molar  per  cent  (G+C)  varies  from  50  per  cent  (G+C)  for  Proteus 
vulgaris  to  59  per  cent  (G  +  C)  for  Alycobacterium  tuberculosis.  The 
values  for  most  bacterial  species  are  close  to  55  per  cent  (G+C). 
The  DNA  nucleotide  values  for  Proteus  vulgaris  and  Mycobacterium 
tuberculosis  are  39  per  cent  and  67  per  cent  (G  +  C),  respectively, 
(Table  II);  and,  as  mentioned  previously,  the  DNA  values  for 
all  species  range  from  32  per  cent  to  73  per  cent  (G  +  C). 

The  RNA  molar  nucleotide  composition  of  algae  varies  from 
52  to  56  per  cent  (G  +  C)  (Table  XI),  whereas  the  corresponding 
DNA  values  manifest  a  much  broader  variation,  that  is,  from  37 
to  64  mole  per  cent  (G  +  C)  (Table  III).  The  RNA  molar  nucleo- 
tide values  of  higher  plants  vary  over  a  4  per  cent  (G+C)  range 
(52.9  to  56.9)  and  the  values  from  fungi  vary  over  a  6  molar  per 
cent  (G  +  C)  range  (50  to  56).  On  the  other  hand,  the  DNA 
values  for  higher  plants  range  from  35  molar  per  cent  (G  +  C)  to 
48  molar  per  cent  (G+C)  (Table  III),  and  the  DNA  values  for 
fungi  vary  from  36  to  54  mole  per  cent  (G  +  C)  (Table  IV).  The 
molar  per  cent  (G  +  C)  is  much  greater  in  protozoan  RNA  than 
in  protozoan  DNA  (Tables  IV,  XI).  The  nucleotide  composition 
of  the  RNA  of  higher  animals  is  extremely  high,  over  60  per  cent 
(G+C);  while  the  DNA  of  the  animal  species  contains  about  40 
to  44  per  cent  (G+C)  (Tables  VI,  XI,  XII).  No  significant 
differences  have  so  far  been  detected  between  the  RNA  base 
composition  of  different  tissues  of  the  same  animals  or  between 
normal  tissues  and  tumors  (34,  39,  44). 

Virus  RNA 

A  third  kind  of  RNA  is  that  found  in  plant,  animal,  and  bacterial 
viruses  (Table  XIII),  RNA  viruses  are  capable  of  replicating 
within  cells  in  the  absence  of  new  DNA  synthesis  (55,  64)  but  the 
mechanisms  by  which  the  RNA  templates  of  the  viruses  are 
replicated  are  not  known.  Conceivably,  an  RNA  strand  might 
serve  as  a  template  for  the  replication  of  a  complementary  strand. 
However,  there  is  no  evidence  that  RNA  duplexes  exist  or  that 
strands  of  complementary  base  composition  occur  in  RNA  viral 
populations.  The  molecular  weight  of  the  RNA  of  viruses  is 
approximately  2  million  (9,  22).  It  is  rather  interesting  that  all 
but  one  of  the  thirteen  plant  and  animal  RNA  viruses  contain 


108  Information  Storage  and  Neural  Control 

TABLE  XIII 

Base  Composition  of  Ribonucleic  Acids  of  Viruses 


%G+C 

Reference 

Bacterial  virus  of  E.  colt  K-12 

52.7% 

(43) 

Plant  Viruses 

Turnip  yellow  mosaic 

55.3 

(59) 

Tobacco  ringspot 

47.9 

(59) 

Tomato  bushy  stunt  (BS3) 

48.7 

(59) 

Southern  bean  mosaic 

49.0 

(59) 

Tobacco  mosaic 

43.8 

(59) 

Aucuba  mosaic 

43.9 

(59) 

Rib  grass 

43.8 

(59) 

Potato  X 

44.6 

(59) 

Animal  Viruses 

Poliomyelitis  (Mahoney) 

46.5 

(57) 

Influenza  A  (PR-8) 

44.1 

(56) 

Influenza  B  (Lee) 

41.4 

(56) 

Encephalomyocarditis 

46.7 

(19) 

Rous  sarcoma 

50.9 

(12) 

41.4  to  50.9  mole  per  cent  (G  +  C)   (Table  XIII).  Many  DNA 
viruses  have  similar  values  (Tables  VII  through  IX). 

Messenger-RNA    (Also   called   Informational-RNA   and   Com- 
plementary-RNA  and  Abbreviated  C-RNA) 

The  lack  of  correlation  between  DNA  nucleotide  composition 
and  ribosomal  nucleotide  composition  provided  a  paradox  for 
some  time.  It  was  believed  that  genetic  information  was  coded  in 
DNA  but  that  the  actual  assembling  of  amino  acids  into  proteins 
occurred  on  the  ribosomes.  The  fact  that  proteins  are  not  syn- 
thesized directly  on  the  genes  demanded  the  existence  of  an  inter- 
mediate information  carrier.  This  intermediate  information  carrier 
was  generally  assumed  to  be  ribosomal-RNA.  However,  it  was 
difficult  to  reconcile  this  with  the  following  facts:  1)  ribosomal- 
RNA  is  relatively  stable;  2)  it  is  remarkably  homogeneous  in  size 
and  in  nucleotide  composition  although  this  homogeneity  reflects 
neither  the  range  of  size  of  polypeptide  chains  nor  the  variation 
in  nucleotide  composition  observed  in  the  DNA  from  different 
sources;  3)  the  capacity  of  bacteria  to  synthesize  a  given  protein 
does  not  survive  beyond  the  integrity  of  the  corresponding  gene; 
and  4)  regulation  of  protein  synthesis  in  bacteria  seems  to  operate 
at  the  level  of  the  synthesis  of  the  information  intermediate  by 
the  gene  rather  than  at  the  level  of  the  synthesis  of  the  protein 
(4,  27). 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  109 

This  paradox  appears  to  have  been  resolved  by  the  hypothesis 
that  ribosomal-RNA  is  not  the  intermediate  carrier  of  information 
from  gene  to  protein,  but  rather  that  ribosomes  are  non-specialized 
structures  wliicii  receive  genetic  information  from  the  genes  in 
the  form  of  an  unstable  intermediate  or  messenger  (4).  Although  this 
radical  revision  in  concepts  was  introduced  as  recently  as  May, 
1961,  an  impressive  array  of  supporting  evidence  has  now  been 
amassed.  This  evidence  will  be  discussed  briefly  in  this  section. 
A  schematic  representation  of  the  mechanism  of  information 
transfer  between  DNA  and  informational-RNA  (messenger-RNA) 
is  shown  in  Figure  8. 

DNA   DOUBLE    HELIX 


NEW  RNA  CHAIN 

Fig.  8.  Hypothetical  representation  of  the  transcription  of  genetic  information 
from  the  DNA  double  heUx  to  "Informational-RNA." 

It  is  assumed  that  DNA  can  act  as  a  template  for  the  syn- 
thesis of  a  new  messenger-RNA  chain.  The  mechanism  is  not 
unlike  tliat  by  which  the  DNA  chains  are  replicated.  Probably, 
the  DNA  double  helix  partially  unwinds.  Each  base  then  attracts 
a  complementary  free  ribonucleotide  already  available  for  poly- 
merization within  the  cell.  The  free  ribonucleotides,  whose  phos- 
phate groups  already  possess  the  free  energy  necessary  for  poly- 
esterification,  then  link  up  with  one  another,  after  being  held  in 
place  by  the  DNA  template  chains,  to  form  a  new  ribopoly- 
nucleotide  molecule.  Thus,  DNA  serves  as  a  template  for  the 
synthesis  of  a  complementary  RNA  strand.   The  newly  formed 


110  Information  Storage  and  Neural  Control 

complementary-RNA  (C-RNA)  strand  then  dissociates  from  the 
DNA,  and  moves  to  the  vicinity  of  nuclear  or  cytoplasmic  ribo- 
somes  where  it  serves  as  a  template  for  protein  synthesis.  Mean- 
while, the  DNx^  strands  probably  spontaneously  return  to  the 
double  stranded  form. 

Messenger-RNA  (C-RNA)  rapidly  becomes  radioactive  when 
cells  are  incubated  with  radioactive  uridine  or  orthophosphate. 
To  account  for  the  high  turnover  of  this  species  of  RNA,  it  has 
been  suggested  that  it  is  very  unstable  and  that  possibly  it  is 
degraded  after  it  has  fulfilled  its  function  in  protein  synthesis  (4). 
The  fact  that  the  bulk  of  the  cellular  RNA  differs  markedly  in 
composition  from  the  DNA  of  a  given  organism  and  also  from  the 
C-RNA  suggests  that  C-RNA  is  present  at  very  low  concentrations 
in  the  cell.  As  commonly  isolated,  C-RNA  has  a  sedimentation 
coefficient  of  only  9  to  12S,  but  this  may  be  due  to  the  fact  that  the 
9  to  12S  molecules  represent  degraded  messenger-RNA  chains. 

The  first  evidence  for  the  existence  of  messenger-RNA  came 
from  the  experiments  of  Volkin  and  Astrachan  (76).  Using  isotope 
labeling,  Volkin  and  Astrachan  were  able  to  show  that  in  bacterial 
cells  infected  with  a  bacteriophage,  such  as  T-2,  there  was  a  high 
turnover  in  a  minor  RNA  fraction.  This  RNA  fraction  had  an 
apparent  nucleotide  composition  which  corresponded  to  that  of 
the  DNA  of  the  phage  and  was  markedly  different  from  that 
of  the  host  DNA.  The  experiments  of  Volkin  and  Astrachan 
were  subsequently  confirmed  and  extended  by  Nomura  and 
co-workers  (49). 

In  1960,  Rich  demonstrated  that  it  was  possible  to  form  a 
specific  and  complementary  helical  complex  involving  a  synthetic 
DNA  strand  (polydeoxyribothymidylic  acid)  and  a  synthetic  RNA 
strand  (polyriboadenylic  acid)  (53).  Schildkraut  et  al.  (62)  em- 
ployed density  gradient  centrifugation  experiments  to  show  that 
a  hybrid  complex  between  polydeoxyguanylic  acid  and  poly- 
ribocytidylic  acid  was  also  possible. 

In  the  same  year,  it  was  shown  by  several  laboratories  (5,  79, 
80,  23,  71)  that  an  RNA  polymerase  enzyme  was  present  in 
bacteria  and  in  animal  cell  nuclei.  The  purified  enzyme  could 
catalyze  a  net  synthesis  of  new  RNA  froin  uridine  triphosphate, 
adenosine    triphosphate,    guanosine    triphosphate,    and    cytidine 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  111 

triphosphate,  but  DNA  exercised  a  directing  role  in  the  enzymatic 
synthesis  of  RNA.  DNA  from  several  sources,  having  widely 
different  base  compositions,  could  serve  as  primers  for  the  synthesis 
of  RNA  by  KNA  polymerase,  but  the  newly  synthesized  RNA 
had  a  composition  which  reflected  the  composition  of  the  DNA 
primer  which  was  einployed.  With  the  purified  enzyme,  RNA 
synthesis  only  took  place  if  DNA  was  present. 

Single  stranded  DNA  from  phage  0X174  could  serve  as  primer 
in  which  case  the  RNA  had  a  composition  complementary  to  the 
DNA  of  0X174  (7).  If  double  stranded  0X174  DNA  was  enzy- 
matically  prepared  with  DNA  polymerase  and  the  double  stranded 
0X174  DNA  was  then  used  as  a  primer  for  the  synthesis  of  RNA 
by  RNA  polymerase,  the  RNA  product  had  a  composition  similar 
to  double  stranded  0X174  DNA.  In  addition,  RNx\  synthesis  was 
primed  by  heat  denatured  DNA  (23,  80).  These  results  suggested 
that  each  of  the  complementary  DNA  strands  could  serve  as  tem- 
plates for  the  synthesis  of  Messenger-RNA.  However,  it  is  not  yet 
definite  that  this  happens  in  vivo.  Possibly,  only  one  of  the  DNA 
strands  serves  to  specify  the  sequence  of  messenger-RNA  and  the 
second  strand  constitutes  "nonsense"  information. 

Not  only  is  the  composition  of  the  newly  synthesized  RNA 
dependent  on  the  composition  of  the  DNA  primer,  but  the  se- 
quence of  the  ribonucleotides  in  the  new  RNA  is  determined  by 
the  sequence  of  deoxyribonucleotides  in  the  DNA.  This  has  been 
demonstrated  in  two  ways.  First,  nearest  neighbor  sequence  studies 
have  been  carried  out  by  Furth,  Hurwitz,  and  Goldman  (24)  and 
by  Weiss  and  Nakamoto  (81).  Second,  it  has  been  demonstrated 
that  enzymatically  synthesized  RNA  formed  with  a  T-2  phage 
DNA  primer  can  be  heated  with  the  T-2  DNA  and  slowly  cooled 
so  as  to  permit  hybrid  formation  by  renaturation  (25).  The  hybrid 
has  been  demonstrated  by  density  gradient  centrifugation  experi- 
ments. Hybrid  formation  occurs  between  C-RNA  specific  to  T-2 
phage  and  T-2  phage  DNA,  but  not  between  G-RNA  of  T-2 
phage  and  E.  coli  DNA  or  sea  urchin  DNA. 

The  existence  of  T-2  specific  RNA,  which  was  initially  inferred 
from  the  isotope  experiments  of  Volkin  and  Astrachan,  was  estab- 
lished by  Nomura,  Hall,  and  Spiegelman  (49).  Newly  synthesized 
RNA  was  separated  from  the  bulk  of  cellular  RNA  using  both 


112  Information  Storage  and  Neural  Control 

zone  electrophoresis  in  starch  columns  and  centrifugation  in 
sucrose  gradients.  T-2  specific  RNA  had  a  higher  electrophoretic 
mobility  and  a  greater  heterogeneity  in  size  than  the  principal 
normal  RNA  components.  The  T-2  specific  RNA  was  found  to 
be  bound  to  the  ribosomes,  but  with  a  linkage  very  sensitive  to 
disruption  by  low  magnesium  levels. 

Renaturation  and  hybrid  formation  experiments  were  per- 
formed to  establish  that  sequence  complementarity  existed  be- 
tween "T-2  phage  specific  RNA''  and  T-2  phage  DNA  (28). 
RNA-DNA  complex  formation  was  demonstrated  in  mixtures  of 
heat  denatured  T-2  phage  DNA  and  purified  T-2  RNA  subjected 
to  the  slow  cooling  process.  The  success  of  the  hybridization 
experiments  suggested  immediately  that  the  original  observation 
by  Volkin  and  Astrachan  (76)  of  a  similarity  in  base  composition 
between  T-2  RNA  and  DNA  was  indeed  a  reflection  of  a  more 
profound  homology.  Hybrid  formation  was  specific.  Heterologous 
DNA  from  Psendomonas  aeruginosa,  E.  coli,  or  phage  T-5  did  not 
yield  DNA-RNA  hybrids  with  T-2  RNA.  This  led  to  the  con- 
clusion that  the  nucleotide  sequences  of  T-2  DNA  and  RNA 
were  complementary. 

In  further  experiments  Spiegelman,  Hall,  and  Storck  (68) 
demonstrated  the  natural  occurrence  of  DNx\-RNA  hybrids  in 
phage  infected  E.  coli  cells.  Finally,  Hayashi  and  Spiegelman  (29) 
and  Gros  et  at.  (27)  have  demonstrated  the  presence  of  natural 
DNA-RNA  hybrids  in  uninfected  bacterial  cells. 

THE  GENETIC  CODE 

The  recent  experiments  of  Nirenberg  and  Matthaei  (48)  and 
of  Ochoa  and  collaborators  (42,  67)  represent  a  major  break- 
through and  give  promise  of  providing  the  key  to  the  entire 
genetic  code  within  one  or  two  years.  Nirenberg  and  Matthaei  (48) 
were  able  to  develop  a  cell  free  ribosomal  system  from  E.  coli  in 
which  the  amount  of  incorporation  of  amino  acids  into  proteins 
was  dependent  upon  the  addition  of  heat  stable  RNA  preparations. 
Transfer-RNA  could  not  replace  the  active  RNA  fraction  which 
presumably  contained  some  messenger-RNA. 

Of  particular  interest  was  the  most  important  observation  that 
the  addition  of  a  synthetic  polyribonucleotide,  polyuridylic  acid. 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  113 

specifically  stimulated  the  incorporation  of  the  amino  acid  L-phen- 
ylalanine  into  a  protein  resembUng  poly-L-phenylalanine.  In  this 
system,  advantage  was  taken  of  the  fact  that  poly-L-phenylalanine 
is  poorly  soluble.  Hence,  it  precipitated  out  of  solution  and  was 
readily  isolated.  The  obvious  implication  of  the  experiments  was 
tliat  polyuridylic  acid  was  functioning  as  a  synthetic  template — 
messenger-RNA:  Hence,  uridylyl-uridylyl-uridylyl  (UUU)  was 
probably  the  nucleotide  triplet  which  coded  for  the  amino  acid 
L-phenylalanine. 

Other  synthetic  polyribonucleotides  were  quickly  tested  in  this 
system  by  Speyer,  Lengyel,  Basilio,  and  Ochoa  (67).  On  the  basis 
of  the  latter  experiments,  the  nucleotide  code  letters  for  the  twenty 
amino  acids  commonly  found  in  proteins  have  been  identified. 
The  letters  for  an  assumed  triplet  code  are  presented  in  Table 
XIV.  It  should  be  pointed  out  that  the  sequence  of  bases  in  the 
triplets  is  known  only  for  UUU  (code  letter  of  L-phenylalanine). 
The  proposed  code  letters  are  in  excellent  agreement  with  amino 
acid  replacement  data  on  nitrous  acid  mutants  of  tobacco  mosaic 
virus    (TMV).    In    experiments   with    nitrous    acid    mutuants    of 

TABLE  XIV 

The  Proposed  Genetic  Code  (42) 


L-Amino  Acid 

RNA  Triplet 

1. 

phenylalanine 

UUU 

2. 

valine 

2U-1G 

3. 
4. 

cysteine 
isoleucine 

2U-1G 
2U-  lA 

5. 
6. 

tyrosine 
lysine 

2U-1A 
1U-2A 

7. 

serine 

2U-  IC 

8. 

leucine 

2U-  IC,  (2U-  1A;2U  -  IG) 

9. 

proline 

1U-2C 

10. 

threonine 

1U-2C,  (lU-  lA-lC) 

11. 

glycine 

1U-2G 

12. 
13. 

tryptophane 
histidine 

1U-2G 
lU-lA-lC 

14. 

arginine 

lU-lC-lG 

15. 

glutamine 

lU-lC-lG 

16. 

alanine 

lU-lC-lG 

17. 

methionine 

lU-lA-lG 

18. 

glutamic 

lU-lA-lG 

19. 

aspartic 

lU-lA-lG 

20. 

asparagine 

1U-2A,  (lU-  lA-  IC) 

Note:  During  the  transcription  of  genetic  infor- 
mation from  DNA  to  RNA,  thymine  (T)  is 
replaced  by  uracil  (U). 


114  Information  Storage  and  Neural  Control 

TMV,  it  has  been  shown  that  in  one  mutant  proHne  is  replaced 
by  leucine,  and  threonine  is  replaced  by  serine.  Since  the  effect 
of  nitrous  acid  is  to  convert  cystosine  to  uracil,  this  would  imply 
that  there  are  changes  of  1U-2C  to  2U-IC  for  the  proline  to 
leucine  replacement  in  the  TMV  protein  and  changes  of  1U-2C 
to  2U-1C  for  the  threonine  to  serine  replacement.  The  replace- 
ment of  serine  by  phenylalanine  and  of  glutamine  by  valine  in 
another  mutant  is  also  consistent  with  the  proposed  genetic  code 
(Table  XIV). 

The  code  need  not  contain  the  triplet  CCA.  The  CCA  is  the 
terminal  sequence  of  transfer-RNA  to  which  the  activated  amino 
acid  is  attached  during  protein  synthesis. 

The  recent  progress  in  the  field  of  genetic  coding  is  impressive. 
However,  many  important  questions  must  still  be  solved.  Three  of 
these  are:  1)  How  a  triplet  of  nucleotides  can  sterochemically 
account  for  the  coding  of  an  amino  acid;  2)  how  ribosomal-RNA, 
transfer-RNA,  and  messenger-RNA  are  held  together  on  the 
ribosomal  particles;  and,  3)  the  mechanisms  by  which  ribosomal- 
RNA,  transfer-RNA,  and  viral  RNA  are  synthesized. 

There  is  every  reason  to  be  optimistic  that  these  and  other 
important  questions  will  soon  be  resolved. 

REFERENCES 

1.  Belozersky,  A.   N.:  Nucleic  Acids  of  Microorganisms.   Nucleoproteins, 

p.  199-229.  Proceedings  of  the  Eleventh  Institute  International  de 
Chimie  Solvay,  Brussels,  June  1-6,  1959,  New  York,  Interscience, 
Inc.    1960 

2.  Belozersky,  A.  N.  and  Spirin,  A.  S.:  Chemistry  of  the  Nucleic  Acids 

of  Microorganisms.  The  Nucleic  Acids  III,  ed.  by  E-  Chargaff  and 
J.  N.  Davidson,  New  York.  Academic  Press,  Inc.,  1960,  p.  147. 

3.  Bergold,  G.  H.:  Biochemistry  of  Insect  Viruses.  The  Viruses  I,  ed.  by 

F.  M.  Burnet  and  W.  M.  Stanley.  New  York,  Academic  press. 
Inc.,  1959,  p.  505. 

4.  Brenner,  S.,  Jacob,  F.  and  Meselson,  M.:  An  unstable  intermediate 

carrying  information  from  genes  to  ribosomes  for  protein  syn- 
thesis. Nature,  79(9.- 576-581,  1961. 

5.  Burma,  D.  P.,  Kroger,  H.,  Ochoa,  S.,  Warner,  R.  C.  and  Weill, 

J.  D.:   Further  studies  on  deoxyribonucleic  acid  dependent  en- 


Pjrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  115 

zymatic  synthesis  of  ribonucleic  acids.  Proc.  Nat.  Acad.  Sci.  USA, 
47:1A9-152,  1961. 

6.  Burton,  K.:  Frequencies  of  nucleotide  sequences  in  deoxyribonucleic 

acids.  Biochem.  J.,  77.-547-552,  1960. 

7.  Chamberlin,  M.  and  Berg,  P.:  Deoxyribonucleic  acid-directed  syn- 

thesis of  ribonucleic  acid  by  an  enzyme  from  Escherichia  coh. 
Proc.  Nat.  Acad.  Sci.   USA,  45.- 8 1-94,  1962. 

8.  Chargaff,  E.:  Isolation  and  Composition  of  the  Deoxypentose  Nucleic 

Acids  and  of  the  Corresponding  Nucleoproteins.  The  Nucleic  Acids  /, 
ed.  by  E.  Chargaff  and  J.  N.  Davidson,  New  York,  Academic 
Press,  Inc.,  1955,  p.  307. 

9.  Cheng,  P.  Y. :  On  the  similarity  of  chain    length  of  nucleic  acids. 

Proc.  Nat.  Acad.  Sci.  USA,  45.-1 557-1 560,  1959. 

10.  Cohen,  S.  S.:  Comparative  biochemistry  and  virology.  Advance  Virus 

Res.,  J.-1-48,  1955. 

11.  Cordes,   S.,   Epstein,   H.  T.,   and  Marmur,  J.:   Some  properties  of 

the  deoxyribonucleic  acid  of  phage  alpha.  Nature,  797:1097-1098, 
1961. 

12.  Crawford,   L.   \'.   and   Crawford,   E.   M.:   The  properties  of  Rous 

sarcoma  virus  purified  by  density  gradient  centrifugation.  Virology, 
7J.-227-232,   1961. 

13.  Crick,  F.  H.  C,  Barnett,  L.,  Brenner,  S.,  and  Watts-Tobin,  R.  J.: 

General  nature  of  the  genetic  code  for  proteins.  Nature,  792.- 1227- 
1232,  1961. 

14.  Davern,    C.    I.:   Bias   in   i^ase-pair  orientation   in   deoxyriljonucleic 

acids.  Nature,  188:2Qd-209,  1960. 

15.  Davison,   P.   F.,   Freifelder,  D.,   Hede,   R.,  and  Levinthal,  C:  The 

structural  unity  of  the  DNA  of  T2  bacteriophage.  Proc.  Nat.  Acad. 
Sci.  USA,  47.- 11 23-1 129,  1961. 

16.  Doty,  P.,  Marmur,  J.,  and  Sueoka,  N.:  The  Heterogeneity  in  Prop- 

erties and  Functions  of  Deoxyribonucleic  Acids.  Brookhaven  Sym- 
posia in  Biology  No.  12,  Structure  and  Function  of  Genetic  Elements, 
1959,  p.   1. 

17.  Dunn,  D.  B.,  Smith,  J.  D.,  and  Spahr,  P.  F.:  Nucleotide  composition 

of  soluble  ribonucleic  acid  from  Escherichia  coli.  J.  Molec.  Biol., 
2.-113-117,   1960. 

18.  Edstrom,   Jan-Erick:   Composition  of  ribonucleic  acid  from  various 

parts  of  spider  oocytes.  J.  Biophys.  Biochem.  Cytol.,  5'.-47-51,  1960. 

19.  Faulkner,  P.,  Martin,  E.  M.,  Sved,  S.,  Valentine,  R.  C.  and  Work, 

T.  S.:  Studies  on  protein  and  nucleic  acid  metabolism  in  virus- 
infected  cells.  2.  The  isolation,  crystallization,  and  chemical  charac- 


116  Information  Storage  and  Neural  Control 

terization  of  mouse  encephalomyocarditis  virus.  Biochem.  J.,  80: 
597-605,  1961. 

20.  Finamore,  F.  J.  and  Volkin,  E.:  Some  chemical  characteristics  of 

amphibian  egg  ribonucleic  acids.  J.  Biol.  Chem.,  2J6.-443-447,  1961. 

21.  Freese,    E.:    On   the    Molecular    Explanation   of  Spontaneous   and 

Induced  Mutations.  Brookhaven  Symposia  in  Biology  No.  12, 
Structure  and  Function  of  Genetic  Elements,  1959,  p.  63. 

22.  Frisch-Niggemeyer,    W.:    Absolute   amount   of  ribonucleic    acid    in 

viruses.  Nature,    778:307-308,   1956. 

23.  Furth,  J.  J.,  Hurwitz,  J.,  and  Goldman,  M.:  The  directing  role  of 

DNA  in  RNA  synthesis.  Biochem.  Biophys.  Res.  Commun.,  4:362-361 , 
1961. 

24.  Furth,  J.  J.,  Hurwitz,  J.  and  Goldman,  M.:  The  directing  role  of 

DNA  in  RNA  synthesis.  Specificity  of  the  deoxyadenylate-deoxyth- 
ymidylate  copolymer  as  a  primer.  Biochem.  Biophys.  Res.  Commun., 
4:431-435,  1961. 

25.  Geiduschek,  E.  P.,  Nakamoto,  T.,  and  Weiss,  S.  B.:  The  enxymatic 

synthesis  of  RNA:  Complementary  interaction  with  DNA.  Proc. 
Nat.  Acad.  Sci.  USA,  47;  1405-1 41 5,  1961. 

26.  Gillies,  N.  E.  and  Alper,  T. :  The  nucleic  acid  content  of  Escherichia 

coU  Strains  B  and  B/r.  Biochim.  Biophys.  Acta,  43:\82-\87,   1960. 

27.  Gros,  F.,  Hiatt,  H.,  Gilbert,  W.,  Kurland,  C.  G.,  Risebrough,  R.  W. 

and  Watson,  J.  D.:  Unstable  ribonucleic  acid  revealed  by  pulse 
labelhng  of  Escherichia  coli.  Nature,  790.-581-585,  1961. 

28.  Hall,  B.  D.  and  Spiegelman,  S.:  Sequence  complementarity  of  T2DNA 

and  T2  specific  RNA.  Proc.  .Nat.  Acad.  Sci.  USA,  47.-137-146,  1961. 

29.  Hayashi,  M.  and  Spiegelman,  S.:  The  selective  synthesis  of  infor- 

mational RNA  in  bacteria.  Proc.  Nat.  Acad.  Sci.  USA,  47:  \  564- 
1580,  1961. 

30.  Hoagland,  M.  B.:  The  Relationship  of  Nucleic  Acid  and  Protein 

Synthesis  as  Revealed  by  Studies  in  Cell-free  Systems.  The  Nucleic 
Acids  III,  ed.  by  E.  ChargafT  and  J.  N.  Davidson,  New  York, 
Academic  Press,  Inc.,  1960,  p.  349. 

31.  Holley,  R.  W.,  Apgar,  J.,  Merrill,  S.  H.  and  Zubkoff,  P.  L.:  Nucleo- 

tide and  oligonucleotide  compositions  of  the  alanine-,  valine-,  and 
tyrosine-acceptor  "Soluble"  ribonucleic  acids  of  yeast.  J.Am.  Chem. 
Soc,  5J.-4861-4862,  1961. 

32.  Iwamura,  T.  and  Myers,  J.:  Changes  in  the  content  and  distribution 

of  the  nucleic  acid  bases  in  chlorella  during  the  life  cycle.  Arch. 
Biochem.,  84:261-211,  1959. 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  117 

33.  Kirby,  K.  S.:  A  new  method  for  the  isolation  of  ribonucleic  acids 

from  mammalian  tissues.  Biochem.  J.,  64.-405-408,  1956. 

34.  Kit,  S.:  Base  composition  of  the  ribonucleic  acids  of  normal  tissues, 

tumors,  and  tissue  subfractions.  Arch.  Biochem.,  55';  1-9,  1960. 

35.  Kit,  S.:  Studies  on  the  structure,  composition,  and  metabolism  of 

tumor  RNA.  Cell  Physiology  of  Neoplasia,  Austin,  University  of 
Texas  Press,  1960,  p.  337. 

36.  Kit,   S.:   Equilibrium  sedimentation  in  density  gradients  of  DNA 

preparations  from  animal  tissues.  J.  Molec.  Biol.,  J;711-716,  1961. 

37.  Kit,  S.:  Species  differences  in  animal  deoxyribonucleic  acids  (DNA) 

as  revealed  by  equilibrium  sedimentation  in  density  gradients. 
Nature,  79J.-274-275,  1962. 

38.  Kit,  S.  and  Dubbs,  D.  R.:  (unpublished). 

39.  Kleinschmidt,    W.   J.:    The   composition    of  ribonucleic    acid    and 

deoxyribonucleic  acid  of  normal  and  neoplastic  tissue.  Cancer  Res. 
79.-966-969,  1959. 

40.  Kok,  I.  P.  and  Opalskii,  A.  F.:  Comparison  with  respect  to  nucleotide 

composition  of  deoxyribonucleic  acids  isolated  from  insects  and 
mammals  by  various  methods.  Biochemistry  {USSR),  25.-839-844, 
1961. 

41.  Lawley,  P.  D.:  The  action  of  Alkylating  Agents  on  Nucleic  Acids 

and  Their  Relation  to  Other  Mutagens.  The  Molecular  Basis  of 
Neoplasia.  Austin,  The  University  of  Texas  Press,  1962,  p.  123. 

42.  Lengyel,  P.,  Speyer,  J.  F.  and  Ochoa,  S.:  Synthetic  polynucleotides 

and  the  amino  acid  code.  Proc.  Nat.  Acad.  Sci.  USA,  ^7.- 1936-1 942, 
1961. 

43.  Loeb,  T.   and   Zinder,   N.   D.:   A   bacteriophage  containing  RNA. 

Proc.  Nat.  Acad.  Sci.  USA,  47:282-289,  1961. 

44.  Magasanik,  B.:  Isolation  and  Composition  of  the  Pentose  Nucleic 

Acids  and  of  the  Corresponding  Nucleoproteins.  The  Nucleic  Acids  I, 
ed.  by  E.  Chargaff  and  J.  N.  Davidson,  New  York,  Academic 
Press,  Inc.,  1955,  p.  373. 

45.  Makhinko,  V.  I.  and  Block,  L.  N.  (USSR):  personal  communication. 

46.  Meselson,   M.,   Stahl,   F.  W.   and  Vinograd,  J.:   Equilibrium  sedi- 

mentation of  macromolecules  in  density  gradients.  Proc.  Nat.  Acad. 
Sci.  USA,  43:581-588,  1957. 

47.  Minagawa,  T.,  Wagner,  B.  and  Strauss,  B.:  The  nucleic  acid  content 

of  Neurospora  Crassa.  Arch.  Biochem.,  50.-442-445,  1959. 

48.  Nirenberg,  M.  W.  and  Matthaei,  J.  H.:  The  dependence  of  cell-free 

protein  synthesis  in  E.  coli  upon  naturally  occurring  or  synthetic 
polyribonucleotides.  Proc.  Nat.  Acad.  Sci.  USA,  47.- 1588-1 602,  1961. 


118  Information  Storage  and  Neural  Control 

49.  Nomura,  M.,  Hall,  B.  D.,  and  Spiegelman,  S.:  Characterization  ol 

RNA  synthesized  in  Escherichia  coli  after  bacteriophage  T2  infec- 
tion. J.  Molec.  Biol,  2.-306-326,  1960. 

50.  Ogur,  M.,  Minckler,  S.,  Lindegren,  G.  and  Lindegren,  C.  C:  The 

nucleic  acids  in  a  polypoid  series  of  Saccharomyces.  Arch.  Biochem., 
40:175-184,  1952. 

51.  Osawa,   S.:  The  nucleotide  composition  of  ribonucleic  acids  from 

subcellular  components  of  yeast,  Escherichia  coli  and  rat  liver 
with  special  reference  to  the  occurrence  of  pseudouridylic  acid 
in  soluble  ribonucleic  acid.  Biochim.  Biophys.  Acta,  42:2AA-25A,  1960. 

52.  Randall,  C.  G.  and  Gafford,  L.  G.:  The  Nucleic  Acid  oj  Fowl  Pox  Virus. 

Abstracts,  Bacteriological  Proceedings,  1961,  p.  148. 

53.  Rich,  A.:  A  hybrid  helix  containing  both  deoxyribose  and  ribose 

polynucleotides  and  its  relation  to  the  transfer  of  information  be- 
tween the  nucleic  acids.  Proc.  Nat.  Acad.  Sci.  USA,  46:  1044-1053 
1960. 

54.  Rolfe,  R.  and  Meselson,  M.:  The  relative  homogeneity  of  micro- 

bial DNA.  Proc.  Nat.  Acad.  Sci.  USA,  45.T 039-1 043,  1959. 

55.  Salzman,  N.  P.:  The  rate  of  formation  of  vaccinia  deoxyribonucleic 

acid  and  vaccinia  virus.   Virology,  70.- 150-1 52,  1960. 

56.  Schafer,  W.:  The  Comparative  Chemistry  of  Infective  Virus  Particles 

and  of  Other  Virus-specific  Products:  Animal  Viruses.  The  Viruses  I, 
ed.  by  F.  M.  Burnet  and  W.  M.  Stanley,  New  York,  Academic 
Press,  Inc.,  1959,  p.  475. 

57.  SchafFer,   F.   L.,   Moore,   H.   F.   and   Schwerdt,   C.    E.:    Base  com- 

position of  the  ribonucleic  acids  of  the  three  types  of  polio  virus. 
Virology,  70.- 530-537,  1960. 

58.  Schuster,   H.:  The  reaction  of  nitrous  acid  with  deoxyribonucleic 

acid.  Biochem.  Biophys.  Res.  Commun.,  2."320-323,  1960. 

59.  Schuster,  H.:  The  Ribonucleic  Acids  of  Viruses.   The  Nucleic  Acids  III 

ed.  by  E.  Chargaff  and  J.  N.  Davidson,  New  York,  Academic 
Press  Inc.,  1960,  p.  245. 

60.  Schildkraut,  C.  L.,  Marmur,  J.  and  Doty,  P.:  Determination  of  the 

base  composition  of  deoxyribonucleic  acid  from  its  apparent 
density  in  CsCI.  J.  Molec.  Biol,   (in  press). 

61.  Schildkraut,   C.   L.,   Marmur,  J.   and  Doty,   P.:   The  formation  of 

hybrid  DNA  molecules  and  their  use  in  studies  of  DNA  homologies. 
J.  Molec.  Biol,  (in  press). 

62.  Schildkraut,  C.  L.,  Marmur,  J.,  Fresco,  J.  R.  and  Doty,  P.:  Forma- 

tion and  properties  of  polyribonucleotide-polydeoxyribonucleotide 
helical  complexes.  J.  Bwl.  Chem.,  236:PC  2-4  (Feb.)   1961. 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  119 

63.  Sibatani,  A.,  Yamana,  K.,  Kimura,  K.,  and  Takahashi,  T.:  Frac- 

tionation of  two  metal^olically  distinct  classes  of  ribonucleic  acids 
in  animal  cells  and  its  l:)earing  on  cancer.  Nature,  786:215-217,  1960. 

64.  Simon,  E.  H.:  Evidence  for  the  nonparticipation  of  DNA  in  viral 

RNA  synthesis.   Virology,   7J.- 105-1 18,  1961. 

65.  Sinsheimer,   R.   L.:  A  single  stranded  deoxyribonucleic  acid  from 

bacteriophage   0X174.  J.  Molec.  Biol.,  7.-43-53,  1959. 

66.  Sinsheimer,  R.  L.:  The  Nucleic  Acids  of  the  Bacterial  Viruses.   The 

Nucleic  Acids  III,  ed.  by  E.  ChargafF  and  J.  N.  Davidson,  New 
York,  Academic  Press  Inc.,  1960,  p.  187. 

67.  Speyer,  J.   F.,   Lengyel,   P.,   Basilio,   C.   and  Ochoa,   S.:   Synthetic 

polynucleotides  and  the  amino  acid  code.  II.  Proc.  Nat  Acad.  Sci. 
USA,  48:63-68,  1962. 

68.  Spiegelman,    S.,   Hall,   B.    D.    and   Storck,   R.:  The   occurrence   of 

natural  DNA-RNA  complexes  in  E.  coli  infected  with  T2.  Proc. 
Nat.  Acad.  Sci.  USA,  47;  11 35-1 141,  1961. 

69.  Spirin,  A.  S.:  The  "temperature  effect"  and  macromolecular  struc- 

ture  of  high-polymer   ribonucleic   acids   of  various   origin.    Bio- 
chemistry (USSR),  2(5.-454-463,  1961. 

70.  Stephenson,    M.    L.    and   Zamecnik,   P.    C:    Purification   of  valine 

transfer  ribonucleic  acid  by  combined  chromatographic  and  chem- 
ical procedures.  Proc.  Nat.   Acad.  Sci.    USA,  47.- 1627-1 635,    1961. 

71.  Stevens,  A.:  Net  formation  of  polyribonucleotides  with  base  com- 

positions analogous  to  deoxyribonucleic  acid.  J.  Biol.  Chem.,  236 
PC43-45  (July)   1961. 

72.  Sueoka,   N.:    Variation   and   heterogeneity  of  base  composition   of 

deoxyribonucleic  acids:  A  compilation  of  old  and  new  data.  J. 
Molec.  Biol.,  J.- 3 1-40,  1961. 

73.  Tessman,  I.:  Mutagenesis  in   Phage  0X174  and  T4  and  properties 

of  the  genetic  material.   Virology,  9.-375-385,  1959. 

74.  Thomas,  R.  S.:  The  chemical  composition  and  particle  weight  of 

tipula  iridescent  virus.   Virology,  74.-240-252,  1961. 

75.  Vendrely,  R.:  The  Deoxyribonucleic  Acid  Content  of  the  Nucleus. 

The  Nucleic  Acids  II,  ed.  by  E.  Chargaff  and  J.  N.  Davidson,  New 
York,  Academic  Press,  Inc.,  1955,  p.  155. 

76.  Volkin,  E.  and  Astrachan,  L.:   RNA  Metabolism  in  T-2  Infected 

Escherichia  coli.    The  Chemical  Basis  of  Heredity,   Baltimore,   The 
Johns  Hopkins  Press,  1957,  p.  686. 

77.  Watson,  J.  D.  and  Crick,    F.  H.  C:   The  Structure   of  DNA.    Cold 

Spring  Harbor  Symposium  on  Qjiantitive  Biology,  /5'."123-131,  1953. 


120  Information  Storage  and  Neural  Control 

78.  Watson,  J.  D.  and  Littlefield,  J.  W.:  Some  properties  of  DNA  from 

Shope  papilloma  virus.  J.  Molec.  Biol.,  2.- 161 -165,   1960. 

79.  Weiss,  S.  B.   and   Nakamoto,  T.:  Net  synthesis  of  ribonucleic  acid 

with  a  microbial  enzyme  requiring  deoxyribonucleic  acid  and  four 
ribonucleic  triphosphates.  J.  Biol.  Chem.,  236:PCA9-20,   1961. 

80.  Weiss,  S.  B.  and  Nakamoto,  T.:  On  the  participation  of  DNA  in 

RNA  synthesis.  Proc.  Mat.  Acad.  Sci.   USA,  47.-694-697,  1961. 

81.  Weiss,  S.  B.  and  Nakamoto,  T.:  The  enzymatic  synthesis  of  RNA: 

Nearest  neighbor  base  frequencies.  Proc.  Nat.  Acad.  Sci.  USA,  47: 
1400-1405,  1961. 

82.  Woese,  C.  R.:  Composition  of  various  ribonucleic  acid  fractions  from 

microorganisms   of  diflferent  deoxyribonucleic   acid   composition. 
Nature,   759;920-921,  1961. 

DISCUSSION  OF  CHAPTER  V 

E.  Roy  John  (Rochester,  New  York):  The  topic  which  you 
have  reviewed  is  of  particular  interest  to  neurophysiologists  be- 
cause of  several  recent  findings  which  seem  to  implicate  RNA 
in  the  storage  of  information  in  the  nervous  system.  I  was  par- 
ticularly interested  by  the  fact  that,  in  the  system  as  you  describe 
it,  RNA  seems  to  be  relatively  insulated  from  the  cellular  environ- 
ment in  which  it  is  located.  Since  a  cellular  information  storage 
mechanism  must  somehow  reflect  events  in  the  environment,  it 
would  seem  that  RNA  would  have  to  be  excluded  from  considera- 
tion if  these  processes  were  as  immutable  as  you  describe  them. 
Therefore,  I  would  like  to  ask  explicitly  what  you  might  expect 
if  for  example,  radical  changes  in  ionic  concentration  were  to 
occur  in  the  environment  where  the  processes  which  you  have 
described  were  taking  place.  Would  the  outcome  still  be  quite 
as  determinate  as  the  impression  that  you  gave?  In  relation  to 
this  question,  I  recall  a  paper  by  Rudenberg  and  Tobias,  in  which 
they  were  suggesting  that  a  certain  amount  of  calcium  is  bound 
to  RNA  in  axoplasm.  Is  it  possible  that  various  ionic  substances 
can  bind  to  RNA  in  certain  stages  of  these  processes,  modifying 
the  processes,  so  that  the  macromolecule  which  will  be  syn- 
thesized is  not  completely  specified  by  the  RNA? 

Saul  Kit  (Houston,  Texas) :  I  think  that  your  point  is  extremely 
interesting.  There  are  two  aspects  on  which  I  might  comment. 


Pyrimidine  Moieties  in  Animals,  Plants,  and  Bacteria  121 

First,  the  whole  question  of  the  ions  is  an  extremely  important 
one.  Ions  are  bound  to  the  phosphoric  acid  groups  of  the  nucleic 
acid  molecules.  Possibly,  they  are  involved  in  the  regulation  of 
gene  function.  A  DNA  molecule  has  two  functions — either  to 
serve  as  a  template  for  its  own  replication — or  to  serve  as  a  tem- 
plate for  the  replication  of  RNA.  What  controls  these  alternative 
functions?  What  are  the  controls  which  determine  which  and  how 
many  of  the  present  cells  will  be  functional?  The  answer  to  these 
questions  possibly  lies  in  the  interaction  between  the  nucleic  acids, 
cations,  and  certain  proteins.  Nucleic  acid  configuration  and 
length  may  be  greatly  modified  by  the  ionic  environment  in  which 
the  nucleic  acid  is  found.  This  process  may  thus  regulate  some  of 
the  cellular  biosynthesis. 

There  was  a  point  brought  out  in  the  discussion  of  Dr.  Echol's 
paper  that  should  be  made  more  explicit.  It  is  possible  that  nucleic 
acids  function  in  the  spread  of  information  among  cells.  This  may 
be  accomplished  either  by  the  free  nucleic  acids  or  by  the  viral 
nucleic  acids.  The  latter  may  be  thought  of  either  as  messengers 
or  as  transducing  viruses.  Herriot,  in  particular,  has  been  in- 
terested in  the  biological  significance  of  extracellular  nucleic  acids. 

Mike  McGlothlen  (Houston,  Texas):  In  the  synthesis  of  RNA, 
the  DNA  would  carry  a  code  dictating  the  formation  of  RNA, 
which  in  turn  dictates  the  formation  of  proteins.  Your  last  slide 
showed  the  RNA  forming  on  the  DNA  template  a  homologous 
process  to  that  proposed  for  reproduction  of  DNA.  Why  do  you 
have  to  have  this  apparent  splitting  of  hydrogen  bonds?  Why  not 
have  a  similar  code  mechanism  in  the  DNA,  such  as  two  or  three 
units  of  the  DNA  chain  dictating  the  reproduction  of  one  unit  of 
the  RNA  chain?  Why  the  same  mechanism  for  the  production 
of  RNA  and  DNA  from  a  DNA  chain,  rather  than  two  different 
methods  involving  dissimilar  codes? 

Heather  D.  Mayor  (Houston,  Texas):  I  think  this  question 
results  from  difficulty  in  understanding  the  diagram.  We  do  not 
really  know  whether  one  part  of  the  molecule  is  busy  transcribing 
RNA,  while  at  the  same  time  the  rest  of  it  is  replicating  DNA. 
Even  though  these  points  are  not  settled,  such  a  diagram  is  useful 
for  the  purpose  of  documenting  what  is  known  to  be  going  on. 
It  simply  indicates  the  double  helix  coming  apart  in  this  manner 


122  Information  Storage  and  Neural  Control 

and  transcribing  in  the  same  letters,  but  perhaps  smaller,  the 
information  which  is  in  the  DNA.  Whether  this  is  done  by  three 
bases  together  in  the  DNA,  or  whether  the  number  is  twenty  is 
not  at  the  moment  important  as  long  as  we  get  the  general  effect 
across. 

Kit:  The  problem  that  faces  us  is  this:  Proteins  vary  greatly 
in  their  length;  they  can  be  very  short  or  they  can  be  very  long. 
How  could  we  obtain  proteins  of  various  sizes  if  the  messenger- 
RNA's  were  all  of  uniform  sizes?  I  was  deliberately  ambiguous 
on  this  point  in  the  diagram  because  we  do  not  really  know  how 
many  proteins  are  coded  by  one  DNA  molecule.  There  might 
be  several.  In  other  words,  we  do  not  know  where  the  periods 
and  where  the  commas  are  on  the  DNA  chain. 

Mayor:  I  am  interested  to  see  that  the  minimal  infective 
amount  of  DNA  in  the  animal  viruses  appears  to  be  around 
4  X  10*^,  whereas,  the  molecular  weight  of  DNA,  for  instance,  in 
T-2  phage  is  about  120  x  10^  Would  this  be  because  more  in- 
formation is  necessarily  contained  in  the  DNA  of  phage  in  com- 
parison with  the  animal  viruses,  or  would  this  be  a  problem  in 
redundancy  again?  Do  you  think  that  there  must  be  more  infor- 
mation contained  in  the  DNA  of  T-2  than  in  rabbit  papilloma 
SV-40,  or  any  of  the  animal  viruses? 

Kit:  I  think  it  is  quite  likely  that  there  is  more  information  in 
vaccinia  or  T-2  phage  than  in  ^X174  polio  or  Shope  papilloma 
virus.  The  minimum  information  that  must  be  present  in  a  virus 
is  the  information  necessary  to  specify  the  protein  coat  of  the 
virus.  For  tobacco  mosaic  virus  this  would  have  to  be  enough 
information  to  specify  a  protein  having  a  molecular  weight  of 
17,500.  This  would  require  about  900  nucleotides  on  the  basis 
of  a  triplet  code,  x^ctually,  there  are  about  6,000  nucleotides  in 
a  TMV-RNA  chain.  Thus,  it  is  very  likely  that  there  is  additional 
information  even  in  tobacco  mosaic  virus,  or  in  other  small 
viruses.  I  think  that  Dr.  Darnell  will  elaborate  on  this  point  in 
connection  with  the  genetic  information  brought  in  by  T-2  phage 
DNA  for  specific  protein  synthesis. 


CHAPTER 
VI 

VIRUS  ACTION  AND  REPLICATION* 

James  E.  Darnell,  Jr.,  M.D. 

INTRODUCTION 

VV  HEN  the  genetic  composition  of  organisms  is  tiiought  of  in 
terms  of  information  storage,  it  is  immediately  apparent  why 
viruses,  which  represent  the  smallest  storehouses  of  biological 
information,  and  thus  probably  the  least  complicated,  have  been 
such  popular  research  tools.  Since  the  discox'cry  of  the  nucleo- 
proteinic  and  molecular  nature  of  viruses  in  1936,  the  study  of 
virus  action  and  replication  has  contributed  greatly  to  the  present 
knowledge  of  how  genetic  information  is  stored  and  expressed. 
I  will  limit  my  discussion  primarily  to  a  summary  of  the  events 
in  bacteriophage  infection,  the  most  thoroughly  understood  virus 
cycle,  and  to  a  brief  discussion  about  recent  work  using  poliovirus 
as  a  model  for  studying  animal  virus  replication. 

MACROMOLECULAR  EVENTS  IN 
BACTERIOPHAGE  INFECTION 

Bacteriophage  infection  is  initiated  by  attachment  of  the  phage 
particle  to  a  susceptible  cell  followed  by  the  injection  of  the  DNA 
of  the  phage  through  a  hole  produced  in  the  cell  wall  by  a  lyso- 
zyme  contained  within  the  phage  tail  (1). 

Two  types  of  response  to  infection  by  phage  may  occur  in 
bacteria:  1)  the  lytic  response  in  which  phage  multiplication  is 
accompanied  by  cell  death  and  lysis;  2)  the  lysogenic  response  in 
which  the  phage  genome  becomes  integrated  with  the  host  cell 


*This  work  was  supported  in  part  by  a  research  grant  from  the  National  Institutes 
of  Health   (C-5789}. 

123 


124  Information  Storage  and  Neural  Control 

and  the  ability  to  make  this  phage  is  transmitted  as  a  heritable 
property  of  the  cell.  Temperate  phage  infections  may  lead  to 
either  of  these  results;  intemperate  phages  can  only  cause  the 
lytic  response.  We  shall  be  concerned  mainly  with  describing  the 
synthetic  capacities  of  cells  undergoing  the  lytic  response  to  infec- 
tion by  the  intemperate  T-even  series  of  phages. 

The  entry  into  the  cell  of  the  DNA  from  such  a  phage  brings 
about  immediate  and  dramatic  changes  in  synthetic  events  within 
the  cell.  Cellular  DNA,  RNA,  and  protein  syntheses  seem  to  be 
stopped  immediately  and  no  further  cell  division  takes  place  (2). 
The  DNA  of  the  cell  is  digested  by  a  DNAse  and  contributes 
nucleotides  to  phage  DNA  (3,  4,  5).  Although  many  enzymes  and 
metabolic  pathways  within  the  cell  are  able  to  function  (2),  the 
integration  of  events  which  previously  led  to  cellular  macro- 
molecular  synthesis  and  continued  cell  growth  is  disrupted.  Recent 
experiments  from  several  different  laboratories  (6)  offer  a  reason- 
able explanation  of  the  subsequent  events  in  the  course  of  synthesis 
ol  new  phage  particles. 

An  outline  of  the  new  work  is  best  begun  by  describing  the 
concept  of  messenger  RNA  as  it  functions  in  phage  infection. 
Current  ideas  of  the  genetic  control  of  protein  synthesis  delegate 
to  DNA  the  role  of  carrier  of  genetic  information.  The  structural 
site  of  protein  formation  has  been  shown  to  be  the  ribosome  (7,  8), 
which,  however,  is  composed  of  RNA  and  protein  (9).  Thus,  it 
has  been  presumed  that  some  RNA  molecule  probably  served  to 
transport  information  from  the  DNA  to  the  ribosome.  The  first 
evidence  of  such  an  RNA  was  obtained  by  Volkin  and  Astrachan 
(10),  who  found  in  phage-infected  cells  that  just  after  infection 
a  species  of  RNA  was  formed  which  had  base  ratios  (substituting 
uracil  for  thymine  and  cytosine  for  5-hydroxymethylcytosine) 
similar  to  those  of  infecting  phage  DNA.  Ribosomal  RNA  does 
not  bear  such  a  relationship  to  DNA  (11).  It  has  since  been  shown 
that  this  newly  fornied  RNA  is  linked  to  ribosomes,  which  form 
phage  proteins,  but  is  not  the  ribosomal  RNA  itself  (12,  13). 
Hall  and  Spiegelman  (14)  have  performed  a  critical  test  of  the 
source  of  this  RNA  by  demonstrating  that  it  can  combine  phys- 
ically by  hydrogen  bonding  with  phage  DNA,  which  directed  its 
formation,   but  not  with  any  other  DNA.  Thus,  one  concludes 


Virus  Action  and  Replication  125 

that  when  DNA  of  an  intemperate  phage  gets  inside  the  ceU,  the 
cellular  DNA  is  destroyed  and  can  no  longer  serve  as  a  source  of 
information.  The  phage  DNA  then  assumes  command  of  the 
synthetic  machinery  in  the  cell  via  messenger  RNA  copied  from 
itself  by  hydrogen  bonding  between  appropriate  base  pairs.  This 
results  in  formation  by  the  cells'  own  ribosomes  of  phage-controlled 
proteins  only. 

The  next  point  of  interest  is:  Wliat  proteins  does  the  virus 
instruct  the  cell  to  make?  The  DNA  of  the  T-even  phages  is 
chemically  peculiar  in  that  it  contains  the  base  5-hydroxymethyl- 
cytosine  (HMC)  in  place  of  the  normally  occurring"  cytosine  (15), 
and  in  that  some  of  the  HMC  residues  hav^e  glucose  attached  to 
them  after  incorporation  into  the  DNA  chain  (16).  Thus,  if 
replication  of  the  bacteriophage  is  to  occur,  a  phage-infected  cell 
must  be  able  to  perform  enzyme  reactions  not  possible  in  an 
uninfected  cell.  In  the  laboratories  of  Dr.  S.  S.  Cohen  and  Dr. 
Arthur  Kornberg  it  has  been  demonstrated  that  phage-infected 
cells  do  indeed  acquire  a  large  number  of  new  enzyme  activities 
within  minutes  after  infection  (17,  18,  19).  These  are  the  first 
proteins  formed  by  the  infected  cell  in  order  that  phage  DNA  may 
be  replicated.  Later  in  the  course  of  infection,  DNA  synthesis  and 
structural  phage  protein  synthesis  begin. 

Cell  lysis,  which  occurs  after  several  hundred  new  phages  per 
cell  have  been  produced,  is  caused  by  the  action  inside  the  cell 
of  a  lysozyme  which  is  also  formed  under  the  genetic  control  of 
the  phage  (20). 

Many  steps  in  the  infectious  cycle  of  the  lytic  phages  are  ob- 
viously well  understood  on  a  molecular  level.  The  intricacies  of 
the  second  type  of  response  to  phage  infection,  the  lysogenic 
response,  have  not  yet  been  .so  thoroughly  elucidated.  In  a  cell 
which  is  infected  by  a  temperate  phage  (one  capable  of  inducing 
lysogeny),  as  soon  as  the  DNx^  enters  the  cell,  differences  from  the 
course  of  events  in  infection  with  a  member  of  the  T-series  of 
phages  can  be  detected.  Even  if  the  pathway  to  the  lytic  response 
is  followed,  the  cell  continues  to  synthesize  cellular  protein  and 
RNA  (21),  and  can  even  be  induced  to  foim  new  enzymes  through- 
out the  latent  period  of  the  virus  (22).  If  the  cell  is  to  become 
lysogenized,  then  the  synthesis  of  RNA,  DNA,  and  protein  stops. 


126  Information  Storage  and  Neural  Control 

In  either  case,  the  DNA  of  the  cell  is  not  destroyed.  In  the  cell 
undergoing  the  lysogenic  response  about  a  two-hour  lag  occurs 
after  infection  before  the  cell  resumes  growth  (23).  The  phage 
genome  has  by  this  time  become  attached  to  the  bacterial  chromo- 
some where  it  is  carried  in  the  form  of  prophage  as  a  new  genetic 
character  of  the  cell.  According  to  current  belief,  phage  genes  in 
a  lysogenic  cell  do  not  function  because  a  repressor  is  formed  by 
one  phage  gene  which  prevents  the  expression  of  the  others  (24). 
Certain  mutations  (virulent)  among  temperate  phages  result  in  a 
change  in  character  of  the  phage  so  that  it  no  longer  can  lysogenize 
cells  but  can  only  lyse  them.  The  mutation  is  presumably  due  to 
a  loss  of  the  repressor.  A  strong  piece  of  evidence  in  support  of 
this  hypothesis  is  that  separately  arising,  virulent  mutants  can 
complement  each  other  during  mixed  infections  to  produce 
lysogenization    (25,    26). 

The  basis  for  the  profound  difference  in  effect  on  the  bacterial 
chromosome  between  temperate  and  intemperate  phages  is  ill 
understood  at  present,  but  may  be  related  in  some  way  to  the 
close  relationship  between  phage  DNA  and  host  DNA  in  the  case 
of  the  temperate  phages.  For  instance,  base  ratios  between  tem- 
perate phages  and  their  hosts  are  similar  (27).  In  addition,  the 
capacity  of  a  bacterial  cell  to  support  multiplication  of  a  temperate 
phage  is  much  more  sensitive  to  inactivation  by  physical  agents 
which  damage  the  nucleic  acids  than  in  the  capacity  to  form  lytic 
phages  (28).  Also,  unirradiated  host  cells  are  able  to  repair  damage 
to  irradiated  temperate  phages,  permitting  growth  of  the  phage. 
It  is  therefore  clear  that  during  the  replication  of  temperate  phages 
there  is  very  close  interrelationship  between  intact  functioning 
cellular  DNA  and  phage  DNA.  It  is  widely  presumed  that  this 
affords  opportunity  for  the  attachment  of  the  phage  DNA  to  the 
chromosome  of  the  cell  and  for  the  establishment  of  lysogeny. 

ANIMAL  VIRUSES:  EFFECT  ON  CELLULAR  SYNTHESIS 

With  this  brief  summary  of  the  possible  interrelationships  be- 
tween bacteriophages  and  their  hosts  in  mind,  we  will  now  discuss 
the  impact  of  viral  infection  on  the  synthetic  processes  in  animal 
cells.    In   the  past   ten  years,   techniques  for   the  study  of  many 


Virus  Action  and  Replication  127 

different  animal  viruses  in  cell  cultures  have  been  developed.  This 
approach  lias,  for  the  first  time,  allowed  animal  virology  to  be 
studied  at  the  cellular  level  with  homogenous  populations  of  cells 
which  can  be  simultaneously  infected  (30). 

It  is  apparent  that  animal  viruses  present  a  vastly  less  homo- 
genous group  of  agents  tlian  bacteriophages.  The  host  range  in- 
cludes virtually  all  animals  from  single  celled  organisms  to  verte- 
brates. These  viruses  range  in  size  froml50A  to  3000A;  they  are 
of  many  different  shapes  and  show  great  variability  in  chemical 
composition.  Moreover,  the  host  cell  for  an  animal  virus  is  con- 
siderably more  complex  than  a  bacterial  cell.  It  is,  tlierefore,  not 
surprising  that  the  interrelationships  between  animal  viruses  and 
their  host  cells  are  quite  varied  and  complex.  This  discussion  will 
center  on  one  animal  virus-cell  system,  the  poliovirus  infected 
HeLa  cell,  which  is  one  of  the  most  thoroughly  studied  animal 
virus  systems. 

First,  a  few  of  the  chemical  and  physical  properties  of  polio- 
virus  should  be  stated  (31).  It  is  a  small  (300  A  in  diameter) 
so-called  spherical  virus  which  is  composed  of  RNA,  25  per  cent, 
and  protein,  75  per  cent.  The  RNA  is  enclosed  within  a  protein 
shell  made  of  subunits  which  are  arranged  in  a  symmetrical  form 
(icosahedral)  on  the  particle  surface  (32,  33).  There  are  no  lipids 
or  cell-derived  macromolecules  attached  to  the  virus  particle  (31). 
The  infectious-unit-to-particle  ratio  is  low  (of  the  order  of  1  in  100) 
in  both  crude  and  purified  virus  suspensions  (31). 

The  initial  event  of  infection  with  poliovirus  in  a  normally 
susceptible  cell  is  adsorption,  which  is,  at  least  to  some  extent, 
dependent  on  the  ionic  strength  of  the  medium  and  on  the  presence 
of  divalent  cations  (34,  35).  After  adsorption  the  infectious  virus 
disappears  rapidly  from  the  surface  of  the  cell.  It  can  be  shown 
with  virus  labeled  by  P'^"  in  its  RNA  that  all  particles  in  a  purified 
suspension  can  adsorb  to  the  cell  but  that  about  50  per  cent  of 
these  come  back  off  the  cell  in  a  non-infectious  and  non-adsorbable 
state  (36).  The  extracted  RNA  in  these  particles  is  still  as  infec- 
tious as  in  an  unexposed  suspension,  however.  The  RNA  of  most 
of  the  particles  which  do  remain  attached  to  the  cell  is  degraded 
to  small  pieces.  About  10  per  cent  of  the  attached  particles  remain 
unchanged  and  about  10  per  cent  are  changed  in  such  a  way  that 


128  Information  Storage  and  Neural  Control 

the  viral  RNA  has  become  susceptible  to  added  RNAse  but  is 
still  in  large  molecular  weight  form.  It  is  these  latter  particles 
which  could  be  expected  to  function  as  units  responsible  for  virus 
replication.  There  is  still  a  great  excess  of  particles  (about  10  fold) 
in  this  state  over  the  number  of  actual  infectious  units,  a  fact  for 
which  an  adequate  explanation  is  lacking  at  present. 

The  HeLa  cell  which  is  growing  exponentially  at  the  time  of 
poliovirus  infection  is  drastically  altered  soon  after  infection. 
Salzman  et  al.  (37)  have  shown  that  net  increases  in  protein,  DNA 
and  RNA  all  cease  upon  infection.  On  the  other  hand,  amino 
acid  incorporation  continues  after  infection,  but  at  a  decreasing 
rate  (38),  and  incorporation  of  radioactive  nucleic  acid  precursors 
into  RNA  goes  on  at  approximately  the  same  rate  but  with  a 
different  pattern.  This  changed  pattern,  however,  is  nonspecific 
in  the  sense  that  halting  growth  by  amino  acid  deprivation  results 
in  the  same  change. 

It  is  important  to  remember  in  considering  experiments  of  this 
nature  that  a  polio  infected  cell  makes  at  most  only  .5  per  cent  of 
its  dry  weight  into  virus  and  that  these  substantial  amounts  of 
incorporation  represent  more  RNA  and  protein  than  eventually 
appear  as  virus  material.  Whether  this  indicates  the  formation 
of  new  material  under  the  direction  of  the  virus  or  the  turnover 
of  pre-existing  cellular  components  is  unknown. 

Late  in  the  course  of  infection,  cellular  RNA  is  degraded  and 
lost  into  the  medium  as  is  a  substantial  amount  of  cellular  protein 
(37).  This  may  represent  a  specific  kind  of  loss.  RNA  is  lost  before 
protein  and  both  are  lost  prior  to  the  liberation  of  virus  into  the 
medium,  a  process  which  occurs  as  a  burst  and  may  be  akin  to 
lysis  of  a  bacterial  cell  after  bacteriophage  production  (39,  40). 

Poliovirus  is  an  extreme  example  of  a  virus  with  a  destructive 
action  on  its  host  cell.  Other  viruses  have  been  shown  to  possess 
lethal  capacity  for  host  cells  without  the  almost  complete  destruc- 
tion which  is  accompanied  by  virus  release.  Examples  of  this  type 
of  interaction  are  found  in  adenovirus  (41),  herpes  simplex  (42), 
and  vaccinia  (43)  infections.  All  of  these  are  held  intracellularly 
to  the  extent  of  about  90  per  cent.  In  each  of  these  cases  the  virus 
apparently  contains  DNA.  In  adenovirus  and  herpes  infections 
there  is  stimulation   of  DNA  synthesis   to  greater  than  normal 


Virus  Action  and  Replication  129 

levels  (44,  45,  46)  during  the  latent  period.  Another  gradation  of 
the  disruptive  effect  of  a  mukiplying  virus  on  its  host  cell  is  found 
in  the  case  of  the  myxoviruses,  influenza  and  Newcastle  disease 
virus.  It  has  been  shown  clearly  that,  although  cell  death  may  be 
the  eventual  outcome  of  encounters  between  these  viruses  and 
HeLa  cells  (47),  infected  cells  can  definitely  divide  even  after 
production  of  viral  protein  has  started  (48). 

One  of  the  most  interesting  and  perhaps  most  important  kinds 
of  relationship  in  animal  virology  is  the  tumor  virus-cell  inter- 
action. Of  the  tumor  viruses  which  have  been  adapted  to  study 
in  cell  culture,  two,  polyoma  virus  and  Rous  sarcoma  virus,  have 
been  most  useful  in  following  the  outcome  of  individual  cells  after 
infection  (49,  50,  51).  The  influence  of  infection  with  these  viruses 
on  the  overall  synthetic  capacities  of  cells  is  not  yet  known.  The 
DNA-containing"  polyoma  virus  has  been  shown  to  have,  initially, 
a  lytic  eff"ect  on  mouse  and  hamster  cells  which,  except  for  being 
slower,  is  similar  in  general  pattern  to  the  action  of  a  virus  like 
polio  (52). 

Eventually,  cultures  which  have  been  infected  with  polyoma 
undergo  a  microscopic  change  and  simultaneous  biologic  transition 
to  cultures  which  no  longer  produce  virus  (or  do  so  at  a  very 
low  rate)  but  which  have  acquired  the  capacity  to  cause  tumors 
in  animals.  In  an  attempt  to  elucidate  the  role  of  the  polyoma 
virus  in  this  transition,  Vogt  and  Dulbecco  have  analyzed  single 
cells  and  have  found  that  transformed  cells  which  continue  to 
divide  indefinitely  do  not  produce  virus  and  have  an  increased 
resistance  to  infection  by  polyoma  (53).  They  were  unable  to 
obtain  any  evidence  that  the  viral  genome  was  present  in  the 
cell  (54).  Thus,  the  vital  question  of  whether  the  virus  has  inte- 
grated with  the  cell  to  cause  the  change  to  a  neoplastic  unit  or 
whether  the  virus,  by  creating  a  selective  pressure  or  by  invoking 
a  developmental  change,  has  aided  in  the  establishing  of  a  line 
of  cells  with  neoplastic  capacity  and  increased  virus  resistance, 
remains  unanswered. 

The  other  tumor  virus  which  has  been  extensively  studied  in 
vitro,  using  quantitative  techniques,  is  the  Rous  sarcoma  virus. 
Rubin  and  Temin  have  analyzed  the  tumor  cell  transition  in 
chicken  fibroblasts  brought  about  by  this  RNA-containing  virus 


1 30  Information  Storage  and  Neural  Control 

(55,  56).  Under  the  usual  experimental  conditions  Rous  infected 
cells  do  not  lyse  but  become  transformed  after  one  to  three  days 
into  tumor  cells  which  grow  as  changed  clones  of  cells.  Each 
transformed  cell  is  capable  of  liberating"  virus  at  a  very  slow  rate 
while  continuing  to  grow.  This  kind  of  integration  between  virus 
and  cell  has  no  counterpart  in  bacteriophage  systems. 

In  none  of  the  animal  virus  systems  studied  is  there  any  detailed 
knowledge  of  how  the  damaging  effect  of  the  invading  virus  is 
mediated,  nor  is  there  any  understanding  of  how  the  tumor 
viruses  become  integrated  with  the  cell. 

ANIMAL  VIRUSES:  FORMATION  OF  VIRAL 
PRECURSOR  MOLECULES 

It  will  be  recalled  from  the  phage  work  that  two  classes  of 
macromolecules  are  formed  after  phage  infection:  1)  messenger- 
RNA  and  "early"  enzyme  proteins  in  which  the  ultimate  function 
is  to  allow  phage  replication,  and  2)  the  phage  precursor  DNA 
and  proteins  which  eventually  form  the  new  particles.  In  the  case 
of  animal  viruses  this  first  class  of  new  products  has  not  been 
proved  definitely  to  exist.  There  are  several  reported  instances 
of  materials  which  are  apparently  formed  by  infected  cells  after 
virus  infection,  but  whether  these  are  genetically  specified  by  the 
infecting  virus  is  unknown.  This  group  of  materials  includes 
interferon,  which  is  produced  by  a  variety  of  cells  after  infection 
by  a  variety  of  viruses  (57),  a  cell  detachment  factor  produced 
by  adenovirus  infected  HeLa  cells,  which  is  serologically  distinct 
from  the  virus  (58)  and  arginase,  which  is  increased  in  cells 
infected  with  papilloma  virus  (59).  There  is  no  apparent  con- 
nection between  the  formation  of  any  of  these  substances  and  the 
replication  of  the  virus  concerned. 

Studies  on  newly  formed  products  in  infected  cells  have,  there- 
fore, been  limited  largely  to  the  study  of  virus  precursor  molecules. 
The  techniques  that  have  been  most  useful  in  identifying  the  time 
and  place  of  both  synthesis  of  viral  precursor  molecules  and 
maturation  of  whole  virus  particles  are:  electron  microscopy, 
flourescent  antibody  staining  and  other  types  of  immunologic 
identification,  extraction  and  measurement  of  infectious  nucleic 
acid,  and  radioisotopic  labeling  and  purification  of  virus. 


Virus  Action  and  Replication  131 

Two  kinds  of  viruses,  both  containing  RNA,  have  been  studied 
most  extensively  in  this  way— myxoviruses  (specifically  influenza 
and  fowl  plague  virus)  and  small  spherical  viruses  (poliovirus, 
encephalomyocarditis  virus  and  Western  equine  encephalitis  virus) . 
I  shall  describe  briefly  the  sequence  of  events  in  the  formation  of 
fowl  plague  virus  and  poliovirus. 

Fowl  Plague  Virus 

Work  in  Schafer's  laboratory  in  Tubingen  has  established  that 
fowl  plague  virus  consists  of  at  least  two  proteins  plus  a  lipid  and 
RNA  (60).  One  protein,  which  is  associated  with  the  RNA  in 
the  S  antigen  and  which  can  be  released  from  the  whole  virus 
particle  by  ether  treatment,  still  in  association  with  the  RNA, 
first  becomes  detectable  by  fluorescent  antibody  staining  in  the 
nucleus  of  cells  three  hours  after  infection.  By  four  hours  after 
infection  it  is  also  found  in  the  cytoplasm.  The  RNA  which  is 
enclosed  by  this  protein  is  also  presumed  to  be  formed  in  the 
nucleus.  The  hemagglutinating  antigen,  on  the  other  hand,  is 
formed  in  the  cytoplasm.  By  the  use  of  5-fluorophenylalanine 
(FPA)  as  an  inhibitor  of  viral  formation,  the  sequence  of  forma- 
tion of  proteins  has  been  determined.  Infected  cells  exposed  to 
FPA  before  one  hour  after  infection  form  no  viral  antigens.  Since 
the  virus  apparently  penetrates  the  cell,  this  may  be  an  indication 
that  a  new  protein  other  than  viral  precursor  protein  must  be 
formed  to  allow  the  subsequent  steps  in  virus  production  to  occur. 
If  two  hours  elapse  before  treatment,  S  antigen  appears  to  be 
formed  normally,  but  neither  the  hemagglutinating  antigen  nor 
the  infectious  virus  is  foimed.  By  three  hours  after  infection  the 
production  of  hemagglutinating  activity  and  infectivity  have  begun 
to  escape  inhibition  and  by  six  to  seven  hours  after  infection  FPA 
has  no  effect.  The  formation  of  infectious  particles  is,  however, 
more  sensitive  to  FPA  than  the  formation  of  hemagglutinating 
antigen,  an  indication  that  these  are  separate  processes  (61). 

The  virus  particle  thus  has  its  origins  in  different  parts  of  the 
cell;  then,  by  a  transport  process  which  also  involves  protein 
synthesis,  the  whole  particle  is  brought  together  at  the  cell  surface 
and  completed.  Nothing  more  is  known  about  events  within  the 
first  hour  of  infection  which  are  necessary  for  the  initiation  of 
replication  of  viral  precursor  molecules. 


132  Information  Storage  and  Neural  Control 

Poliovirus  Biosynthesis:  Source  and  Time  Course  of  Synthesis 
of  Viral  Constituents 

I  will  now  describe  the  source  and  time  course  of  synthesis  of 
the  RNA  and  protein  of  poliovirus  and  then  discuss  some  new 
evidence  relating  to  the  questions  of  1)  the  necessity  for  "early" 
protein  formation  prior  to  actual  virus  replication,  and  2)  the  role 
of  poliovirus  RNA  as  a  "messenger"  RNA. 

By  differentially  labeling  the  macromolecules  (protein  or  RNA) 
and  the  acid-soluble  pool  (amino  acids  or  nucleotides)  of  HeLa 
cells,  and  observing  the  production  of  poliovirus  under  these  con- 
ditions, it  was  shown  that  viral  macromolecules  were  constructed 
de  novo  in  the  infected  cell  from  the  acid-soluble  pool.  This  was 
true  for  both  viral  RNA  and  viral  piotein  (38,  62). 

To  determine  the  time  at  which  viral  macromolecules  were 
synthesized  relative  to  the  maturation  cycle,  radioactive  precursors 
of  either  protein  or  RNA  were  added  to  the  medium  of  infected 
cells  at  various  times  after  infection  and  virus  purified  at  the  end 
of  maturation  (38,  63).  From  the  isotope  content  of  the  purified 
virus  could  be  determined  how  much  of  the  virus  had  been  syn- 
thesized at  the  time  of  addition  of  the  radioisotope.  Virus  protein 
and  virus  RNA  were  shown  to  be  formed  between  two  and  one- 
half  and  six  hours  after  infection  and  there  was  very  little  lag 
between  the  onset  of  formation  of  viral  macromolecules  and  of 
the  whole  virus.  An  independent  confirmation  of  this  last  state- 
ment was  obtained  by  determining  the  times  of  formation  of 
infectious  RNA  (ribonuclease  sensitive  plaque-forming  activity) 
and  of  whole  virus.  Here  the  very  earliest  increases  of  infectious 
material  could  be  determined  (the  first  0.1  per  cent  of  new  virus 
or  infectious  RNA).  It  was  found  that  infectious  RNA  began  to 
increase  at  about  2-2.5  hours  while  an  increase  in  whole  virus 
began  approximately  thirty  minutes  later.  It  would  appear,  then, 
that  for  the  first  2-2.5  hours  of  the  infectious  cycle  the  cell  does 
not  make  virus  precursor  molecules. 

The  essential  findings  of  the  above  experiments  are  borne  out 
in  electron  microscopic  investigations  of  infected  cells.  Home  and 
Nagington  (33)  found  evidence  in  electron  photomicrographs  of 
circumscribed  areas  of  apparent  multiplication  of  protein  sub- 
units  in  the  cytoplasm  of  HeLa  cells,  beginning  about  three  hours 


Virus  Action  and  Replication  133 

after  infection.  Fogh  and  Stuart  (64)  published  beautiful  pictures 
of  crystalline  arrays  of  whole  virus  particles  in  the  cytoplasm  of 
several  kinds  of  cells.  These  crystals  first  appeared  five  to  six 
hours  after  infection. 

In  an  effort  to  determine  whether  protein  synthesis  is  required 
for  any  steps  in  poliovirus  multiplication  prior  to  the  foimation 
of  virus  precursor  molecules,  experiments  using  inhibitors  of  pro- 
tein synthesis  were  performed  during  that  time  period.  The  effect 
of  the  inhibitors  on  whole  virus  multiplication  and  on  infectious 
RNA  formation  was  measured  (65). 

The  amino  acid  analog"  5-fluorophenylalanine,  which  had 
originally  been  shown  by  Ackerman  et  al.  (66)  to  inhibit  polio- 
virus  multiplication,  was  found  to  be  effective  at  a  concentration 
of  about  0.05  mM  in  completely  preventing  whole  virus  synthesis 
while  affecting  synthesis  of  infectious  RNA  only  slightly,  if  at  all. 
If  FPA  was  left  in  a  culture  past  the  usual  time  for  the  onset  of 
maturation  and  then  the  effects  of  the  drug  were  reversed  by 
addition  of  a  large  excess  of  L-phenylalanine,  maturation  and 
virus-protein  synthesis  began  within  an  hour,  indicating  that  all 
possible  steps  except  the  formation  of  viral  coat  protein  had 
occurred  in  a  normal  fashion  in  the  presence  of  FPA.  Puromycin,  a 
drug  which  inhibits  protein  synthesis  specifically  in  several  systems 
in  a  rather  different  manner  from  that  of  FPA  (67,  68,  69),  was 
found  to  inhibit  both  infectious  RNA  and  whole  virus  synthesis 
when  added  prior  to  two  hours  after  infection.  If  the  puromycin 
was  added  at  2.5  hours,  then  about  15  per  cent  as  much  infectious 
RNA  was  formed  as  in  controls,  without  the  formation  of  any 
wliole  virus;  and  if  puromycin  was  added  at  three  hours,  the 
cells  produced  30  to  100  per  cent  of  the  normal  amounts  of  infec- 
tious RNA  but  only  about  5  to  10  per  cent  of  the  normal  yield 
of  whole  virus.  Addition  of  puromycin  at  the  outset  of  infection 
and  removal  1.5  hours  later  resulted  in  a  corresponding  delay  in 
the  beginning  of  infectious  RNA  synthesis. 

All  these  experiments  taken  together  provide  suggestive  evidence, 
but  not  proof,  that  the  synthesis  of  some  material,  probably  protein 
in  nature,  is  necessary  for  initiating  the  replication  of  poliovirus 
RNA.  This  step  in  virus  formation  is  not  sufficiently  complete 
by  two  hours  after  infection  to  allow  RNA  replication. 


134  Information  Storage  and  Neural  Control 

The  discovery  by  Nirenberg  and  Matthaei  (70)  of  an  in  vitro 
system  for  synthesizing  protein  which  is  dependent  on  the  addition 
of  a  messenger-RNA  offers  the  opportunity  to  attack  this  problem 
directly.  In  this  system  which  is  derived  from  Escherichia  coli  cells, 
these  workers  found  that  tobacco  mosaic  virus  RNA  stimulated 
the  formation  of  TMV  coat  protein,  indicating  that  the  virus  RNA 
was  itself  the  active  messenger  in  protein  synthesis.  We  have  found 
recently  that  this  system  also  responds  to  the  addition  of  poliovirus 
RNA  by  producing  material  which  will  specifically  precipitate 
with  poliovirus  antiserum.  Thus,  if  the  poliovirus  RNA  specifies 
the  formation  of  other  proteins,  these  should  also  be  formed  in 
this  in  vitro  system. 

We  can  speculate  as  to  the  type  of  protein  tiiat  poliovirus  might 
require  to  assure  its  own  synthesis  in  HeLa  cells.  The  virus  con- 
tains no  nucleic  acid  bases  (31)  or  amino  acids  (38)  foreign  to  the 
cell.  Moreover,  with  tiie  possible  exception  of  guanine  and  cytosine 
nucleotides,  the  cell  contains  more  than  enough  acid-soluble 
material  to  provide  for  the  synthesis  of  all  viral  material  formed. 
Tlius  some  enzyme  protein  of  a  type  the  cell  already  possesses, 
but  which  is  either  under  cellular  control  or  is  located  at  a  position 
in  the  cell  which  is  inaccessible  to  the  virus,  might  be  used  by  the 
virus  as  a  mechanism  for  escaping  cellular  control. 

REFERENCES 

1.  Kozloff,  L.  M.,  Lute,  M.,  and  Henderson,   K.:   Viral   Inv^asion  I, 

II,  III.  J.  Bio.  Chem.,  225.-51 1-546,  1957. 

2.  Cohen,  S.  S.:  Growth  requirements  of  bacterial  viruses.  Bad.  Rev., 

7J.-1-24,  1949.  Cohen.  S.  S.:  Studies  on  Controlling  Mechanisms 
in  the  Metabolism  of  Virus-Infected  Bacteria.  CSH  Symp.  Qjtant. 
Biol.,  7<§.-221-235,  1953. 

3.  Luria,  S.  E.,  and  Human,  M.  L.:  Chromatin  staining  of  bactei'ia 

during  bacteriophage  infection.  J.   Bact.,  59.-551-560,   1950. 

4.  Kozloff,  L.  M.:  Origin  and  Fate  of  Bacteriophage  Material.  CSH 

Symp.  Quant.  Biol,  75;209-220,  1953. 

5.  Hershey,  A.  D.:  Bacteriophages  as  genetic  and  biochemical  systems. 

Advance  Virus  Res.,  4:25-6\,  1957. 

6.  Cold   Spring  Harbor  Symposia   on   Quantitative  Biology,   vol.   26, 

1961,  in  press. 


Virus  Action  and  Replication  135 

7.  McQuillen,   K.,   Roberts,   R.   B.,   and   Britten,   R.  J.:   Synthesis  of 

nascent  protein  by  ribosomes  in  E.  coli.  Proc.  Nat.  Acad.  Sci.  USA, 
45:\437,   1959. 

8.  Tissieres,  A.,  Schlessinger,  D.,  and  Gros,  F.:  Amino  acid  incorpora- 

tion into  proteins  by  E.  coli  ribosomes.  Proc.  Nat.  Acad.  Sci.  USA, 
4(5.- 1450-1 463,  1960. 

9.  Tissieres,  A.,   Watson,  J.   D.,   Schlessinger,   D.,   and  Hollingworth, 

B.  R.:  Ribonucleoprotein  particles  from  E.  coli.  J.  Molec.  Biol., 
7.-221-233,  1959. 

10.  Volkin,   E.   and  Astrachan,   L.:   RNA  Metabohsm  in  T-2-Infected 

E.  coli.  In  McElroy  and  Glass,  The  Chemical  Basis  of  Heredity. 
Baltimore,  Johns  Hopkins  Press,   1957. 

11.  Spahr,   P.   F.,   and  Tissieres,   A.:   Nucleotide  composition  of  ribo- 

nucleoprotein particles  from  E.  coli.  J.  Molec.  Biol.,  7:237-239, 1959. 

12.  Gros,  F.,  Hiatt,  H.,  Gilbert,  W.,  Kurland,  C.  G.,  Risebrough,  R.  W., 

and  Watson,  J.  D.:  Unstable  ribonucleic  acid  revealed  by  pulse 
labelling  of  E.  coli.  Nature,  790:581-584,  1961. 

13.  Brenner,  S.,  Jacob,  F.,  and  Meselson,  M.:  An  unstable  intermediate 

carrying  information  from  genes  to  ribosomes  for  protein  synthesis. 
Nature,  190:576-580,  1961. 

14.  Hall,    B.    D.,    and   Spiegelman,    S.:    Sequence   complementarity   of 

T2-DNA  and  T-2  specific  RNA.  Proc.  Nat.  Acad.  Sci.  USA,  47: 
137-146,   1961. 

15.  Wyatt,  G.  R.,  and  Cohen,  S.  S.:  A  new  pyrimidine  base  from  bac- 

teriophage nucleic  acids.  Nature,  770:1072-1073,  1952. 

16.  Sinsheimer,  R.  L.:  The  glucose  content  of  the  deoxyribonucleic  acids 

of  certain  bacteriophages.  Proc.  Nat.  Acad.  Sci.  USA,  42:502,  1956. 

17.  Flaks,  J.    G.,    Cohen,    S.    S.,    and   Lichtenstein,  J.:    Virus-induced 

acquisition  of  metabolic  function  I,  H.  J.  Biol.  Chem.,  2J4:1501- 
1511,  1959. 

18.  Kornberg,  A.,  Zimmerman,  J.  B.,  Kornberg,  S.  R.,  and  Josse,  J.: 

Enzymatic  synthesis  of  DNA,  influence  of  bacteriophage  T2  on 
the  synthetic  pathway  in  host  cells.  Proc.  Nat.  Acad.  Sci.  USA, 
45:772-785,  1959. 

19.  Aposhian,  H.  V.,  and  Kornberg,  A.:  The  DNA  polymerase  formed 

after  bacteriophage  infection  of  E.  coli — A  new  enzyme.  Fed.  Proc, 
20:361,  1961. 

20.  Streisinger,  G.:  CSH  Symp.  on  Quant.  Biol.,  vol.  28,  1961.  In  press. 

21.  Siminovitch,    L.:    Biochemical   modifications   of  the   bacterial   host 

during  bacteriophage  development.  Ann.  Inst.  Pasteur,  84:265-212, 
1953. 


136  Information  Storage  and  Neural  Control 

22.  Siminovitch,   L.   and  Jacob,   F.:   Biosynthese  induite  d\m  enzyme 

pendant  le  developpement  des  bacteriophages  chez  E.  coli  K12. 
Ann.  Inst.  Pasteur,  83:1  AS,   1952. 

23.  Frey,  B.  A.,  and  Gros,  F.:  The  metaboUc  activities  of  E.  coli  during 

the  establishment  of  lysogeny.  J.  Gen.  Microbiol.,  27.-655-692,  1959. 

24.  Jacob,  F.:   Genetic  Control  of  Viral  Functions,   in  Harvey  Lectures. 

1958-1959,  New  York,  Academic  Press,  Inc.,   1960. 

25.  Kaiser,  A.  D.:  Mutations  in  a  temperate  phage  P22  and  lysogeny  in 

salmonella.    Virology,  3; 42-61,   1957. 

26.  Levine,  M.:  Mutations  in  the  temperate  phage  P22  and  lysogeny 

in  salmonella.     Virology,  3:22-4\,  1957. 

27.  Luria,  S.  E.:  Control  of  Bacterial  Properties  by  Bacteriophage  Genes. 

Symposium  on  Virus-Cell  Relationship,  Oholo,  Israel,  1961. 

28.  Garen,  A.,  and  Zinder,   N.   D.:   Radiological  evidence  for  partial 

genetic    homology    between    bacteriophage    and    host    bacteria. 
Virology,   h'iAl-^ie,  1955. 

29.  Luria,    S.    E.:    The   multiplication   of  viruses,    Protoplasmatologia. 

Handbuch  der  Protoplasmqforschung,  4:1-63,  1958. 

30.  Dulbecco,  R.:  Interactions  of  viruses  and  animal  cells.  Physiol.  Rev 

J5.-301-335,   1955. 

31.  Schaffer,  F.  L.,  and  Schwerdt,  C.  E.:  Purification  and  properties 

of  poliovirus.  Advance  Virus  Res.,  6;159-204,  1959. 

32.  Fmch,  J.  T.,  and  Klug,  A.:  Structure  of  poliomyelitis  virus.  Nature, 

7.5J.-1709-1714,   1959. 

33.  Home,  R.  W.,  and  Nagington,  J.:  Electron  microscope  studies  of 

the  development  and  structure  of  poliomyelitis  virus.  J.   Molec. 
Biol.,  7.-333-338,  1959. 

34.  Bachtold,  J.  G.,  Bubel,  H.  C,  and  Gebhardt,  L.  P.:  The  primary 

mteraction  of  poliomyelitis  virus  with  host  cells  of  tissue  culture 
origin.   Virology,  4.- 582-589,  1957. 

35.  Holland,  J.  J.,   and   McLaren,   L.   C:   The  mammalian  cell-virus 

relationship  II.  Adsorption,  reception  and  eclipse  of  poliovirus  by 
HeLa  cells.  J.  Exp.  Med.,  109:487,  1959. 

36.  Joklik,  W.  K.,  and  Darnell,  J.  E.:  The  adsorption  and  early  fate  of 

purified  poliovirus  in  HeLa  cells.   Virology,   13:439-447,  1961. 

37.  Salzman,  N.  P.,  Lockart,  R.  Z.,  and  Sebring,  E.  D.:  Alterations  in 

HeLa  cell  metabolism  resulting  from  poliovirus  infection.  Virology. 
P.-244-259,  1959. 

38.  Darnell,  J.  E.,  and  Levintow,  L.:  Poliovirus  protein:  Source  of  amino 

acids  and  time  course  of  synthesis.  J.  Biol.  Chem.,  235:74-77,  1960. 


Virus  Action  and  Replication  1 37 

39.  LwofF,  A.,  Dulfiecco,  R.,  Vogt,  M.,  and  Lwoff,  M.:  Kinetics  of  the 

release  of  poliovirus  from  single  cells.  Virology,  /.•128-139,  1955. 

40.  Howes,  D.  W.:  The  growth  cycle  of  poliovirus  in  cultured  cells  III. 

The  asynchronous  response  of  HeLa  cells  multiply  infected  with 
Type  I  poliovirus.   Virology.  9.'1 10-126,  1959. 

41.  Boyer,  G.  S.,  Leuchtenberger,  C,  and  Ginsberg,  H.  S.:  Cytological 

and  cytochemical  studies  of  HeLa  cells  infected  with  adenovirus. 
J.  Exp.  Med.,  7(95.- 195-2 16,  1957. 

42.  Stoker,  M.  G.  P.,  Smith,  K.  M.,  and  Ross,  R.  W.:  Electron  micro- 

scopic studies  of  HeLa  cells  infected  with  herpes  vn-us.  J.   Gen. 
Microbiol.,   79.-244-249,  1958. 

43.  Salzman,  N.  P.:  The  rate  of  formation  of  vaccinia  DNA  and  vaccinia 

virus.   Virology,  70.T50-152,  1959. 

44.  Ginsberg,   H.   S.,   and  Dixon,   M.   K.:   Deoxyribonucleic   acid   and 

protein  alterations  in  HeLa  cells  infected  with  Type  4  adenovirus. 
J.  E.xp.  Med.,  109:407-422,  1959. 

45.  Green,  M.,  and  Daesch,  G.  E.:  Biochemical  studies  on  adenovirus 

multiplication  H.   Kinetics  of  nucleic  acid  and  protein  synthesis 
in  suspension  cultures.   Virology,  /J.T69-176,  1961. 

46.  Newton,  A.,  and  Stoker,  M.  G.  P.:  Changes  in  nucleic  acid  content 

of  HeLa  cells  infected  with  herpes  virus.  Virology,  5.-549-560,  1958. 

47.  Marcus,   P.    L,   and   Puck,  T.   T.:   Host-cell  interaction  of  animal 

viruses  L  Titration  of  cell-killing  by  viruses.   Virology,  6.- 405  423, 
1958. 

48.  Wheelock,  E.  F.,  and  Tamm,  L:  Mitosis  and  division  in  HeLa  cells 

infected   with  influenza   or  Newcastle  disease   virus.    Virology,   8: 
532-536,  1959. 

49.  Temin,  H.  M.,  and  Rubin,  H.:  Characteristics  of  an  assay  for  Rous 

sarcoma  virus  and  Rous  sarcoma  cells  in  tissue  culture.   Virology, 
6.-669-688,  1958. 

50.  Winocour,  E.,  and  Sachs,  L.:  A  plaque  assay  for  the  polyoma  virus. 

Virology,  <?.-397-400,   1959. 

51.  Dulbecco,  R.  and  Freeman,  G.:  Plaque  production  t^y  the  polyoma 

virus.  Virology,  5.-396-397,  1959. 

52.  Vogt,  M.,  and  Dulbecco,  R.:  Virus-cell  interaction  with  a  tumor- 

producing  virus.  Proc.  Nat.  Acad.  Sci.  USA,  46:365-370,  1960. 

53.  Dulbecco,  R.,  and  Vogt,  M.:  Significance  of  continued  virus  pro- 

duction in  tissue  cultures  rendered  neoplastic  by  polyoma  virus. 
Proc.  Nat.  Acad.  Sa.  USA,  46:\6\7-\623,  1960. 

54.  Vogt,  M.  and  Dulbecco,  R.:   Studies  on  cells  rendered  neoplastic 

by  polyoma  virus.   Virology,  16:41-51,  1962. 


1 38  Information  Storage  and  Neural  Control 

55.  Temin,  H.  M.,  and  Rubin,  H.:  A  kinetic  study  of  infection  of  chick 

embryo  cells  in  vitro  by  Rous  sarcoma  virus.  Virology,  5;209-222, 
1959 

56.  Rubin,  H.,  and  Temin,  H.  M.:  A  radiological  study  of  cell-virus 

interaction  in  the  Rous  sarcoma.   Virology,  7:75-9\,  1959. 

57.  Isaacs,  A.:  The  Nature  and  Function  of  Interferon.  Perspectives  in 

Virology,  New  York,  Wiley,  1960,  vol.  2. 

58.  Rose,  W.  R.,  Hartley,  J.  W.,  Roizman,  B.,  and  Levy,  H.  B.:  Charac- 

terization of  a  factor  formed  in  the  course  of  adenovirus  infection 
of  tissue  cultures  causing  detachment  of  cells  from  glass.  J.  Exp. 
Med.,   108:713-729,  1958. 

59.  Rogers,  S.:  Induction  of  arginase  in  rabbit  epithelium  by  the  Shope 

rabbit  papilloma  virus.  Nature,   75J.T815-1816,  1959. 

60.  Schafer,  VV.:   Some  Observations  Concerning  the  Reproduction  of 

RNA-Containing  Animal  Viruses.  In  Virus  Growth  and  Variation, 
New  York,  Cambridge  University  Press,  1959,  p.  61. 

61.  Zimmerman,  T.,  and  Schafer,  W.:  Effect  of  fluorophenylalanine  on 

fowl  plague  virus  multiplication.   Virology,  7/.-676-698,  1960. 

62.  Saizman,  N.  P.,  and  Sevring,  E.  D.:  The  source  of  poliovirus  ribo- 

nucleic acid.  Virology,  7J.-258-260,  1961. 

63.  Darnell,  J.   E.,   Levintow,   L.,  Thoren,   M.,  and  Hooper,   L.:  The 

time  of  synthesis  of  poliovirus  RNA.    J^irology,   13:271-279,   1961. 

64.  Fogh,  J.,  and  Stuart,  D.  C,  Jr.:  Intracellular  crystals  of  polioviruses 

in  HeLa  cells.  Virology,  77.-308-311,  1960. 

65.  Levintow,   L.,  Thoren,   M.,  Darnell,  J.   E.,   and  Hooper,   L.:   The 

Effect  of  5-fluorophenylalanine  and  Puromycin  on  the  Replication 
of  Poliovirus  (in  press,  1961). 

66.  Ackerman,  W.,  Rabson,  A.,  and  Kurtz,  H.:  Growth  characteristics 

of  poliomyelitis  virus  in  HeLa  cell  cultures:  Lack  of  parallelism  in 
cellular  injury  and  virus  increase.  J.  Exp.  Med.,  7(9O.-437-450,  1954. 

67.  Cohen,  G.  N.,  and  Munier,  R.:  Effects  des  Analogues  structuraux 

d'amino  acides  sur  la  crossance,  la  synthese  des  proteines  et  la 
synthese  d'enzymes  chez  E.  coli.  Biochim.  Biophys.  Acta.,  31:347- 
356,  1959. 

68.  Nathans,  D.,  and  Lipmann,  F.:  Amino  acid  transfer  from  amino- 

acyl-ribonucleic  acids  to  protein  on  ribosomes  of  E.  coli.  Proc. 
Nat.  Acad.  Sa.  USA,  45.T721-1729,  1961. 

69.  Yarmolinsky,  M.  B.,  and  De  La  Haba,  G.  L.:  Inhibition  by  puro- 

mycin of  amino  acid  incorporation  into  protein.  Proc.  Nat.  Acad. 
Sci.  USA,  45:1721-1729,  1959. 


Virus  Action  and  Replication  1  39 

70.  Nirenberg,  M.,  and  Matthaei,  J.  H.:  The  dependence  of  cell-free 
protein  synthesis  in  E.  coli  upon  naturally  occurring  or  synthetic 
polyriljonucleotides.  Proc.  Nat.  Acad.  Sci.  USA,  47. ■  \  588 -\  602,  1961. 

DISCUSSION  OF  CHAPTER  VI 

Saul  Kit  (Houston,  Texas):  I  sincerely  congratulate  Dr.  Darnell 
on  his  very  interesting  and  informative  paper.  He  may  have 
already  settled  this  point,  but  it  is  one  that  is  extremely  important 
to  us.  It  has  been  postulated  that  the  messenger-RNA  is  destroyed 
at  the  time  the  proteins  are  synthesized  and  that,  as  a  result  of  this, 
there  is  a  need  for  continual  synthesis  of  messenger-RNA.  Dr. 
Darnell's  system  would  seem  to  be  an  excellent  one  for  testing 
this  possibility,  since  after  synthesis  of  the  polio  proteins  in  the 
reconstructed  ribosome  system,  one  could  re-extract  the  polio  RNA 
and  test  for  infectivity.  I  wonder  whether  he  has  already  done  this. 

James  E.  Darnell,  Jr.  (Cambridge,  Massachusetts):  It  turns 
out  not  to  be  necessary  to  try  to  extract  the  polio  RNA  and  to 
look  for  any  infectivity.  Rather,  what  happens  is  this:  The  Niren- 
berg system  is  made  from  E.  coli,  and  E.  coli  ribosomes  contain  quite  a 
large  amount  of  ribonuclease.  It  has  been  shown  in  Watson's 
laboratory  (J.  D.  Watson,  personal  communication)  that  natural 
messenger-RNA  from  E.  coli,  which  can  be  attached  in  vitro  to  the 
ribosomes,  is  broken  down  completely  during  protein  synthesis. 
This  is  the  case  with  polio  also.  However,  we  cannot  say  definitely 
that  this  is  linked  to  protein  synthesis  because  the  breakdown  goes 
on  if  the  system  is  incubated  in  the  absence  of  an  ATP  generating" 
system.  One  possible  way  to  get  around  this  is  to  use  ribosomes 
from  cells  which  do  not  contain  ribonuclease,  such  as  Bacillus 
megaterium.  An  additional  point  of  interest  in  this  general  area 
is,  of  course,  whether  messenger-RNA  from  animal  cells  is  handled 
difTerently  from  bacterial  messenger-RNx'\. 

Kit:  Are  the  E.  coli  ribosomes  not  stabilized  sufficiently  by  the 
amount  of  magnesium  used  in  your  medium  so  that  there  is  no 
release  of  ribonuclease  activity? 

Darnell:  Not  completely.  Incubation  of  polio  RNA  in  the 
synthesizing  system  without  an  energy  source,  and,  therefore,  with 
no  resulting  protein  synthesis,  still  results  in  degradation  of  the 
viral  RNA. 


CHAPTER 
VII 

THE  INFORMATION  CONCEPT  IN  ECOLOGY: 
SOME  ASPECTS  OF  INFORMATION -GATHER- 
ING BEHAVIOR  IN  PLANKTON** 

Bernard  C.  Patten,  Ph.D. 


T. 


INTRODUCTION 


.HE  subject  of  community  energetics  concerns  the  processes 
by  which  ecological  coinmunities  achieve  a  favorable  balance  be- 
tween energy  gains  and  losses.  Referred  to  as  the  study  of  pro- 
ductivity, or  trophodynamics,  this  is  one  of  the  most  active  areas 
of  investigation  in  modern  ecology.  A  central  concept  in  this  work 
is  that  of  the  food  chain  (1),  or  food  web  (2),  beginning  with 
photosynthesizing  plants  and  proceeding  through  various  trophic 
levels  toward  a  terminal  consumer,  or  consumers.  Such  a  network 
is  illustrated  in  Figure  1,  where  the  Si  represent  "species":  Si 
being  the  sun,  So  and  Ss  producers  (plants),  ^'4-^7  primary  con- 
sumers (herbivores),  and  .^7-^^11  secondary  consumers  (carnivores). 
Note  that  ^^7  is  an  omnivore  since  it  consumes  both  plant  and 
animal  material.  For  simplicity,  decomposers  are  not  shown.  The  qj 
denote  pathways  of  energy  flux:  for  example  qs  signifies  that  energy 
is  gained  by  Si  through  eating  S2. 

In  the  mid-1 940's  when  energy  ecologists  were  involved  in 
working  out  the  intricacies  of  food  chain  relationships  on  a  who- 
eats-whom-and-how-much  basis,  a  physicist,  E.  Schrodinger,  as- 
serted  that  living  organisms  feed  upon  "negative  entropy"    (3). 


*Contribution  No.  120  from  the  Virginia  Institute  of  Marine  Science. 

f  The  following  biochemical  abbreviations  are  used  in  this  paper. 
ADP  =  adenosine  diphosphate;  ATP  =  adenosine  triphosphate;  DPN  =  diphospho- 
pyridine  nucleotide;  FMN  =  flavin  mononucleotide;  PN  =  pyridine  nucleotide;  PNH2 
=  reduced  pyridine  nucleotide;  TPN  =  triphosphopyridine  nucleotide. 

140 


Information  Concept  in  Ecology 


141 


Since  everyone  in  ecology  understood  that  organisms  ate  food  and 
that  they  did  so  for  the  energy  it  contained,  Schrodinger's  sug- 
gestion was  ignored.  In  1949,  however,  another  physicist,  Brillouin 
(4),  elaborated  further  in  a  manner  more  harmonious  with  the 
climate  of  trophodynamic  thought:  "The  earth  ...  is  constantly 
receiving  energy  and  negative  entropy  from  outside  .  .  .  life  feeds 
on  high  grade  energy  or  'negative  entropy'  ...  all  experimental 
measures  show  that  the  entropy  of  the  refuse  is  larger  than  that 
of  the  food."  This  established  a  definite  connection  with  ecological 
thinking,  and  further  rapport  developed  in  1953  with  the  appear- 
ance of  papers  by  Branson  (5)  and  Linschitz  (6)  which  amplified 
the  theme  along  thermochemical  lines.  This  work  was  soon  fol- 
lowed by  several  publications  in  ecological  journals  which  at- 
tempted to  relate  information  theory  to  community  trophociy- 
namics  (7,  8). 

In  the  present  paper  some  aspects  of  the  organization  and 
behavior  of  plankton  communities  are  discussed  in  an  informational 
and  trophodynamic  context.  In  the  development,  such  communities 
will  come  to  take  the  logical  form  of  a  rational  utility-seeker  (9), 


'7  "'8  ~9  ^10  ^11 

Fig.  1.  Schematic  diagram  of  a  hypothetical  food  chain. 


142  hiformation  Storage  and  Neural  Control 

where  utility  is  understood  to  have  a  hedonistic  value  in  either 
1)  reducing"  the  community's  uncertainty  about  nature,  or  2)  per- 
mitting purchase  of  a  measure  of  certainty.  In  the  latter  connection, 
energy  will  be  identified  as  a  universal  currency.  We  begin  with 
a  result  from  communication  theory. 

SHANNON'S  THEOREM  10 

Consider,  following  Shannon  (10),  a  discrete  communication 
channel  fed  by  an  information  source.  If  H{x)  is  the  input  entropy 
and  H{y)  tnat  of  the  output,  H{x,j)  is  the  joint  entropy  of  input 
and  output,  and  H{y\x)  and  H{x\y)  are  conditional  entropies,  then 

[1]  H{x,y)  =  H{x)  +  H{,j\x)  =  H{y)  +  H{x\ij). 

For  such  a  system  SJiannon  proved  his  Theorem  10:  If  a  correction 
channel  has  a  capacity  H{x\y),  correction  data  can  be  encoded  in 
such  a  manner  that  all  but  an  arbitrarily  small  fraction  of  errors 
induced  by  noise  can  be  corrected.  This  is  not  possible  if  the 
channel  capacity  is  less  than  //(.vjj),  which  represents  the  amount 
of  information  which  must  be  supplied  to  correct  the  message. 

This  theorem  has  been  exploited  by  Ashby  (11,  12)  as  the  basis 
for  a  cybernetic  theory  of  biological  homeostasis.  Roughly,  the 
organism  is  regarded  as  bomJDarded  by  information  from  an 
environment  which  tends  to  drive  the  organism  into  states  outside 
the  limits  which  permit  survival.  To  achieve  stability,  therefore, 
it  becomes  necessary  in  light  of  Shannon's  theorem  for  the  organism 
to  provide  information  to  a  "regulator"  (analogue  of  correction 
channel)  in  amounts  at  least  as  great  as  the  disturbances.  Ashby 
calls  this  the  law  of  requisite  variety.  It  implies  that  an  organism 
must  continually  be  concerned  with  having  sufficient  information 
available  (accumulated  against  the  gradient  imposed  by  the 
second  law  of  thermodynamics)  to  meet  particular  environmental 
threats.  The  compatibility  of  this  theory  with  the  Schrodinger- 
Brillouin  thesis  is  evident.  Also  apparent  is  the  fact  that  its  basic 
applicability  is  unaltered  by  a  conversion  from  the  scale  of  organism 
to  that  of  ecological  community:  both  units  metabolize,  both  are 
subject  to  the  same  "heat  death"  and,  consequently,  both  have 
similar  problems  of  homeostasis. 


Information  Concept  in  Ecology  143 

COMMUNITY  STABILITY 

It  is  almost  axiomatic  in  ecology  that  structurally  complex 
communities  are  intrinsically  more  stable  than  simpler  ones.  The 
standard  illustration  is  to  contrast  the  highly  stable  biota  of 
tropical  rain  forests  with  the  comparatively  unstable  assemblages 
of  the  Arctic  tundra.  Community  stability  and  complexity  can  be 
related  in  the  following  manner. 

At  any  specified  stage  in  its  development,  a  community  con- 
tains m  species,  s,,  of  plants  and  animals  with  frequencies  A'',-  such 
that 

[21  Y.Ni  =  N.  {i=\,2,...,m) 

1=1 

The  uncertainty  per  individual   of  selecting  the  i  '  species  is 

where  P  is  probability.  The  total  uncertainty,  N  <  D  >  ,  referred 
to  as  community  diversity,  is 

f4]  D  =  -XlAMogPCs,). 

(=1 

Although  the  diversity  problem  is  not  of  specific  concern  here,  it 
should  be  mentioned  that  there  has  been  considerable  develop- 
ment of  this  subject  along  informational  lines  (13,  14,  15),  repre- 
senting the  most  extensive  application  of  information  theory  which 
has  so  far  been  made  to  an  ecological  problem. 

MacArthur  (16)  has  proposed  that  community  stability  be 
equated  to  the  complexity  of  the  food  web  as  given  by  an  entr'opy 
measure: 

[5J  S=  -Z^^(7.)logP(r/,), 

where  S  is  stability  and  Piqj)  the  probability  of  energy  traversing 
a  particular  path  qj.  The  rationale  of  this  suggestion  is  that  removal 
of  a  species  and  consequent  destruction  of  the  pathways  leading 
to  and  from  it  would  be  less  disruptive  to  a  community  with  a 
high  value  for  S  than  to  one  with  a  lower  value.  Since  the  P{qj)'s 
are  obviously  functions  of  the  A^,'s,  it  follows  that  stability  and 
diversity  are  related,  the  more  diverse  systems  being  the  more 
stable.  Since  greater  stability  implies  greater  success  in  meeting" 


144  Information  Storage  and  Neural  Control 

the  imperative  of  Shannon's  Theorem  10,  and  since  stabihty  is  a 
function  of  compositional  complexity,  it  follows  that  a  natural 
tendency  of  ecological  communities  should  be  to  develop  to  maxi- 
mum proportions  within  the  limitations  imposed  by  particular 
environments.  This  conclusion  is  consistent  with  empirical  ob- 
servations. 

One  measure  of  the  extent  to  which  a  given  community  has 
expanded  to  fill  a  physical  space  is  the  total  quantity  of  organic 
matter  contained  in  that  space.  This  variable  will  be  referred  to 
here  as  the  community's  biomass.  Because  community  ontogeny 
(ecological  succession)  proceeds  by  means  of  niche  (17)  prolifera- 
tion (more  species  make  more  species  possible),  a  reasonable  way 
to  assess,  in  a  quantitative  sense,  the  extent  of  organization  of  a 
community  might  be  to  oxidize  a  suitable  sample  in  a  calorimeter 
and  to  equate  heat  evolution  with  intrinsic  complexity.  Though 
admittedly  crude,  such  an  approach  would  not  be  entirely  without 
basis  since  all  information,  even  that  which  is  abstract,  is  under- 
stood to  be  physically  based  and  is  therefore  referable  to  thermo- 
dynamic negative  entropy  (18,  19).  This  broaches  the  problem  of 
the  relationship  between  information  and  energy — the  reason  why 
information  theory  is  of  interest  to  energy  ecologists. 

ENERGY  AS  CURRENCY 

The  connection  between  energy  and  information  has  been 
well  established  in  the  context  of  macroscopic  thermodynamics 
(19)  where  adiabatically  accessible  system  states  are  generally 
regarded  as  informationally  equivalent,  while  those  attainable 
only  non-adiabatically  are  not  (20).  In  the  usual  Boltzmann- 
Gibbs  treatments,  the  role  of  matter  in  determining  a  system's 
entropy  is  obscure;  however,  the  recently  introduced  formalism 
of  Jaynes  (21)  and  Tribus  (22)  offers  considerable  clarification, 
as  follows. 

Consider  a  system  of  /?«,  rib,  ■  ■  ■  particles  of  matter  of  kinds 
a,  b,  .  .  .  in  a  phase  space  with  coordinates  Xi,  X2,  ....  When  the 
coordinates  are  prescribed  and  the  number  of  particles  known, 
the  system  consists  of  a  finite  number  of  discrete  quantum  states, 
J,  with  energies  e^.- 


Injormation  Concept  in  Ecology  145 

|6]  ey  =  eO';  Tia,  Ub,  .  .  .;Xi,  X2,  .  .  .). 

If/?,  is  the  probability  for  a  particular  small  subsystem  to  be  in 
state  I,  then, 

[7]  Z  V,  =  1 

i 
[8]  11V^^^    =     <f> 

[y]  Y^V^na.i  =  <>L>,  (a,l>,c,  .  .  .) 

and 

[10]  *S:=  -kY^P^np,, 

where  S  is  the  entropy  of  the  system,  and  <  >  denotes  expected 
values.  The  maximum  uncertainty  of  selecting"  a  subsystem  in  state 
I  is  obtained  when  the  p^s  are  all  ecjual  (10).  Maximizing   [10] 

[11]  "T=?  ^^''^'^  l)dp,  =  0; 

difTerentiating  [7],  [8]  and  [9]  and  introducing  the  undetermined 
Lagrangian  multipliers  ««,  «&,  •   .   ■  ,  1^,  ^a,  we  obtain 

[12]  (Qo-  l)T.dp.  =  0 

[13]  (3Y.e.dp,=0 

[14]  a^X)  na.,dpi  =  0.  {a,h,c,  .  .  .) 

Adding"  [11-14]  and  collecting"  terms: 

[15]         XI  Oil  P'  +  ^^0  -\-  ^U  +  a,,  Ha.,  +  abni,,  +  .  .  .)  fZ/J/  =  0, 

from  which 

[16]  p,  =  exp  (  — Qo  —  (Se,-  —  aa  nn.i  —  ab  Hb.i  —...). 

Thus,  a  system's  entropy  is  maximal  when  the  e,'s,  na,i's,  «6, /s, 
etc.,  of  all  its  subsystems  are  identical,  signifying  a  homogeneous 
distribution  of  matter  and  energy  throughout  the  system.  It  can 
be  shown  (22)  that  the  rate  of  entropy  change  is 


146  Information  Storage  and  Neural  Control 

dS  =  k  (J3d  <e>  +  aa  d  <Ha>  +  "&  d  <nb>  +  .  .  . 

If  two  systems  with  different  values  of  /3,  «„,  a^,  .  .  . ,  and  witli 
total  energy  and  matter  constant  between  them,  are  allowed  to 
interact  irreversibly,  then  the  energy  gain  of  one  must  be  equiva- 
lent to  the  loss  of  the  other,  and  the  gain  in  n,j,  n,^^  .  .  .  by  one 
corresponds  to  that  lost  by  the  other.  In  the  language  of  game 
theory  (23)  such  a  relationship  is  zero-sum.  \idXi  =  dX-i  =  .  .  .  =0, 
then  the  connected  system's  entropy  change  is,  from  [17], 

[18]  dS  -  k  [(j3  -  13')  d  <€>  +  {aa  -  aa)  d  <Ha> 

+  (ab  —  ab')  d  <nb>  +  .  .  .] 

Hence,  it  is  possible  for  one  of  the  systems  (call  it  community) 
to  decrease  its  entropy  at  the  expense  of  the  other  (environment) 
since  the  only  requirenient  is  that  dS  >  0  overall. 

Details  of  energy-matter  exchange  between  such  systems  are 
very  complex  because  the  parameters  3,  aa,  at,  ■  .  .  may  become 
reciprocally  coupled  within  a  system 

d<e>  _  _    a"Qo 

[19]  )  {a,b,c,  .  .  .) 


d<na>   _ 

6/3        ~   '     daa  dl3l 
SO  that 

[20]         d<e>   =   -^  dl3  -  -^  daa  (a,b,C,  ...) 


and 


d^-     "        dl3  daa 


d''QiO     ,,,        3"Qo 


[21]         d<na>   =  -  T ~dl3  -  ~r—;daa  {a,h,c,  .  .  .). 

daa  dp  daa" 

The  important  point  to  distinguish  for  our  present  purpose  is  that 
for  one  system  to  diminish  its  entropy  with  respect  to  another 
with  which  it  is  coupled  in  communication,  it  must  establish  and 
maintain  physical  barriers  to  the  free  exchange  of  energy  and 
matter.  In  short,  it  must  proliferate  structural  heterogeneity  by 
maximizing  the  inequality  of  the /^,'s  in  [10]. 

At  the  community  level,  such  heterogeneities  are  maintained 
by  a  graded  series  of  discrete,  functional  barriers  ranging  from,  at 


Information  Concept  in  Ecology  147 

the  lower  end  of  the  scale,  quantum  states,  atoms,  molecules, 
membranes,  cells  and  their  ultrastructural  components,  tissues  and 
organs,  to,  at  the  upper  end,  individual  organisms,  species,  popu- 
lations, multi-specific  evolutionary  units  (supraorganisms)  (24) 
and  finally  the  community  itself.  The  construction,  maintenance 
and  operation  of  such  barriers  (with  all  the  morphology  and 
physiology  that  this  implies)  are  achieved  by  physical  and  chem- 
ical processes  which,  in  net,  are  endergonic.  Without,  therefore, 
a  continuous  input  of  energy,  the  barricades  would  fail  to  function 
and  would  ultimately  be  eroded  away,  with  the  result  that  com- 
munity and  environment  would  become  one. 

This  is  a  trivial  conclusion,  of  course.  After  all,  it  is  one  of  the  most 
obvious  statements  which  could  be  made  regarding  bio-systems. 
Yet  its  articulation  seeins  necessary  to  provide  a  basis  for  the  follow- 
ing restatement  of  the  Schrodinger-Brillouin  proposition:  Energy 
may  be  regarded  as  a  universal  currency  with  which  organisms  pur- 
chase utility,  as  negative  entropy,  from  the  environment. 

COMMUNITY  BIOENERGETICS 

In  view  of  this  proposition,  the  ultimate  source  of  negativt 
entropy  to  an  ecological  community  may  be  regarded  to  be 
photons.  When  a  photon  strikes  an  atom  an  electron  is  lifted  from 
ground  state  to  a  higher  empty  orbital  (vertical  arrow  ^~  -^^*  in 
Fig.  2).  For  most  molecules  excited  electrons  usually  drop  back 
to  ground  state  immediately,  dissipating  the  excess  energy  as 
electromagnetic  radiation  (broken  arrow,  Fig.  2).  Living  systems 
to  paraphrase  Szent-Gyorgi  (25),  have  shoved  themselves  between 
these  two  processes  by  shunting  the  excited  electrons  into  different 
downhill  pathways  in  which  their  energy  can  be  released  slowly 
and  put  to  useful  work.  The  first  step  in  the  process  is  excitation 
(by  photons,  or  indirectly  via  accessory  plant  pigments)  of  pi 
electrons  in  the  conjugated  portion  of  chlorophyll  a.  In  cyclic 
photosynthetic  phosphorylation  (26)  the  chlorophyll  provides  these 
electrons  directly,  thereby  acting  both  as  electron  donor  and 
acceptor  (Fig.  2).  In  noncyclic  photophosphorylation  the  electrons 
come  from  H2O,  which  the  excitation  energy  decomposes  to 
oxygen,  freed  as  O2,  and  H  atoms.  The  hydrogen  electrons  sub- 


148 


Information  Storage  and  Neural  Control 


Fig.  2.  Schematic  diagram  of  the  energy  cycle  of  an  ecosystem,  modified    and 

expanded  after  Szent-Gyorgi  (25)  and  Arnon  (26).  Anaerobic  and    chemosyn- 

thetic  processes  are  not  indicated. 


sequently  reduce  one  of  two  pyridine  nucleotides  {PN  -^  PNHi, 
Fig.  2).  Concurrently,  ATP  is  synthesized,  incorporating  into  its 
terminal  "high  energy"  phosphate  bond  some  of  the  original 
photon  energy.  Neither  ATP,  DPN,  nor  TPN  is  stable  enough 
to  function  in  energy  storage.  This  is  accomplished  by  reducing 
CO 2  to  carbohydrates  and  water,  then  to  lipids  (Fig.  2). 

Energy  so  stored  may  be  utiHzed  directly  by  the  primary  pro- 
ducer, or  it  may  be  transmitted  to  other  organisms  in  the  food 
chain.  The  retrieval  of  energy  from  storage  is  accomplished  by 
transferring  electrons  (in  H  atoms)  to  PN,  releasing  the  carbon 
as  CO  2.  The  PNHi  then  transfers  electrons  to  flavin  mononucleo- 
tide (FMN),  whence  they  cascade  down  the  oxidative  chain  of 
cytochromes,  generating  heat  at  every  step.  Most  of  the  energy 
remaining  is  converted  (in  oxidative  phosphorylation)  to  .4  TP, 
in  which  form  it  is  available  for  the  performance  of  cellular  work. 
Finally,  the  electrons  are  transferred  to  Oo  which  then  binds 
protons  to  form  HoO.  Water  represents  ground  state,  where  the 
cycle  e~ —>  e* -^  e^  is  completed.  If,  at  a  specified  time,  the 
system  contains  more  free  energy  than  it  did  at  a  prior  time,  we 
say  that  a  favorable  balance  between  inputs  and  expenditures  has 


Injormation  Concept  in  Ecology  149 

been  achieved.  This  enables  the  system  to  maintain  or  further 
diminish  its  entropy.  If  the  system  possesses  less  free  energy  after 
a  passage  of  time,  we  say  the  balance  is  unfavorable  and  the  system 
is  less  able  to  forestall  an  entropy  gain.  These  relationships  can 
be  summarized  by  a  simple  transfer  function  which  will  be  termed 
cost.  This  variable  represents  the  amount  of  energy  which  must 
be  expended  to  gain  a  unit  of  energy  from  the  environment: 

P7r~    <  1  (biomass  gain) 

[22]  piv~^  =  1  (steady  state) 

P7r~^  >  1  (biomass  loss) 

where  tt  denotes  total  energy  gained  by  the  community  and  p 
represents  total  energy  lost. 

Let  us  now  consider  some  specific  behavioral  and  organiza- 
tional attributes  of  planktonic  systems  which  exemplify  goal- 
adaptability,  the  goal  being  biomass  maximization. 

PROCEDURES 

The  plankton  communities  under  consideration  occupied  the 
York  River,  Virginia,  during  the  summer  of  1960.  For  ten  con- 
secutive weeks,  from  June  23  to  August  25,  in  situ  dark  and  light 
bottle  differential  oxygen  studies  (27)  were  performed  weekly  to 
assess  energy  flux  through  the  community.  The  sampling  station 
was  located  about  300  yds.  off  the  end  of  the  Virginia  Institute  of 
Marine  Science  pier  where  the  approximate  depth  of  mean  low 
water  was  thirty  feet. 

Hydrographic  determinations  included  vertical  profiles  of  chlor- 
inity,  temperature,  dissolved  oxygen,  total  dissolved  phosphorus 
and  total  nitrate.  Temperature  was  recorded  with  a  thermistor 
unit.  Chlorinity  was  titrated  with  silver  nitrate.  Dissolved  oxygen 
was  measured  by  the  unmodified  Winkler  method.  Dissolved 
organic  and  inorganic  phosphorus  were  obtained  as  follows: 
Fractionation  into  dissolved  and  adsorbed  inorganic  and  dissolved 
and  particulate  organic  components  was  achieved  by  Millipore 
(type  HA)  filtration.  Inorganic  fractions  were  assayed  directly; 
organic  fractions  were  obtained  by  digesting  samples  for  twelve 
hours  at  20  psi;  the  molybdate  method  (corrected  for  salt  interfer- 
ence) was  employed  to  estimate  the  orthophosphate  in  both  cases. 


1 50  Information  Storage  and  Neural  Control 

For  energy  flux  determinations,  paired  dark  and  liglit  bottles 
containing  water  samples  from  two,  six  and  ten  feet  were  sus- 
pended at  various  depths  in  the  water  column  for  twenty-four 
hours  (beginning  0730  EST),  and  then  fixed  for  Winkler  titration. 
The  suspension  depths  included  all  combinations  of  the  collection 
depths:  (2,2),  (2,6),  (2,10);  (6,2),  (6,6),  (6,10);  (l0,2),  (l0,6), 
(10,10),  where  the  left  member  of  each  pair  designates  collection 
depth  and  the  right  member  suspension  depth.  Additional  dark 
bottles  for  (l4,14)  and  (l8,18)  were  also  included.  Production 
variables  were  detei  mined  from  the  initial  and  final  dissolved 
oxygen  concentrations  in  the  bottles: 

TT  =  I  —  d  (photosynthesis) 

[23]  p  =  i  —  d  (respiration) 

IT  —  ,0  =  I  —  i  (net  production) 

where  /,  d  and  /  are,  respectively,  light  bottle,  dark  bottle,  and 
initial  oxygen  concentrations.  The  differential  oxygen  concen- 
trations were  converted  to  gram  calories  (gcal)  using  suitable 
conversion  factors  derived  from  the  stoichiometry  of  the  photo- 
synthesis and  respiration  reactions. 

Incident  solar  radiation  at  the  water  surface  was  measured  in 
gcal  cm~"  by  an  Eppley  10-junction  pyrheliometer  installed  a  few 
hundred  yards  from  the  station,  the  output  of  the  thermopile 
being  electronically  integrated  and  automatically  printed-out 
every  thirty  minutes.  Extinction  coefficients  for  "white"  light 
were  detei  mined  on  samples  obtained  from  the  various  depths 
at  the  beginning  and  end  of  each  experiment.  The  optical  densities 
were  measured  with  a  Klett-Summerson  colorimeter,  using  a 
neutral  filter.  From  these  values  a  mean  was  obtained  for  the 
upper  ten  feet  and  was  employed  to  estimate  the  light  intensity 
at  any  depth. 

Total  chlorophyll  was  assayed  by  Millipore-filtering  samples 
from  different  depths,  grinding  filters  and  residues  with  sand  and 
extracting  the  pigment  in  90  per  cent  acetone  (A/^COs-saturated), 
then  Seitz-filtering  to  remove  sand  and  undissolved  millipore  frag- 
ments. Absorbancies  were  determined  with  a  red  filter,  and  were 
converted  to  chlorophyll  concentrations  by  comparison  with  a 
standard  curve  prepared  from  chlorophyll  a. 


Information  Concept  in  Ecology 


151 


Biomass  was  estimated  as  ash-free  dry  weight  of  suspended 
soHds.  The  inethod  involved  filtering  water  through  tared  Millipore 
filters  (type  HA),  desiccating"  filters  plus  residues,  weighing  for 
total  solids,  then  ashing  at  600 °C.,  rehydrating  the  ash,  desic- 
cating, and  weighing  again  to  obtain  the  ash  weight. 

Counts  of  phytoplankton  units  (chains,  colonies  or  individual 
cells)  were  made  from  Sedgwick-Rafter  mounts  of  fresh  samples 
obtained  from  the  various  depths.  All  flagellates  and  diatoms  were 
counted;  ciliates  and  other  animals,  when  present,  were  excluded. 

COMMUNITY  ADAPTATIONS  FOR  MAXIMUM  BIOMASS 

The  York  River  water  column  at  the  station  sampled  was 
comparatively  unstratified  from  surface  to  bottom  during  the 
summer  of  1960.  This  is  illustrated  by  the  graphs  in  Figure  3. 


MAN  DIUOLVIO  OXTMII 


1 


MEM   eitMLVto  rHOtrORU* 


MCAN    NITRATE 


Fig.  3.  Mean  vertical  distribution  of  chlorinity,  temperature,  and  the  dissolved 
substances  oxygen,  phosphorus  and  nitrate. 


152 


Information  Storage  and  Neural  Control 


Chlorinity  varied  from  8.54  to  12.60  parts  per  thousand,  with  the 
surface  water  generally  a  little  less  saline  than  that  near  the 
bottom.  The  mean  gradient  for  the  ten  experiments  was  only 
0.75  parts  per  thousand.  Temperature  ranged  from  24°C.  in  June 
to  over  27  °C.  in  mid-August,  the  surface  waters  being  somewhat 
warmer  than  the  lower  strata.  The  mean  temperature  gradient 
was  but  0.29  °C.  These  two  variables,  clilorinity  and  temperature, 
are  determinants  of  water  density.  In  this  case,  they  indicate  a 
very  small  gradient  of  increasing  density  with  depth,  thus  assuring 
a  fair  amount  of  vertical  mixing  in  the  water  column.  This  con- 
clusion is  underscored  by  the  vertical  distribution  patterns  of 
other  dissolved  substances  for  which  data  were  obtained — dissolved 


SURFACE 


2  - 


6- 


DEPTH      10  _ 
(ft) 


14- 


18- 


BOTTOM 


4000 


6000 


MEAN    NUMBER  OF  CELLS 
(  per  ml  ) 

Fig.   4.   Mean  concentrations  of  living  phytoplankters,   in  counting  units  ml, 

at  various  depths. 


Information  Concept  in  Ecology 


153 


oxygen  diminished  only  slightly  with  depth,  and  dissolved  phos- 
phorus and  nitrate  increased  slightly  (Fig.  3). 

In  sharp  contrast  to  these  physical  relationships,  the  living 
organisms  of  the  phytoplankton  were  markedly  stratified  in  the 
upper  layers  (Fig.  4).  To  account  for  this  we  note  that  the  domi- 
nant organisms  of  the  summer  flora  were  motile  flagellates  closely 
related  to  forms  which  are  known  to  be  positively  phototactic. 
Thus,  swimming  is  the  probable  primary  mechanism  involved. 
Other  factors  of  possible  influence  include  rapid  cell  division  in 
the  lighted  surface  layers,  and  manufacture  of  low  specific  gravity 
(lipid)  storage  products. 

The  mean  daily  vertical  distribution  of  light  in  the  ten  experi- 
ments is  graphed  in  Figure  5,  and  shows  typical  exponential 
extinction  with  virtually  complete  absence  of  light  at  the  bottom. 


SURFACE 


DEPTH 
(ft) 


10  - 


MEAN   EXTINCTION 

COEFFICIENT  =  0.97 


BOTTOM 

I  I  I  I 

0  200  400  600 

MEAN   SUBMARINE   ILLUMINATION 
(   gcal  cm-2day-'  ) 
Fig.  5.  Mean  vertical  distribution  of  light  in  the  ten  experiments. 


154 


Information  Storage  and  Neural  Control 


PHOTOSYNTHESIS 
(    gcol  cm-^doy-' ) 


DEPTH 
(ft) 


RESPIRATION 
(    jcQl  cm-*do>-'  ) 


Fig.  6.  Mean  photosynthesis  and  mean  respiration  in  the  water  column. 

Mean  photosynthesis  and  respiration  are  depicted  in  Figure  6. 
As  expected,  production  was  highest  near  the  surface  and  atten- 
uated in  exponential  fashion  with  depth.  Respiration  was  about 
equal  throughout  the  upper  ten  feet,  but  was  only  half  as  great 
below  this  level. 


MEAN    ASH-FREE    SOLIDS  ("BIOMASS") 
(  mg  cm-2  ) 


DEPTH   ,0- 
(ft) 


MEAN  TOTAL  CHLOROPHYLL 
(  ug  cm-2  J 


Fig.  7.  Mean  ash-free  solids  and  mean  total  chlorophyll  concentration  at  various 

depths. 


Information  Concept  in  Ecology  155 

The  concentration  of  ash-free  solids  (Fig.  7)  was  observed  to 
increase  markedly  with  depth.  This  variable  may  be  equated  to 
community  biomass  since  even  the  non-living  detrital  material 
which  it  includes  represents  a  source  of  energy  to  certain  hetero- 
trophic components  of  the  living  plankton.  The  inverse  relationship 
between  the  vertical  distribution  of  these  ash-free  solids  and  that 
of  living  cells  is  a  consequence  of  the  detrital  rain  from  the  zone 
of  production  at  the  top  of  the  water  column,  and  also  of  the 
upwelling  of  bottom  materials.  Since  even  dead  organic  material 
of  this  type  has  an  oxygen  demand,  a  significant  (though  un- 
specifiable)  fraction  of  what  was  represented  in  Figure  6  as  com- 
munity "respiration"  is  a  product  of  non-biological  oxidations 
attending  decomposition.  Since  such  oxidations  cost  the  com- 
munity biomass  energy,  it  is  proper  that  they  be  included  in 
determinations  of  energy  loss. 

As  in  the  case  of  ash-free  seston,  the  vertical  distribution  of 
total  chlorophyll  was  different  from  that  which  would  be  antici- 
pated on  the  basis  of  the  cell-count  data  (Fig.  7).  Chlorophyll 
concentration  increased  gradually  with  depth.  The  explanation 
is  that  large  quantities  of  chlorophyll- and  its  degradation  products 
(many  of  which  would  be  included  in  this  assay)  are  associated 
with  non-living  detritus  (28,  29)  and  sediments  (30,  31,  32).  The 
two  curves  of  Figure  7  strengthen  the  conclusion  that  we  are 
dealing  with  a  fairly  well-mixed  water  mass  since  a  certain  amount 
of  upwelling  is  indicated. 

Let  us  now  consider  some  of  the  photosynthetic  characteristics 
of  the  plankton  community  at  various  depths. 

Recall  from  the  description  of  procedures  that  water  samples 
for  the  measurement  of  photosynthesis  were  collected  at  depths 
of  2,  6  and  10  ft.  and  were  resuspended  so  that  data  for  all  com- 
binations of  collection  and  suspension  depths  could  be  obtained. 
The  graphs  in  Figure  6  are  for  results  from  the  particular  com- 
binations (2,2),  (6,6)  and  (lO,10).  In  Figure  8,  ail  of  the  combina- 
tions are  graphed  in  3-space  with  coordinates  (collection  depth,  sus- 
pension depth,  mean  photosynthesis).  The  surface  depicted  is 
concave  upward,  slopes  downward  toward  the  viewer,  and  curves 
markedly  upward  on  the  left.  Consider,  first,  photosynthesis  as 
a  function  of  suspension  depth,  by  looking  at  the  surface  from  back 


156  Information  Storage  and  Neural  Control 

to  front.  The  downward  sloping  represents  the  attenuation  of 
photosynthesis  with  increased  depth  of  suspension  due  to  decreased 
illumination.  The  curve  of  Figure  6  is  the  locus  obtained  on  this 
surface  by  connecting  the  points  (2,2,4.61),  (6,6,1.46),  and  (lO,10, 
0.85).  Now  view  the  surface  from  left  to  right.  This  gives  photosyn- 
thesis as  a  function  of  collection  depth.  Regardless  of  the  depth  of 
suspension,  the  populations  collected  at  2  ft.  always  photosyn- 
thesized  more  than  those  obtained  from  6  and  10  ft.;  the  latter 
samples  appear  to  give  very  similar  results.  These  relationships 
indicate  that  the  organisms  taken  from  deeper  layers  of  the  water 
column  have  less  capacity  for  photosynthesis  than  those  which 
normally  occupy  the  surface  waters.  This  may  be  a  reflection  of 
the  fact  that  the  deeper  plankters  are  senescent  and  sinking; 
microscopic  examination  usually  revealed  the  surface  organisms 
to  be  far  more  active  in  swimming  than  their  counterparts  from 
below. 

Consider  now  the  thermodynamic  efficiency  of  photosynthesis 
as  reflected  by  the  ratio  of  mean  photosynthesis  per  mean  illumi- 
nance at  each  depth.  These  data  are  presented  in  Figure  9.  The 
surface  generated  is  concave  "upward,  slants  upward  approaching 
the  viewer  and  toward  the  left,  and  rises  sharply  on  the  right  in 
front.  Studying  from  back  to  front  first,  we  observe  that  photo- 
synthetic  efficiency  increases  with  depth  of  suspension,  hence  with 
diminished  light  intensity.  This  result  is  in  accord  with  the  photo- 
synthesis literature  (33).  Now  studying  the  surface  from  left  to 
right,  we  observe  that  plankters  living  nearer  the  surface  are 
generally  more  efficient  in  light  utilization  when  compared  at  the 
same  suspension  depths  with  those  from  deeper  layers,  except  that 
organisms  collected  from  10  ft.  appear  to  be  almost  as  efficient 
as  those  from  2  ft.  when  both  are  suspended  at  the  deeper  level. 
In  general,  then,  the  relationships  of  Figure  9  are  consistent  with 
those  of  Figure  8  in  denoting  greater  productive  capacity  of  surface 
populations  compared  to  those  from  farther  down.  The  observation 
that  efficiency  increases  as  light  decreases  can  be  interpreted  to 
be  adaptively  significant  in  respect  to  the  goal  of  biomass  maxi- 
mization. The  extent  of  this  dark-adaptability  under  natural 
conditions  is  emphasized  by  comparing  efficiencies  of  the  popu- 
lations at  the  depths  they  naturally  occupy.  Thus  the  points  (2,2, 


Information  Concept  in  Ecology 


157 


MEAN 
PHOTOSYNTHESIS 

(gcal  cm"2day"') 


5-1 


(10,2,257) 


(2,10,1.06) 


COLLECTION      DEPTH   (ft) 

Fig.  8.  Photosynthesis  as  a  function  of  sample  collection  and_suspension  depths 
(means  for  ten  experiments). 


18.5),  (6,6,19.5),  and  (10,10,44.7)  in  Figure  9  indicate  a  2.4-fold 
efficiency  increase  at  10  ft.  compared  to  2  ft. 

Two  phenomena  may  be  involved  in  this  increased  efficiency 
of  the  deeper  populations:  1)  the  purely  numerical  "swamping" 
effect  of  more  photons  in  the  upper  layers  of  the  water  column 
than  can  possibly  be  absorbed  by  the  plant  pigments  (34),  and 
2)  actual  increase  in  the  thermodynamic  efficiency  of  chlorophyll 
with  depth.  The  latter  is  illustrated  in  Figure  10  in  which  mean 
photosynthesis  per  unit  mean  illumination  per  unit  mean  initial 
concentration  of  total  chlorophyll  in  the  ten  experiments  is  graphed. 


158 


Information  Storage  and  Neural  Control 


MEAN 

PHOTOSYNTHESIS 

PER  UNIT 

ILLUMINATION 

(gcal    kcal'i] 


10,53.1) 


•50 


-28 


(2,2,18.5) 


--I0 
6  10 

COLLECTION         DEPTH     (ft) 
Fig,  9.  Photosynthesis  per  unit  illumination  as  a  function  of  sample  collection 
and  suspension  deptlis  (means  for  ten  experiments). 


This  graph  represents  chlorophyll  efficiency.  Although  the  surface 
shown  is  fairly  similar  to  that  of  Figure  9,  the  greatest  similarities 
are  on  the  left  (2  ft.  collection  depth)  and  in  the  rear  (2  ft.  sus- 
pension depth).  The  forward  part  of  the  Figure  10  surface  (10  ft. 
suspension  depth)  and  the  rigiit-hand  side  (10  ft.  collection  depth), 
however,  are  considerably  more  elevated  than  those  of  Figure  9. 
These  relationships  appear  at  a  glance  by  noting  that  much  more 
of  the  underside  of  the  Figure  10  surface  is  visible  than  that  of 
Figure  9. 


Information  Concept  in  Ecology 


159 


MEAN 

PHOTOSYNTHESIS 

PER   UNIT 

ILLUMINATION    8 
CHLOROPHYLL 
( gcal  kcal"'  ug"') 


(2,10,10.9) 


f-IO 


-5 


(2,2,3.9)     (10,10,9.2) 


(  10,6,6.1) 


10,2,2.6) 


COLLECTION      DEPTH    (ft) 

Fig.   10.  Photosyntliesis  per  unit  illumination  and  chlorophyll  as  a  function  of 
sample  collection  and  suspension  depths  (means  for  ten  experiments). 


To  test  the  significance  of  these  relationships,  the  vertical  co- 
ordinate of  the  Figure  10  points  was  divided  into  that  of  the 
Figure  9  points  to  obtain  the  ixiean  chlorophyll  concentrations 
required  to  give  unit  efficiency:  4.86,  4.41  and  4.05  /xg",  respec- 
tively, for  samples  collected  at  2,  6  and  10  ft.  These  values  indicate 
less  chlorophyll  to  be  required  as  sample  depth  is  increased.  Only 
the  first  two  means  were  significantly  difTerent,  however,  estab- 
lishing that  the  chlorophyll  at  2  ft.  was  less  efficient  that  that  at 
6  ft.  Because  of  the  high  variance  associated  with  the  10  ft.  samples, 


160  Information  Storage  and  Neural  Control 

the  2 — 10  ft.  and  6 — 10  ft.  means  could  not  be  distinguished.  There- 
fore, the  tendency  toward  increased  chlorophyll  efficiency  with 
depth  of  collection  cannot  be  formally  accepted  as  a  generalization. 
We  may  accept  it  on  intuitive  grounds,  however,  noting  the  high 
likelihood  for  a  Type  II  (35)  biometrical  error  due  to  small  sample 
size,  i.e.,  an  error  such  that  the  null  hypothesis  is  accepted  when 
in  fact  it  is  false. 

Although  physiological  mechanisms  {e.g.,  photoinhibition)  are 
doubtlessly  involved  in  the  observed  increase  of  chlorophyll 
efficiency  with  depth,  ecological  factors  are  also  implicated.  One 
of  the  striking  features  about  the  vertical  organization  of  summer 
estuarine  plankton  communities  is  variability  in  species  com- 
position and  in  cell  concentrations.  The  numerical  stratification 
of  the  York  River  phytoplankton  has  already  been  described 
(Fig.  4).  Table  I  is  provided  to  illustrate  the  nature  of  species 
changes  with  depth.  It  is  a  list  of  phytoplankton  species  and  their 
concentrations  obtained  from  the  third  experiment  (July  6)  of  the 
series  under  consideration.  There  is  nothing  especially  atypical 
about  this  particular  list;  it  is  fairly  representative. 

The  table  shows  that  two  flagellates  (in  decreasing  order  of 
importance:  Massartia,  Chilojrtorms)  were  dominant  at  the  surface. 
Two  feet  below,  three  species  dominated  in  a  diff'erent  order  of 
abundance  {Alassartio,  Gjrodimum,  Chilomonas).  These  forms  are 
all  highly  motile;  Massartia  and  Gyrodinium  are  dinoflagellates, 
Chilomonas  is  a  yellow-green  flagellate.  Both  of  these  groups  typically 
photosynthesize  at  maximal  rates  under  conditions  of  high  light 
intensity  (36).  At  6  ft.  the  surface  forms  were  no  longer  of  sig- 
nificance (dominants  being  Eutreptia,  Gyrodinium),  and  at  10  ft. 
they  were  entirely  absent  (dominants:  Eutrepfia,  Pyramimonas, 
Leplocylindricus) .  Eutreptia  is  a  euglenoid,  Pyramimonas  a  flagellated 
green  alga,  and  Leplocylindricus  an  immotile  diatom.  The  latter 
two  groups,  in  estuaries,  are  generally  adapted  to  photosynthesize 
maximally  under  conditions  of  low  or  medium  illumination  (36). 

It  would  seem  from  these  few  general  observations  that  main- 
tenance of  a  suitable  vertical  diversity  structure  might  constitute 
a  significant  segment  of  community  strategy  in  implementing  the 
goal  of  biomass  maximization.  That  a  very  definite  vertical 
diversity  pattern  is  maintained  in  summer  is  illustrated  in  Figure  1 1 


Information  Concept  in  Ecology 


161 


5  ^ 


<^ 


[2  w 


z  >■ 

OP 
Z    H 


§1 

§0 

<  s 


CN       o  CN  o  o 
I    o    I    r--  m  r~-  r- 


■--, 


m  cn  00  CN 
o  m  ■*  CO 
\o  -^  en  ■r- 


r^  \o  c?\  CN 
I   ^  in  o  o    1 

■^   un   (N   CM 


00  00  so  -^  M  c^ 
O  CM  t-^  O  CO  CO 
in  o  CM  -^  M  CM 


t~-   I    1  --   I    I    I 


I     I    r~-    I   r--    I     I    o    I     1 


rv]  CM  \o  <^]  \o 

ro   CO   ^-c   •^   ^-' 
CM   CM   T-H  ,-. 


I    I    i 


00    00    00    00   CN)    CM    CM   V3   O    C^]    C^I    C^l    C^l    CV) 

r^cM'*-<^rOcOtOT-H.^T-.'rH.^.^T-<     I 

OOCncOcOCMCMCMt-ht-i 


I  I  I  I  I  I  1        T-H      ^ 


^      — 


br.  t« 


g   _.  o     u 


i-,    ^      -o    - 


c    tj: 

U     TO     ft 

w  §:! 

•  —    o 
en  -o     g 


o 


-  S3-?  i« 

«j    O     1^     i- 
— »  -—     u     03 


J2  «      ^  O 


^  S 


■a,  C"':=    2  5 


OCi 


5    5^    c; 

O  h2  a. 


^ 

<u 

^      W 

5 

s 

>. 

^^ 

:^ 

s 

^ 

G. 

-S  a. 

o 

■~ 

0^ 

_0 

'■p.   o 

c3 

-G   -2 

.5 

j^ 

a  — 

ft 

T3 

o 

u 

>^ 

Q 

T3 

■*^ 

-a 

dT3 

1 

^V 

s 

jj 

M       U 

■  S  S 

E: 

C 

ft 

G 

~  c 

'Z 

lj 

">      5J 

Q 

-a 

^ 

c 

-0 

^:a 

162 


Information  Storage  and  Neural  Control 
SURFACE 


DEPTH       10- 
(ft) 


BOTTOM 


5  10 

MEAN    COWMUNITY    DIVERSITY 
I  bits  X  lO-yml  ) 


Fig.  11.  Mean  community  diversity,  D  =   — ^  Nj  log  P(Si),  at  various  depths. 


in  which  mean  community  diversity  for  the  ten  experiments,  as 
defined  in  Equation  [4],  is  plotted  against  depth.  The  figure  shows 
maximum  diversity  at  the  2  ft.  leveL 

From  the  foregoing  data  (Figs.  4  and  11)  it  is  concluded  that 
the  living  organisins  of  the  York  River  summer  plankton  are 
vertically  stratified  in  excess  of  the  extent  attributable  to  cor- 
responding heterogeneity  of  the  physical  environment  (Fig.  3). 
Maintenance  of  such  a  concentration  gradient  against  the  mixing 
forces  of  the  environment  inust  therefore  be  endergonic — the 
organisms  must  expend  biomass  energy  to  reduce  the  entropy  of 
their  distribution  in  space. 


Injormatwn  Concept  in  Ecology 


163 


In  oceanography,  there  is  a  prominent  and  widely  accepted 
theory  that  energy  accrual  by  a  planktonic  system  cannot  exceed 
respiratory  losses  in  a  uniformly  mixed  water  column  (37).  Vertical 
stratification  of  the  organisms  is  therefore  necessary  for  positive 
energy  balance.  This  theory  has  never  been  rigorously  developed, 
however,  and  as  a  matter  of  fact  has  recently  (38)  been  invalidated 
by  proof  for  a  countertheorem:  Vertically  homogeneous  plankton 
communities  are  energetically  feasible.  Stratification  is  therefore 
not  essential  to  positive  energy  balance.  This  conclusion  makes  the 
foregoing  York  River  observations  difficult  to  understand.  Why 
should  a  community  expend  energy  to  achieve  and  maintain  pro- 
nounced vertical  stratification  if  it  is  not  thermodynamically 
essential  for  it  to  do  so?  Consider  the  following. 

The  important  variable  relating  to  community  energy  balance  is 
the  cost  as  defined  in  Equation  [22].  In  Figure  12  mean  cost  data 
are  graphed  as  a  function  of  collection  and  suspension  depths; 


(6,10,2.671 


(10,10,2.99), 


SUSPENSION      DEPTH    (ft) 

Fig.  12.  Cost,  p  TT-i  as  a  function  of  collection  and  suspension   deptlis  (means 

for  ten  experiments). 


164  Information  Storage  and  Neural  Control 

this  particular  grapii  is  rotated  90°  clockwise  around  the  vertical 
axis  (compared  to  previous  figures  of  this  type)  to  improve  the 
perspective  in  which  the  surface  is  viewed.  Studying  the  surface 
from  back  to  front  first,  we  see  that  cost  increases  in  a  generally 
hyperbolic  or  logarithmic  fashion  with  depth  of  collection;  the 
surface  is  saddle-shaped,  being  convex  upward  from  back  to  front. 
The  fact  that  it  rises  toward  the  viewer  supports  the  previous 
conclusion  that  the  deeper  populations  are  less  viable  than  those 
nearer  the  surface — their  cost  of  operation  is  higher.  The  ribbon- 
shaped  segment  in  the  figure  denotes  the  loci  on  the  surface  and 
on  the  horizontal  plane  where  the  ratio  p7r~^  is  unity,  i.e.,  where 
an  exact  balance  between  energy  inputs  and  expenditures  is 
achieved.  For  any  specified  collection  depth,  this  ribbon  indicates 
the  depth  at  which  the  sample  must  be  suspended  to  achieve  a 
steady  state  between  inputs  and  losses.  This  depth  is  seen  to  become 
shallower  as  the  collection  depth  increases — another  indication  of  the 
intrinsically  higher  vitality  of  populations  found  nearer  the  surface. 

Viewing  the  surface  of  Figuie  12  from  right  to  left,  cost  is  shown 
to  increase  as  suspension  depth  is  increased.  In  this  direction  the 
surface  is  concave  upward.  Thus,  despite  the  measure  of  dark- 
adaptability  demonstrated  earlier,  the  price  to  a  population  of 
inhabiting  deeper  layers  in  the  water  mass  is  unequivocally  in- 
creased cost  of  operation.  This  datum  appears  to  provide  an 
economically  logical  reason  for  stratification.  A  well-known  doc- 
trine from  marginal  analysis  in  economics  (39)  states  that  the 
scale  of  an  activity  should  be  expanded  so  long  as  marginal 
profitability  (increase  in  net  utility  gain)  is  a  positive  value,  and 
carried  to  a  point  where  marginal  yield  is  zero.  This  corresponds 
to  the  procedure  in  calculus  of  maximizing  a  function  by  setting 
its  first  derivative  to  vanish.  Applied  to  the  plankton,  this  law 
demands,  in  view  of  observed  depth-cost  relationships,  that  the 
community  should  invest  biomass  energy  to  concentrate  its  com- 
ponent organisms  near  the  surface  up  to  the  point  where  additional 
return  becomes  zero.  It  would  appear  that  the  stratification 
behavioi  of  the  York  River  plankton  is  consistent  with  sound 
economic  policy. 

The  converse  of  the  marginal  profitability  law  would  be:  If 
marginal  gains  are  negative,  the  scale  of  an  activity  should  be 


Information  Concept  in  Ecology  165 

reduced  at  least  until  a  point  of  no  further  loss  (zero  return)  is 
reached.  Let  us  examine  the  behavior  of  the  York  plankton  in 
respect  to  this  proposition.  Referring  to  Figure  6,  mean  photo- 
synthesis is  observed  to  exceed  mean  respiration  in  the  upper 
water  column  {pir'~^  <  1)  but  not  in  the  lower  (ptt"^'  >  1)-  This 
relationship  is  so  typical  in  aquatic  communities  that  the  depth 
at  which  the  photosynthesis  and  respiration  curves  cross  (ptt"^  =  1) 
is  a  standard  variable — the  compensation  depth.  The  mean  depth 
of  compensation  at  the  York  sampling  station  during  the  summer 
of  1960  was  6.5  ft.;  this  level  is  denoted  by  broken  lines  in  Figures 
3-7  and  in  Figure  1 1 .  When  phytoplankters  drift  beneath  the 
instantaneous  compensation  depth  they  experience,  on  the  average, 
a  shift  fi'om  positive  to  negative  energy  balance.  If  a  net  positive 
balance  is  to  be  achieved  for  the  whole  water  columii  it  is  necessary 
that  the  community  reduce  energy  losses  in  the  lower  part  of  the 
column.  This  implies,  by  the  mathematical  nature  of  the  cost 
variable,  increasing  the  rate  of  photosynthesis  and /or  depressing 
the  rate  of  respiration.  Community  behavior  in  accordance  with 
tlie  former  imperative  has  already  been  described  as  dark-adap- 
tability. We  consider  now  the  attenuation  of  respiration. 

The  data  which  have  been  presented  indicate  that  althougli 
photosynthetic  capacity  of  the  plankters  was  irreversibly  (in  24 
hours)  less  at  the  6  and  10  ft.  levels  than  at  2  ft.  (Fig.  9),  vigorous 
respiration  equivalent  to  that  of  surface  populations  persisted 
down  to  10  ft.  (Fig.  6).  Below  10  ft.,  however,  oxygen  uptake  was 
sharply  reduced.  The  extent  of  actual  metabolic  failure  must  be 
even  gi^eater  than  indicated  by  Figure  6  since  the  concentration 
of  oxidizable  detritus  increased  with  depth  (Fig.  7)  producing  a 
continually  increasing  oxygen  demand  (reflected  in  the  oxygen 
curve  of  Fig.  3).  This  underscores  the  conclusion  that  metabolism 
is  sharply  curtailed  soon  after  the  organisms  drift  beneath  the 
compensation  depth.  This  phenomenon  constitutes  a  pei'fect 
response  on  the  part  of  tlie  community  to  the  converse  marginal 
profitability  principle,  and  is  an  example  of ''beneficial  death"  (24) 
at  the  community  level.  Beneficial  death  is  u.sually  thought  of  in 
connection  with  individuals  {e.g.,  dead  cells  forming"  the  matrix 
of  a  functional  tissue,  as  in  plant  xylem  or  some  insect  wings)  or 
populations  {e.g.,  annual  plants,  some  social  insect  castes,  genetic 


166  Information  Storage  and  Neural  Control 

lethals).  That  the  comparatively  loosely  organized  coinmunity 
may  also  derive  profit  through  death  of  its  constituent  organisms 
at  an  appropriate  time  is  an  interesting  speculation. 

First  of  all,  planktonic  systems  such  as  these  have,  of  course, 
evolved.  One  of  the  important  taxonomic  characteristics  of  the 
algal  phyla  is  the  nature  of  food  storage  products.  It  would  seem 
that  with  such  a  capability  already  well  developed  generally  in 
these  groups  there  could  have  evolved,  in  the  time  available, 
species  able  to  maintain  robust  metabolic  activity  right  down  to 
the  bottom  if  it  were  consistent  with  community  design.  This 
would  be  especially  adaptive  in  water  masses  where  expectation 
for  return  to  the  trophogenic  zone  (above  the  compensation  depth) 
through  vertical  turbulence  would  be  good.  Indeed,  evolutionary 
theory  asserts  that  such  foims  would  enjoy  a  selective  advantage 
over  more  labile  ones.  In  phytoplankton,  the  smaller  motile 
species  typically  possess  rapid  dynamics  and  short  generation 
times,  but  lack  the  capacity  for  sustained  yields  characteristic  of 
larger  forms  with  more  conservative  dynamics  (15).  Clearly,  from 
the  standpoint  of  the  York  system  in  summer  with  its  slight  vertical 
density  gradient,  the  latter  type  of  organism  would  not  be  nearly 
so  satisfactory  a  component  as  the  former.  Their  production  rates 
per  unit  of  biomass  would  be  slower  in  the  upper  water  column, 
and  their  collective  respiration  higher  in  the  depths;  the  result 
might  be  a  net  energy  loss  to  the  conmiunity.  Under  winter  con- 
ditions when  hydrography  is  such  that  vertical  turbulence  is 
extreme  and  the  water  column  thoroughly  mixed,  larger  species 
with  longer  generation  times,  lower  light  optima,  and  greater 
capabilities  for  food  storage  niight  be  more  serviceable  con- 
stituents. Perhaps,  therefore,  the  seasonal  replacement  of  summer 
flagellate  floras  by  diatomaceous  communities  in  winter  and  spring 
may  be  taken  to  reflect  community  adaptability  in  response  to  a 
changing  environment.  In  view  of  this,  it  does  not  seem  unreason- 
able that  particular  species  may  be  selected  for  occupancy  in  a  com- 
munity under  a  specific  environmental  regime,  not  only  for  their 
Darwinian  competence  in  competition,  but  as  well  for  their  compati- 
bility as  functional  components  of  a  goal-adapted  "machine"  (40). 

This  possibility  would  seem  to  add  another  dimension  to  the 
classical  concept  of  ecological  community  because  of  its  implicit 


Information  Concept  in  Ecology  167 

demand  that  the  success  or  failure  of  species  be  related  to  and 
interpreted  in  a  broader  sociological  context.  Acceptance  of  such 
a  context  carries  with  it  the  important  advantage  of  making  some 
of  the  elegant  formalisms  (9,  23)  developed  in  connection  with  the 
study  of  situations  of  conflict  available  for  ecological  analysis.  The 
theory  of  games  and  decisions  is,  however,  notoriously  teleological 
in  basis:  litigants  come  to  odds  through  mutual  impairment  of 
purposive  behavior.  This  objection  can  be  ameliorated  to  a  very 
large  extent  by  regarding  community  goal-adapted  behavior  in 
a  teleonomic,  not  teleological,  sense;  i.e.,  the  community  is 
"programmed"  for  goal  achievement  though  possessing  no  "con- 
scious" knowledge  of  the  goal.  This  kind  of  thinking  is  widely 
accepted  in  connection  with  the  problem  of  DNA  coding,  and 
it  has  been  formalized  in  Bellman's  (40)  concept  of  information 
pattern.  In  such  a  framework,  the  mechanism  of  natural  selection 
may  still  be  construed  to  operate  at  an  infraspecies  level;  for 
example,  by  acknowledging"  that  the  information  pattern  of  a 
species  (a  program  containing  the  accumulated  history  of  its  past 
and  rules  for  decision  making)  can  enable  the  latter  to  make,  in  a 
completely  mechanistic  manner,  a  choice  between  alternative 
strategies  such  as  those  embodied  in  a  recent  theorem  (41)  due 
to  Rashevsky:  If  two  individuals  work  on  the  production  of  some 
object  of  satisfaction  (utility)  and  if  their  cooperative  efforts  result 
in  an  increased  overall  productivity,  then  each  individual  will 
have  less  of  the  object  of  satisfaction  if  each  adopts  a  strategy  of 
maximizing  his  own  satisfaction  (egoism,  competition)  than  if  each 
tries  to  maximize  the  sum  of  the  satisfactions  of  both  individuals 
(altruisin,  cooperation). 

The  importance  of  epistemological  bearing  in  determining  the 
character  of  questions  which  one  may  ask  of  biosystems  and, 
consequently,  that  of  the  answers  elicited  can  be  illustrated  as 
follows.  Consider  a  proposition  of  the  form,  "The  organism 
(species,  community)  is  adapted  to  .  .  . ."  This  is  completely 
acceptable  biological  rhetoric.  Constructed  in  the  passive  voice, 
the  statement  carries  the  implication  that  it  is  the  fortuitous 
environment  which  does  the  selecting.  If  we  go  to  the  active  form, 
"The  organism  adapts  to  .  .  .,"  we  provide  the  biological  sub- 
ject with  a  degree  of  initiative  in  the  process.  This  is  still  quite 


1 68  Information  Storage  and  Neural  Control 

acceptable.  If  now  we  change  the  verb  akogether  and  posit, 
"The  organism  adopts  a  strategy  for  .  .  .,"  we  pass  for  many 
readers  rather  too  abruptly  into  the  realm  of  purpose.  Thus,  it 
might  be  more  suitable  to  say  instead,  "The  organism  is  pro- 
grammed for  a  strategy  of  .  .  .."  The  important  point  to  im- 
press here  is  that  all  of  these  statements  mean  essentially  the 
same  thing  mechanistically,  though  epistemologically  they  are 
poles  apart.  Consequently,  they  give  rise  to  very  different  ways  of 
asking  questions,  therefore  to  divergent  investigational  approaches, 
and  finally  to  quite  different  classes  of  answers. 

To  illustrate,  if  in  the  present  instance  the  hrst-mentioned 
point  of  view  is  adopted,  then  only  the  empirical  sections  of  this 
paper  would  have  relevance,  and  its  content  might  be  summarized 
by  saying:  The  York  River  plankton  community  appears  to  be 
eminently  adapted  to  its  environment  as  indicated  by  1)  stratifica- 
tion of  organisms  near  the  surface  where  there  is  more  light, 
2)  increase  of  chlorophyll  efficiency  with  depth  due  to  both 
physiological  and  species  compositional  reasons,  and  3)  sharp 
curtailment  of  respiration  in  the  lower  part  of  the  water  column 
as  the  organisms  die  and  sink,  making  possible  a  positive  balance 
between  energy  gains  and  losses  in  the  community.  This  is  a 
descriptive  approach,  and  it  yields  purely  descriptive  answers  with 
limited  power  to  provide  real  insight  into  the  marvel  of  organiza- 
tion and  behavior  which  is  the  community. 

Contrast  this  with  the  summary  which  might  result  from 
acceptance  of  the  last  point  of  view:  Based  on  the  above-mentioned 
observations,  the  York  River  community  appears  to  be  pro- 
grammed for  a  strategy  of  maximizing  its  biomass,  therefore  its 
energy  content,  therefore  its  ability  to  purchase  utility  and  increase 
its  information  reserves,  therefore  its  diversity  or  richness  of  form, 
and  therefore  its  stability  in  a  variable  environment.  In  the 
process,  the  community,  inchoate  a  biological  system  as  it  is, 
meets  some  fundamental  thermodynamic  and  economic  impera- 
tives, as  well  as  the  dictum  of  Shannon's  Theorem  10. 

The  new  level  of  abstraction  so  attained  may  or  may  not  qualify 
the  community  as  a  Wienerian  (42)  machina  ratiocinatrix,  but  if 
there  is  a  distinction,  it  would  seem  to  lie  largely  in  the  realm  of 
logic  and  semantics,  not  of  biology. 


Information  Concept  in  Ecology  169 

REFERENCES 

1.  Lindeman,  R.  L.:  The  trophic-dynamic  aspect  of  ecology.  Ecology, 

2J.-399-418,   1942. 

2.  MacFadyen,  A.:  The  meaning  of  productivity  in  biological  systems. 

J.  Anim.  EcoL,  77.-75-80,  1948. 

3.  Schrodinger,    E.:    What   is   Life?    Cambridge,    England,    University 

Press,  1945. 

4.  Brillouin,  L.:  Life,  thermodynamics,  and  cybernetics.  Sci.  Amer.,  37: 

554-568,  1949. 

5.  Branson,  H.  R. :  A  definition  of  information  from  the  thermodynamics 

of  irreversible  processes  and  its  application  to  chemical  communi- 
cation in  biological  systems,  in  Quastler,  H.  (ed.),  Information 
Theory  in  Biology.  Urbana,  Univ.  Illinois  Press,  1953,  p.  25. 

6.  Linschitz,  H. :  The  information  content  of  a  bacterial  cell,  in  Quastler, 

H  (ed.),  Information  Theory  in  Biology.  Urbana,  Univ.  Illinois  Press, 
1953,  p.  251. 

7.  Patten,  B.  C:  An  introduction  to  the  cybernetics  of  the  ecosystem: 

the  trophic-dynamic  aspect.  Ecology,  40:22\-23\,   1959. 

8.  Patten,  B.  C:  Negentropy  flow  in  communities  of  plankton.  Limnol. 

Oceanogr.,  6.-26-30,   1961. 

9.  Thrall,  R.   M.,   Coombs,   C.  H.  and  Davis,   R.   L.   (eds.):  Decision 

Processes.  New  York,  Wiley,  1954. 

10.  Shannon,   C.   E.:   A  mathematical   theory  of  communication.   Bell 

Syst.   Tech.  J.,  27.-379-423,  623-656,   1948. 

11.  Ashby,  W.  R.:  An  Introduction  to  Cybernetics.  New  York,  Wiley,  1956. 

12.  Ashby,  W.  R.:  Design  for  a  Brain,  ed.  2.  New  York,  Wiley,  1960. 

13.  Margalef,   D.    R.:    Informacion  y  diversidad   especifica  en  las  com- 

munidades  de  organismos.  Invest.  Pesquera,  3.-99-106,  1956. 

14.  Margalef,  D.  R.:  La  teoria  de  la  informacion  en  ecologia.  Mem.  R. 

Acad.  Cien.  Artes,  Barcelona,  J2.-373-449,   1957. 

15.  Margalef,  D.  R.:  Temporal  succession  and  spatial  heterogeneity  in 

phytoplankton,  in  Buzzati-Traverso,  A.  A.  (ed.),  Perspectives  in 
Marine  Biology.  Berkeley,  LIniv.  California  Press,  1958,  p.  323. 

16.  MacArthur,  R.:  Fluctuations  of  animal  populations,  and  a  measure 

of  community  stability.  Ecology,  J(5." 533-536,  1955. 

17.  Hutchinson,  G.  E.:  Concluding  remarks.  Cold  Spring  Harbor  Symp. 

Qimnt.   BioL,   22:415-427,   1957. 

18.  Wiener,  N.:  The  Human  Use  of  Human  Beings.  New  York,  Doubleday, 

1950. 

19.  Brillouin,   L.:  Science  and  Information    Theory.   New  York,  Academic 

Press,  1956. 


170  Information  Storage  and  Neural  Control 

20.  Rothstein,  J.:    Communication,   Organization,   and  Science.    Indian   Hills, 

Colorado,   Falcon's  Wing  Press,   1958. 

21.  Jaynes,  E.  T.:  Information  theory  and  statistical  mechanics.  P/owVa/. 

Rev.,    7O6.-620-630;    7C5.-171-190,   1957. 

22.  Tribus,  M.:   Information  theory  as  the  basis  for  thermostatics  and 

thermodynamics.   Gen.  Syst.,  6.-127-138,   1961. 

23.  Von  Neumann,  J.  and  Morgenstern,  O.:  Theory  of  Games  and  Economic 

Behavior.  Princeton,  N.  J.,  Princeton  Univ.  Press,  1947. 

24.  Allee,  W.  C,  Emerson,  A.  E.,  Park,  O.,  Park,  T.  and  Schmidt,  K.  P.: 

Principles  of  Animal  Ecology.  Philadelphia,  Saunders,  1949. 

25.  Szent-Gyorgi,  A.:  Submolecular  biology.  Radiation  Res.,  1960,  supp. 

2,  p.  4. 

26.  Arnon,  D.   I.:  The  role  of  light  in  photosynthesis.  Sci.  Amer.,  203: 

105-118,   1960. 

27.  Gaarder,  T.  and  Gran,  H.  H.:  Investigation  on  the  production  of 

plankton  in  the  Oslo  Fjord.   Rapp.  et  Proc. —  Verb.,   Cons.   Internal. 
Explor.   Mer.,  ^i.-l-48,   1927. 

28.  Gilli)richt,  M.:  Untersuchungen  zur  Productions  Biologic  des  Plank- 

tons in  der  Kieler  Bucht.   I.  h'ieler  Meeresf.,  <§.•  173-1 91,   1951. 

29.  Krey,  J.:    Untersuchungen  zum   Sestongehalt  des   Meerwassere   1: 

Der  Sestongehalt  in  der  westlichen  Ostsee  und  unter  Helgoland. 
Ber.   Deutsch  Komm.   Meeresf.,    72:431-456,    1952. 

30.  Vallentyne,   J.    R.:    Sedimentary   chlorophyll    determination    as    a 

paleobotanical  method.   Canadian  J.   Bot.,  Ji.-304-313,   1955. 

31.  Vallentyne,  J.  R.,  and  Bidwell,  R.  G.  S.:  The  relation  between  free 

sugars   and   sedimentary  chlorophyll   in  lake   muds.   Ecology,   37: 
495-500,  1956. 

32.  Vallentyne,  J.    R.,   and   Craston,   D.    F.:    Sedimentary  chlorophyll 

degradation   products  in   surface   muds  from   Connecticut   lakes. 
Canadian  J.  Bot.,  35:35-42,  1957. 

33.  Rabinowitch,   E.   I.:  Photosynthesis  and  Related  Processes,   New  York, 

Interscience,  1951,  vol.  2,  pt.  1,  p.  603. 

34.  Odum,  H.  T.,  McConnell,  W.,  and  Abbott,  VV.:  The  chlorophyll 

"A"  of  communities.  Publ.  Inst.  Mar.  Sci.,  Univ.  Texas,  5.-65-96,  1958. 

35.  Snedecor,   G.  W.:  Statistical  Methods  Applied  to  Experiments  in  Agri- 

culture and  Biology,  ed.  5,  Ames,  Iowa,  Iowa  State  College  Press, 
1956. 

36.  Ryther,  J.  H.:   Photosynthesis  in  the  ocean  as  a  function  of  light 

intensity.  Limnol.  Oceanogr.,   7.-61-70,   1956. 


Information  Concept  in  Ecology  171 

37.  Sverdrup,  H.  U.:  On  conditions  for  the  vernal  blooming  of  phyto- 

plankton.  J.  dii  Conseil,   7<§;287-295,   1953. 

38.  Patten,  B.   C:   Energy-depth  relationships  in  plankton.    In  manu- 

script. 

39.  Baumol,   VV.  J.:  Economic   Theory  and  Operations  Analysis.  Englewood 

Chffs,  N.  J.,  Prentice-Hall,  1961. 

40.  Bellman,  R.:  Adaptive  Control  Processes:  A  Guided  Tour.  Princeton,  N.  J., 

Princeton  University  Press,  1961. 

41.  Rashevsky,  N.:  A  contribution  to  the  theory  of  egoistic  and  altruistic 

interactions.  Bull.  Math.  Biophys.,  2J.-115-134,  1961. 

42.  Wiener,  N.:  Cybernetics.  New  York,  Wiley,  1948. 

DISCUSSION  OF  CHAPTER  VII 

Walter  Abbott  (Houston,  Texas) :  You  are  undoubtedly  aware 
of  the  criticisms  of  the  light-dark  bottle  technique,  mainly  because 
of  the  reduction  in  turbulence.  Does  this  mean  that  your  data 
represent  minimum  estimates? 

Bernard  C.  Patten  (Gloucester  Point,  Virginia):  I  do  not 
know  what  it  means,  actually.  There  are  four  or  five  classes  of 
criticism  of  the  light  and  dark  bottle  method.  Some  of  them  work 
in  opposition,  i.e.,  an  error  in  one  direction  may  be  cancelled  or 
partially  mollified  by  another  error  in  the  reverse  direction.  I 
think  the  technique  is  marginal,  at  best,  for  obtaining  absolute 
measures  of  energy  flux,  but  quite  adequate  for  relative  com- 
parisons of  the  activities  of  different  populations,  which  is  how  we 
used  it.  If  you  perform  enough  experiments  and  observe  that  a 
fairly  consistent  pattern  emerges,  you  can  begin  to  feel  confident 
of  the  reality  of  the  pattern  even  though  the  data  may  represent 
minimal  estimates. 

Heather  D.  Mayor  (Houston,  Texas):  Would  you  have  to 
allow  for  extra  energy  gain  to  your  system,  brought  about  by  the 
process  of  measurement?  Is  there,  for  example,  additional  "noise" 
added  to  the  respiration  term  because  of  the  measurements? 

Patten:  I  am  not  quite  certain  what  you  mean,  but  we  do  make 
corrections  of  the  type  you  suggest.  We  make  a  correction  for  the 
fact  that  the  photosynthetic  quotient  is  generally  greater  than 
unity  in  marine  phytoplankton,  but  I  do  not  believe  that  this  is 
the  kind  of  thing  to  which  you  are  referring. 


1 72  Injormatwn  Storage  and  Neural  Control 

Mayor:  Using  a  quantum  analogy,  I  know  you  would  be 
introducing  an  additional  perturbation  by  measuring  your  param- 
eters. I  was  wondering"  whether  you  have  considered  this  problem, 
or  whether  you  are  working  at  a  level  where  this  type  correction 
is  not  necessary. 

Patten:  If  there  is  an  analogous  problem  here,  I  am  not  aware 
of  it.  I  believe  we  may  be  thinking  as  well  as  working  at  different 
levels. 


CHAPTER 
VIII 

EXCHANGE  OF  INFORMATION  ABOUT 
PATTERNS  OF  HUMAN  BEHAVIOR 

Gregory  Bateson,  M.A, 


A 


.T  THE  outset  I  wish  to  make  two  acknowledgments.  First, 
I  would  like  to  credit  Attneave's  work  (1)  in  which  he  points 
out  the  synonymy  between  what  in  information  theory  is  called 
"redundancy"  and  what  in  popular  parlance  is  called  "pattern." 
You  will  see,  as  I  develop  what  I  have  to  say,  that  this  synonymy 
is  basic.  Second,  I  want  to  acknowledge  a  less  definable  debt  to 
conversations  with  Alex  Bavelas  about  his  experiments  involving 
varieties  of  contingency  in  learning  contexts.  I  had  hoped  that 
the  outcome  of  these  conversations  would  be  a  paper  in  which 
his  name  would  be  included  as  joint  author.  Since  our  diverse 
professional  commitments  have  prevented  our  getting  together  on 
this,  I  must  take  responsibility  for  the  thoughts  which  his  work 
has  stimulated  in  me.  A  major  part  of  this  paper  will  be  devoted 
to  defining  that  order  of  information  which  I  regard  as  "infor- 
mation about  patterns  of  human  behavior."  This  involves  a 
restructuring  of  learning  theory. 

Let  us  assume  that  all  receipt  of  information  is  "learning."  This 
will  bring  within  a  single  theoretical  spectrum  the  whole  range  of 
phenomena,  beginning  with  the  receipt  of  a  pip  by  a  receiving 
machine  at  the  end  of  a  wire,  up  to  and  including  such  complex 
phenomena  as  the  development  of  neurosis  or  psychosis  under 
environmental  stress.  Notice  first  of  all  that  the  receipt  of  a  bit, 
a  yes  or  no  answer  to  a  question,  is  not  usually  called  "learning" 
if  the  receiver  already  knows  to  what  question  the  bit  is  an  answer. 
Psychologists  who  perform  what  are  usually  called  learning  experi- 

173 


1 74  Information  Storage  and  Neural  Control 

ments  generally  ignore  phenomena  of  this  order.  Their  experiments 
are  concentrated  upon  a  change  in  the  way  the  receiving  entity 
responds  to  what  is  supposed  to  be  the  same  bit  when  this  bit  is 
presented  on  successive  occasions. 

"Learning,"  as  the  word  is  used  by  psychologists,  denotes  the 
receipt  of  a  meta-bit,  i.e.,  a  piece  of  information  which  will  change 
the  subject's  response  to  some  bit.  Over  and  over  the  psychological 
experimenter  presents  the  stimulus,  a  buzzer,  followed  by  meat 
powder.  He  observes  that,  after  a  number  of  trials,  the  animal 
which  formerly  did  not  salivate  when  it  heard  the  buzzer  now 
does  salivate.  This  change  is  called  "learning."  But  when  this 
process  has  approached  completion,  if  the  psychologist  again 
presents  the  buzzer  and  the  animal  salivates,  this  receipt  of  infor- 
mation— the  receipt  of  the  sound  of  the  buzzer  as  a  yes  or  no 
answer  to  a  question  which  the  animal  can  now  identify — is  not 
usually  regarded  as  learning.  By  including  this  simplest  phenom- 
enon, i.e.,  the  receipt  of  the  single  bit,  within  the  spectrum  of 
learning,  the  question  as  to  whether  a  computer  is  or  is  not  learning 
when  it  receives  appropriate  input  is  answered  out  of  hand.  This 
is  learning  of  the  simplest  order. 

Second  order  learning  arises  when  the  subject  changes  his  ability 
to  receive  the  yes  or  no  answer  to  a  question.  This  is,  in  fact,  the 
phenomenon  which  psychologists  have  studied  maximally  in  learn- 
ing experiments;  the  dog  learns  that  the  buzzer  means  future 
meat  powder. 

But  beyond  this,  there  is  obviously  a  third  order  of  learning 
called  the  acquisition  of  "test  wisdom,"  or  "set  learning."  Here 
the  subject  learns  that  he  is  to  be  on  the  lookout  for  sequences 
of  a  certain  sort  in  his  universe,  which  include  both  external 
events  and  his  own  behaviors.  For  example,  he  learns  to  behave 
instrumentally  in  order  to  solve  the  problems  presented  by  stimuli. 
If  the  laboratory  is  Pavlovian,  he  learns  to  expect  the  stimuli  to 
be  direct  predictions  of  future  reinforcements  which  will  come 
regardless  of  his  action.  I  shall  speak  of  this  as  third  order  learning, 
referring  to  those  changes  whereby  the  subject  who  encounters 
and  solves  repeated  problems  of  a  certain  sort  comes  to  expect 
his  universe  to  be  structured  in  ways  related  to  the  formal  struc- 
turing of  these  previous  problems. 


Patterns  of  Human  Behavior  175 

To  this  formal  structuring  of  contexts,  we  can  apply  the  language 
which  invokes  contingency.  We  shall  then  say,  for  example,  that 
the  animal  which  has  undergone  recurrent  classical  Pavlovian 
experimentation  will  expect  his  universe  to  be  so  structured  that 
reinforcements  are  contingent  only  upon  stimuli,  not  upon  his  re- 
sponses. If  his  universe  is  totally  structured  in  this  way,  all  he  can 
do  is  to  prepare  for  the  coming  reinforcement,  e.g.,  by  autonomic 
measures  such  as  salivation.  He  can  predict  but  he  cannot  control. 

Note  that  a  subject,  acting  in  terms  of  this  philosophy  or  in 
terms  of  any  philosophy  of  this  order,  will,  in  general,  have  such 
experience  of  his  universe  as  will  validate  his  philosophy.  If  he 
does  not  believe  it  is  worthwhile  to  behave  instrumentally,  he 
will  never  engage  in  behavior  which  would  disprove  or  test  the 
philosophy.  And,  conversely,  if  he  has  had  past  experience  only 
of  instrumental  contexts,  he  will  have  learned  to  behave  instru- 
mentally and  will  encounter,  as  it  seems  to  him,  a  universe  in 
which  instrumental  behavior  is  appropriate.  Attempting  to  make 
a  reinforcement  come,  he  will  try  out  various  courses  of  action; 
and  when  the  reinforcement  does  come,  he  will  believe  that  the 
action  which  immediately  preceded  it  was  an  effective  instru- 
mental action.  His  experience  of  his  universe  will  validate  his 
theory  of  instrumental  magic,  even  though  the  causal  contin- 
gencies assumed  by  this  magic  may  be  mythological  or  delusory. 

Let  me  now  extend  what  I  have  said  about  individual  learning 
to  what  would  superficially  seem  to  be  much  more  complex 
phenomena— those  of  interpersonal  exchange.  To  do  this,  we 
have  only  to  personify  the  experimenter  as  well  as  the  learning 
subject  and  to  see  the  learning  experiment  as  a  small  segment 
of  an  interchange  between  two  persons:  A,  the  experimenter, 
provides  the  stimulus;  B,  the  subject,  responds  to  the  stimulus; 
and  A  follows  B's  response  with  a  reinforcement. 

Notice  that  these  categories  {stimulus,  response,  and  reiti  for  cement) 
which  we  are  putting  upon  the  behaviors  cannot  be  empty.  If 
the  experimenter  does  not  provide  a  reinforcement,  this  in  itself 
is  a  reinforcement;  and,  if  the  subject  does  not  respond  to  the 
stimulus,  this  failure  to  respond  represents  the  subject's  response 
to  the  stimulus.  Notice  also  that  if  there  were  no  stimulus,  this  in 
itself  would  be  the  stimulus  to  which  the  subject  responded.  In 


1 76  Information  Storage  and  Neural  Control 

the  world  of  communication,  a  message  does  not  have  to  be  an 
event  or  an  object  in  order  to  be  a  message.  As  my  friend  Ray 
Birdwhistell  says,  "Nothing  never  iiappens." 

If  we  look  at  an  on-going  interchange  between  persons  who 
behave  alternatively,  they  can  never  "not  behave."  The  inter- 
change has  been  going:  .  .  .  A,  B,  A,  B,  A,  B,  .  .  .  From  this 
we  cut  out,  for  our  analysis,  any  triad:  a  sequence  A,  B,  A,  or, 
if  you  like,  a  sequence  B,  A,  B.  Within  any  such  triad,  we  can  now 
recognize  that  the  third  item  is  necessarily  a  reinforcement  be- 
cause, in  this  triad,  if  the  third  item  had  been  something  other 
than  what  it  was,  or  if  it  had  been  something,  for  example,  which 
made  the  second  item  inappropriate,  it  would  obviously  have 
been  a  negative  reinforcement.  So  if  the  third  item  is  appropriate, 
it  is,  in  fact,  a  positive  reinforcement  of  the  second  item.  By  the 
same  token,  the  second  item  can  always  be  regarded  as  a  "response" 
since  it  follows  a  first  item  and  is  reinforced  by  a  third.  Cor- 
respondingly, the  first  item  is  necessarily  a  stimulus  since  it  precedes 
the  second,  which  is  reinforced  by  the  third.  These  are  purely 
formal  relations  between  items  and  must  necessarily  obtain  in  any 
triad  of  an  interchange  between  learning  entities. 

It  follows  that,  in  a  long  interchange  of  this  kind,  any  behavior 
of  B  is  necessarily  simultaneously  a  stimulus,  a  response,  and  a 
reinforcement,  according  to  how  we  slide  our  identification  of  the 
triad  up  and  down  the  series.  The  same  is  true  for  any  behavior 
of  A.  Such  a  scheme  has  the  advantage  of  presenting  to  the  scien- 
tist all  the  possibilities  for  punctuating  a  sequence  of  interchange 
at  the  level  of  complexity  of  the  triad.  It  is,  however,  arbitrary  in 
that  it  excludes  the  simpler  (dyadic)  units  of  interchange  and  also 
the  more  complex  (polyadic)  units. 

The  arbitrary  selection  of  the  trigram,  however,  does  raise  a 
number  of  interesting  problems.  Note  that  each  item  in  any  tri- 
gram is  also  a  member  of  two  other  trigrams.  Clonsider  such  a 
sequence  as  the  following: 

A  .   .   .  23  25  27  29  .   .   . 

B  ...      22  24  26  28  30  .   .   . 

In  the  sequence,  the  odd  numbers  represent  items  of  A's  behavior 
while  the  even  numbers  represent  those  of  B.   The  sequence  is 


Patterns  of  Human  Behavior  \11 

deliberately  imagined  to  be  far  from  the  beginning  and  from  the 
end  of  the  total  interchange.  It  will  be  observed  that  B's  item  26 
is  a  response  in  the  trigram  25-26-27,  but  it  is  also  a  reinforce- 
ment in  the  trigram  24-25-26  and  a  stimulus  in  the  trigram 
26-27-28.  The  formal  truth,  however,  may  not  represent  the 
natural  history  of  the  relationship  as  it  is  perceived  by  the  par- 
ticipants. They  are  busy  putting  their  labels,  imposing  their 
Gestalten,  on  the  items  and  on  the  trigrams.  It  is  perfectly  possible, 
for  example,  for  A  to  punctuate  this  interchange  in  such  a  way 
that  he  will  see  only  the  trigrams  23-24-25  and  27-28-29  and 
ignore  or  brush  off  B's  items  22  and  26,  creating  a  picture  in  which 
A  always  provides  the  stimuli  and  reinforcements  while  B  provides 
only  the  responses.  If  A  succeeds  in  maintaining  this  system  and 
in  making  B  see  the  relationship  in  the  same  way,  we  may  say 
that  A  is,  in  this  particular  sense,  the  dominant  participant  in 
the  relationship.  On  the  other  hand,  B,  by  pulling  his  punches 
on  items  22  and  26,  may  succeed  in  forcing  A  to  think  that  he 
(A)  has  the  initiative.  It  may  then  be  difficult  to  decide  who  is 
"dominant." 

At  this  point  it  is  not  appropriate  to  go  into  all  the  possible 
details  of  the  punctuation  of  such  sequences.  However,  a  part  of 
this  matter  has  been  explored  in  earlier  publications  (2)  in  which 
the  formal  resemblances  and  differences  between  dominance, 
dependency,  and  spectatorship  were  discussed.  It  was  pointed  out 
that  these  themes  of  relationship  could  be  reduced  to  paradigms 
of  learning  and  that  various  types  of  "end-linkage"'  could  occur. 
For  example.  A,  in  his  relationsliip  to  B,  could  take  the  dominant 
end  of  a  dominance-submission  relationship  and  the  succoring  end 
of  a  succoring-dependence  one.  These  patterns  could  also  be 
reversed,  in  which  case  A  would  combine  dominance  with  de- 
pendency. Very  basic  differences  between  cultures,  e.g.,  between 
the  cultures  of  England  and  America,  might  be  expressed  as 
contrasts  of  end-linkage  in  parent-child  relationships. 

But,  if  it  is  true  of  human  natural  history  that  people  punctuate 
their  interchanges  into  sequences  which  are,  in  fact,  contexts  of 
learning,  it  follows  that  in  interpersonal  interchange  we  must  also 
face  at  least  the  three  levels  of  learning  which  have  already  been 
defined    in    the   learning   experiments.    That   is,    each    person   is 


178  Information  Storage  and  Neural  Control 

receiving  bits  of  information,  and  these  bits  are  already  falling 
into  place  as  yes  or  no  answers  to  questions  of  which  the  person 
already  has  understanding.  But  the  second  order  learning  must 
also  be  occurring,  i.e.,  he  must  be  changing  his  identification  and 
understanding  of  the  questions  to  which  the  bits  are  answers;  and 
third  order  learning  must  also  be  going  on,  namely,  he  must  be 
learning  the  characteristic  patterns  of  contingency  in  this  re- 
lationship. 

The  reality  of  these  three  levels  of  learning,  especially  the 
reality  of  the  third  level  and  perhaps  of  higher  levels,  can  only 
be  demonstrated  convincingly  from  phenomena  of  pathology. 
Wlien  all  is  going  smoothly,  it  is  not  possible  to  get  a  clear  picture 
of  what  orders  of  learning  are  operating.  It  is  when  certain  orders 
of  learning  are  disturbed  that  it  becomes  possible  to  analyze  and 
recognize  these  orders. 

For  a  long  time  psychologists  have  been  performing  various 
experiments  which  amply  demonstrate  what  I  am  trying  to  say. 
Unfortunately,  the  conventional  phrasings  used  in  the  psycho- 
logical laboratories  are  not  along  the  lines  I  am  advocating  here. 
The  experiments  to  which  I  refer  are  those  called  experiments  in 
"experimental  neurosis."  Traditionally,  these  are  described  with- 
out invoking  any  theory  of  levels  of  learning.  For  example,  we  are 
told  that  the  dog  starts  to  exhibit  psychotic  or  other  sympto- 
matology when  his  "discrimination  breaks  down."  Let  me  dissect 
a  typical  experiment  for  you  so  that  you  may  see  that  what  happens 
is  not  necessarily  a  matter  of  breakdown  of  discrimination  but 
can  be  seen  as  a  matter  of  disruption  of  the  learning  process  at 
what  I  am  calling  the  third  level. 

Classically,  the  animal  is  presented  with  an  ellipse,  which 
means  x,  and  with  a  circle,  which  means  y.  If  the  dog  performs 
X  in  response  to  the  ellipse  and  y  in  response  to  the  circle,  it  either 
gets  its  reward  or  avoids  its  punishment.  But,  if  the  dog  fails  to 
"discriminate"  between  these  stimulus  objects,  it  receives  punish- 
ment or  fails  to  get  a  reward.  Having  taught  the  dog  this  dis- 
crimination, the  experimenter  begins  to  fatten  the  ellipse  and  to 
flatten  the  circle.  The  dog  responds  by  exerting  greater  effort  to 
tell  the  difference  between  the  symbols,  and  at  first  these  eff"orts 
will  be  successful.  As  a  further  stage  is  reached  and  the  discrimina- 


Patterns  of  Human  Behavior  1 79 

tion  becomes  more  difficult,  the  psychologist  makes  a  pencil  mark 
on  the  back  of  the  ellipse  in  order  to  distinguish  it  from  the  "circle." 
He  also  uses  a  coin  or  some  other  randomizing  device  to  decide 
which  of  the  stimulus  objects  he  is  going  to  administer  next.  He 
cannot  afford  to  administer  them  in  any  patterned  order  which 
the  dog  might  learn.  Finally,  these  two  objects  become  indis- 
tinguishable; i.e.,  from  the  point  of  view  of  the  dog  they  are  one 
object  or,  rather,  they  would  be  one  object  if  the  dog  had  not 
been  told  previously,  "This  is  a  context  for  discrimination."  This 
message  was  underlined  during  the  period  when  discrimination  was 
difficult  but  still  possible. 

The  message,  "This  is  a  context  for  discrimination,"  is  carried 
partly  by  the  earlier  training  and  partly  by  every  circumstance 
of  the  laboratory,  the  harness,  the  smell  of  the  experimenter,  and 
so  forth.  All  these  ancillary  stimuli  are,  in  fact,  indications  to  the 
dog  that  he  is  now  in  a  context  for  discrimination.  At  this  point, 
the  dog  starts  to  show  grossly  disturbed  behavior;  it  may  bite  its 
keeper,  refuse  food,  become  comatose,  etc. 

If  the  experiment  is  started  with  a  naive  dog  and  the  preliminary 
training  in  discrimination  is  omitted,  the  dog  does  not  go  crazy. 
If  you  start  with  a  dog  untrained  in  discrimination  and  present 
a  single  stimulus  object  (flipping  a  coin  to  decide  what  this  object 
shall  mean),  the  dog  has  to  guess  and  will  do  the  appropriate 
thing;  it  will  gamble  on  the  difference.  The  dog  cannot  toss  a 
coin,  but  it  settles,  in  general,  to  approximately  the  probabilities 
which  it  experiences.  If  the  stimulus  object  means  \  70  per  cent  of  the 
times  and  )'  30  per  cent  of  the  times,  the  dog  will  settle  to  guessing 
at  .V  70  per  cent  of  the  time  and  guessing  at  y  30  per  cent  of  the  time. 
This  is  not  the  ideal  course  which  the  sophisticated  gambler  would 
follow;  he,  of  course,  would  bet  on  x  100  per  cent  of  the  times  be- 
cause it  gives  more  frequently  the  positive  reinforcement. 

What  happens,  it  seems  to  me,  in  the  pathogenic  experiment  is 
that  the  experimenter  succeeds  in  communicating  to  the  dog  a 
message  about  the  contingency  patterns  in  which  it  is  to  find 
itself,  and  this  message  happens  to  be  an  untrue  message.  The 
dog  is  in  a  probabilistic  situation,  but  the  experimenter  has  con- 
vinced the  dog  that  it  is  in  a  discrimination  situation,  at  which 
point  very  severe  pathological  changes  start  to  appear. 


1 80  Information  Storage  and  Neural  Control 

These  are  the  situations  which,  in  our  work  on  schizophrenia, 
have  come  to  be  called  "double-binds."  These  may  now  be  defined 
very  simply  as  pathological  alterations  of  communication  at  the 
third  level. 

Let  me  illustrate  this  pathogenic  pattern,  or  perhaps  I  should 
say  broken  pattern,  rather  briefly  with  an  excerpt  from  a  book 
entitled  Mary  Poppins  (3).  This  is  an  English  children's  book  by 
P.  L.  Travers  about  an  English  nanny  Mary  Poppins.  She  has 
taken  the  two  children  to  a  little  old  gingerbread  shop  owned 
by  Mrs.  Corry,  a  tiny  old  woman  with  two  large  "sad"  daughters: 

"I  suppose  you've  come  for  some  gingerbread?" 

"That's  right,  Mrs.  Corry,"  said  Mary  Poppins  politely. 

"Good.  Have  Fannie  and  Annie  given  you  any?"  She  looked  at 
Jane  and  Michael  as  she  said  this. 

"No,  Mother,"  said  Miss  Fannie  meekly. 

"We  were  just  going  to,  Mother — -"  began  Miss  Annie  in  a 
frightened  whisper. 

At  that  Mrs.  Corry  drew  herself  up  to  her  full  height  and  regarded 
her  gigantic  daughters  furiously.  Then  she  said  in  a  soft,  fierce, 
terrifying  voice,  "Just  going  to?  Oh,  indeed!  That  is  very  interesting. 

And  who,  may  I  ask,  Annie,  gave  you  permission  to  give  away  my 
gingerbread — ?" 

"Nobody,  Mother.  And  I  didn't  give  it  away.  I  only  thought — " 

"You  only  thought!  That  is  very  kind  of  you.  But  I  will  thank  you 
not  to  think.  I  can  do  all  the  thinking  that  is  necessary  here!''  said 
Mrs.  Corry  in  her  soft,  terrible  voice.  Then  she  burst  into  a  harsh 
cackle  of  laughter. 

"Look  at  her!  Just  look  at  her!  Cowardy-custard !  Cry-baby!"  she 
shrieked,  pointing  her  knotty  finger  at  her  daughter. 

Jane  and  Michael  turned  and  saw  a  large  tear  coursing  down 
Miss  Annie's  huge,  sad  face,  but  they  did  not  say  anything,  for, 
in  spite  of  her  tininess,  Mrs.  Corry  made  them  feel  rather  small 
and  frightened  .   .   . 

In  this  episode  Mrs.  Corry  indicates  that  this  is  a  context  in 
which  to  have  given  gingerbread  to  the  children  would  be  re- 
warded and  not  to  have  given  gingerbread  might  be  punished. 


Patterns  of  Human  Behavior  181 

The  daughter  Annie  tries  to  alibi  for  not  giving  gingerbread,  and 
Mrs.  Corry  promptly  punishes  her.  This  is  not,  was  not,  that  sort 
of  context  at  all;  it  was  one  in  which  the  daughter  had  no  right 
to  give  away  gingerbread  and  was  wicked  to  even  think  of  doing  so. 

The  problem,  then,  for  every  individual  in  every  interchange 
is  to  maintain  an  up-to-the-minute  grasp  of  understanding  of  the 
state  of  the  contingency  patterns  between  himself  and  his  vis  a  vis. 
Consciously  or  unconsciously,  he  has  to  be  able  to  recognize  what 
sorts  of  trigrams,  or  more  complex  sequences,  should  characterize 
the  relationship  at  every  moment  and  to  act  in  terms  of  these 
recognitions.  The  individual  has  to  predict  from  what  occurred 
previously  which  pattern  is  appropriate  at  the  moment.  This  is 
what  we  call  understanding  between  persons.  Without  it  or  when 
such  understanding  is  traumatized  or  punished,  very  severe  patho- 
logical behavior  may  follow. 

But  such  understanding  is  only  possible  because  we  are  able  to 
predict,  to  guess  correctly  at  a  given  moment,  within  what  pattern 
we  are  operating  and  within  what  pattern  the  other  person  is 
operating.  Prediction  is  the  essence  of  the  matter,  and  it  is  at  this 
point  that  double-bind  theory  links  up  with  information  theory. 

Redundancy,  as  the  term  is  technically  used,  is  that  charac- 
teristic of  the  sequence  of  events  that  enables  an  observing  subject 
to  make  a  better  than  probable  guess  at  the  next  item  in  the 
sequence,  so  that  this  next  item,  when  it  actually  occurs,  does 
not  provide  100  per  cent  new  information.  It  is  rather  unfortunate 
that  the  word  redundancy  has  been  used  in  this  sense,  because 
coinfortable  communication  between  people  (we  may  even  say 
efficient  communication  between  people)  depends  entirely  upon 
such  ability  to  predict.  It  might  have  been  happier  to  describe 
the  phenomenon  of  redundancy  as  a  necessary  condition  of 
efficiency  rather  than  as  a  characteristic  excess  since  it  is  economical 
to  deal  with  patterns  rather  than  with  multiple  bits. 

It  is  now  appropriate  to  think  for  a  moment  about  the  place 
in  human  natural  history  of  patterns  of  this  order.  Bavelas  (per- 
sonal communication)  has  shown  that  these  orders  of  learning  are 
singularly  difficult  to  modify  when  erroneous  learning  has  occurred. 
The  experimental  material  is  somewhat  as  follows:  The  subject 
is  presented  with  a  board  on  which  there  are  a  number  of  buttons 


1 82  Information  Storage  and  Neural  Control 

and  is  told  to  find  the  correct  way  to  press  these  buttons.  He  is 
told  that  when  he  presses  them  correctly,  a  bell  will  ring.  The 
subject  proceeds  to  press  buttons,  and  after  he  has  pressed,  say, 
fifty  buttons,  the  bell  rings.  The  experimenter  now  asks  him  if 
he  knows  how  to  do  it  and  if  he  will  do  it  again.  The  subject 
again  presses  buttons,  and  after  he  has  pressed  about  forty-five 
buttons,  the  bell  rings.  He  is  again  asked  to  repeat  the  task,  and 
this  time  after  about  forty  pressings  the  bell  rings.  The  subject 
is  doing  better  and  better.  When  the  subject  has  reduced  the 
number  of  pressings  to  about  twenty,  Bavelas  stops  the  experi- 
ment and  tells  him  that  there  is  no  connection  between  the  buttons 
and  the  bell,  that  the  bell  is  only  geared  probabilistically  to  a 
hypothetical  learning  curve. 

The  subject  will  then  look  Bavelas  firmly  in  the  eye  and  tell 
him  he  is  lying.  This,  of  course,  is  true  except  that  the  subject  is 
wrong  as  to  which  lie  he  is  attributing  to  Bavelas.  The  truth  is 
that  Bavelas  was  lying  initially  when  he  told  the  subject  there 
was  a  connection  between  the  bell  and  the  buttons,  but  he  is 
now  telling  the  truth.  The  subject,  however,  cannot  be  convinced 
of  this  and  will  reassert  his  theory  of  the  interrelation  between 
the  buttons,  usually  quite  a  complex  theory  with  a  lot  of  paren- 
thetical cautions  in  it:  "At  this  part  of  the  sequence  you  should 
not  go  too  fast";  "If  you  go  too  fast,  you  can  only  correct  it  by 
going  back  to  the  beginning  of  the  sequence,"  etc.  The  subject 
is  perfectly  certain  that  what  he  was  doing  was  related  to  the 
theory  he  built  up  and  that  his  experience  has  validated  this 
theory.  He  has  been,  after  all,  well  reinforced  in  this  belief  by  his 
steadily  increasing  success. 

There  is,  I  understand  from  Bavelas,  only  one  way  of  dis- 
illusioning the  subject  in  regard  to  his  theories  about  these  buttons. 
This  is  by  asking  him  to  perform  the  experiment  upon  a  second  sub- 
ject. As  he  does  this  and  sees  the  second  subject  develop  analogous 
but  dissimilar  illusions,  he  realizes  the  nature  of  the  situation  and 
the  process  through  which  he  has  gone. 

The  point  I  want  to  make  is  that  these  impressions,  illusions 
at  the  third  level,  are  held  very  deeply  and  are  exceedingly  difficult 
to  disturb;  the  same  must  be  true  of  knowledge  and  wisdom  at 
the  third  level.  I  have  mentioned  that  the  subject  trained  in  an 


Patterns  oj  Human  Behavior  183 

instrumental  philosophy  will,  of  course,  encounter  a  universe  which 
will  seem  to  him  to  validate  that  philosophy  and  that  a  subject 
trained  in  a  Pavlovian  universe  will  correspondingly,  as  it  seems 
to  him,  encounter  a  universe  in  which  the  Pavlovian  philosophy 
is  appropriate. 

It  is  a  formal  characteristic  of  this  level  that  opinions  about  it 
are,  in  general,  self-validating,  and,  of  course,  a  great  deal  of  the 
difficulty  in  psychotherapy  occurs  in  wrestling  with  this  particular 
fact.  The  interchange  between  therapist  and  patient  always  seems 
to  the  patient  to  validate  those  third  level  premises  with  which 
he  entered  the  therapy  room.  This  is  the  phenomenon  of  "trans- 
ference." The  therapist's  task  is  to  endeavor  to  break  up  those 
learnings  at  the  third  level  for  which  the  patient  has  been  deeply 
reinforced  in  the  past,  those  learnings  of  which  he  is,  in  general, 
almost  unconscious  and  which  necessarily  have  this  characteristic 
of  being  self  validating  ...  no  mean  task. 

At  this  point,  we  are  approaching  a  fourth  learning  level:  the 
problem  of  changes  at  the  third  level.  I  have  said  that  this  is  no 
mean  task  for  the  therapist,  and  I  think  it  is  worth  noting  that 
this  is  a  task  in  which  considerable  meanness,  in  another  sense  of 
the  word,  may  be  a  necessary  ingredient.  To  change  one's  basic 
premises  at  this  third  level  is  always  in  some  degree  painful  and 
always  difficult,  and  the  therapist  may  be  compared,  if  you  will, 
to  Mrs.  Clorry.  He  must,  of  necessity,  put  the  patient  in  the  wrong 
at  the  third  level.  It  is  therefore  essential  that  psychotherapy  shall 
be  double-binding  in  the  sense  in  which  the  word  is  defined  here. 
Mrs.  Corry  is  pathogenic  because  she  goes  on  doing  this  without 
mercy.  The  therapist  is  curative  insofar  as  he  does  it  with  wisdom 
and  with  consistency.  After  all,  Mrs.  Corry,  is  inconsistent  even 
in  her  inconsistency  and  can,  therefore,  always  surprise  her  victim; 
whereas,  the  therapist  must  instruct  his  patient,  albeit  by  implicit 
methods,  so  that  new  expectations  may  replace  the  old  and  may 
be  rewarded. 

This  problem  of  fourth-level  learning,  of  changes  at  the  third 
level,  is  a  necessary  part  of  human  life.  It  obtains  in  courtship;  it 
obtains  in  initiation;  it  obtains  in  psychotherapy;  it  obtains,  in 
fact,  wherever  important  reconstruction  of  relationship  must  occur. 
We  know  very  little  about  such  phenomena,  and   I  cannot  tell 


1 84  Information  Storage  and  Neural  Control 

you  much  today.  Certain  aspects,  however,  are  conspicuous 
enough  to  be  worth  mentioning.  First,  it  seems  that  such  deep 
clianges  and  the  processes  by  which  they  occur  are  almost  in- 
variably cloaked  with  unconsciousness  and  with  amnesia.  The 
ability  of  any  couple  to  tell  you  what  it  really  was  that  they  went 
through  in  courtship  is  approximately  zero.  They  can  tell  you 
dates,  times,  and  places.  They  may  be  able  to  identify  a  single 
striking  episode,  something  that  he  did  or  she  did  which  struck 
the  other  with  a  moment's  flash;  but,  in  general,  such  processes  are 
not  subject  to  recall  and  have  not  been  investigated.  Wliile  there 
is  a  great  deal  of  fantasy  about  courtship,  there  is,  as  a  matter  of 
fact,  no  recorded  data  regarding  it  in  any  culture  of  the  world. 
Similarly,  the  patient  and  the  therapist  are  both  virtually 
unable  to  tell  you  what  happened  that  led  to  psychotherapeutic 
change.  Theories  are  many;  fantasies  are  many;  recipes  are  many 
and  are  always  unsatisfactory.  It  is  not  too  much  to  say  that  this 
is  a  region  of  almost  total  scientific  ignorance.  I  believe,  however, 
that  it  has  to  be  analyzed,  has  to  be  studied,  and  will  be  studied 
in  the  next  twenty  years,  and  that  in  this  study,  the  branch  of 
information  theory  dealing  with  patterns  of  patterns,  redundancies 
about  redundancies,  will  be  a  central  tool. 

REFERENCES 

1.  Attneave,  Fred:  Applications  of  Information   Theory  to  Psychology,  New 

York,  Henry  Holt  and  Co.,  1959. 

2.  Bateson,  Gregory:  Morale  and  national  character,  in,  Civilian  Morale, 

2nd  Yearbook  of  the  Society  for  the  Psychological  Study  of  Social 
Issues,  edited  by  Goodwin  Watson,  New  York,  Houghton  Mifflin 
and  Co.,  1942,  p.  71. 

3.  Travers,   P.   L.:  Mary  Poppins,  New  York,   Reynal  and  Hitchcock, 

1934,  p.  121. 

DISCUSSION  OF  CHAPTER  VIII 

Herman  Blustein  (Chicago,  Illinois):  Doesn't  an  adequate 
communication  system  actually  preclude  the  knowledge  of  the 
rules  of  the  game  by  both  communicators  and  receptors  of  the 
communication  system? 


Patterns  of  Human  Behavior  1 85 

Gregory  Bateson  (Palo  Alto,  California):  I  think  there  are 
two  questions  combined  here.  One  concerns  the  case  where  com- 
munication is  going  along  "smoothly,"  as  I  called  it  earlier.  Is  a 
knowledge  of  the  rules  necessary?  Obviously  it  is  not.  The  rules 
are  provided;  they  are  built  in,  and  that  is  all  we  ask.  To  be  able 
to  cough  them  up  and  inspect  them  is  not  necessary.  So  far  I 
think  we  are  in  agreement;  however,  behind  this  is  the  question 
of  "rules  about  rules"  and  "rules  about  rules  about  rules."  I  think 
we  always  walk  around  wishing  to  be  in  the  state  of  "things  going 
along  smoothly,"  and  wishing,  therefore,  not  to  turn  over  all  this 
disturbing  stuff,  i.e.,  unwilling  to  raise  questions  about  the  rules.  We 
may  be  forced  to  do  this  when  things  go  wrong.  We  want  some 
of  the  rules  to  be  steady.  We  hope  we  can  operate  on  the  common 
assumptions  of  the  culture  which  we  share,  and  we  hope  to  try 
to  get  mutual  understanding  at  that  level.  If  we  cannot,  we  may 
be  pushed  into  reexamining  blemishes  of  the  culture,  but  this  will 
be  painful  and  always  at  an  upper  level  which  we  do  not  want 
to  disturb. 

Yasuhiko  Taketomo  (New  York,  New  York):  In  the  com- 
ments on  expectation  in  relationships,  were  you  referring  to 
something  like  role-taking  in  psychiatric  communication? 

Bateson:  I  was  doing  so  in  a  terribly  loose  context.  I  think  the 
evidence  is  going  to  come  from  such  work  as  that  of  Birdwhistell, 
studying  expressive  movement  and  expressive  posture.  This  is  not 
a  study  of  those  movements  which  are  quasi-linguistic,  such  as 
thumbing  a  ride,  but  the  study  of  those  much  less  conscious  and 
much  less  voluntary  elements  in  our  movements.  I  think  it  is 
going  to  appear  that,  while  we  talk  with  words,  mathematical 
equations,  and  other  highly  sophisticated  devices,  we  are,  in  fact, 
either  leaning  forward  on  the  rostrum  or  scratching  in  our  pockets 
looking  for  a  cigarette  or  some  other  object.  All  these  movements 
can  be  interpreted  and  handled,  and  are  going  to  be  interpreted 
and  handled,  at  this  third  level  as  sequence  markers  or  signals 
about  the  relationship.  But  when  I  lean  forward  or  draw  back 
from  you,  these  movements  indicate  to  you  whether  I  want  you 
to  come  forward  and  shoot  me  with  questions  or  whether  you 
should  beware  of  my  defenses,  and  so  on.  I  think  the  implementa- 
tion is  going  to  come  from  this  area  and  from  the  field  of  micro- 


1 86  Information  Storage  and  Neural  Control 

linguistics  in  which  modulations  of  loudness  of  voice,  emphasis, 
rasp,  etc.,  are  going  to  be  the  key  signals. 

Myron  F.  Weiner  (Dallas,  Texas) :  Assuming  that  somebody 
comes  to  you  because  he  has  had  a  breakdown  of  relationships 
because,  in  turn,  his  metacommunications  or  metaconcepts,  or 
what  he  expects  of  the  world,  are  somewhat  different  from  what 
he  says  he  expects,  do  you  think  it  would  be  of  some  value  in 
correcting  his  behavior  to  bring  to  his  consciousness  the  fact  that 
his  metaperception  is  quite  different  from  what  he  thinks  he 
perceives? 

Bateson:  This  is  a  problem  of  technique  of  psychotherapy. 
Let  nie  reword  Dr.  Weiner's  question:  "Does  it  help  to  give  him 
insight?"  I  would  not  agree  that  insight  is  necessary  and  sufficient. 
It  may  be  sufficient,  but  I  do  not  think  it  is  necessary.  I  think  that 
experiences  of  effect  in  communication  at  these  levels  probably 
are  therapeutically  necessary,  but  I  do  not  think  it  is  necessary 
that  these  communications  take  the  form  of  providing  a  guide  to 
conscious  insight  into  the  mechanics  of  these  levels.  Surely  it 
never  happens.  I  do  not  know  of  any  school  of  psychotherapy  that, 
as  yet,  has  enough  language  for  talking  about  these  levels  to  even 
attempt  to  give  insight  at  these  levels.  We  just  do  not  have  the 
language  to  give  that  insight.  I  think  we  know  that  psychotherapy 
occurs;  but  since  it  occurs  in  a  culture  which  does  not  have  sufficient 
language  to  say  what  is  happening,  it  follows  that  linguistic  insight 
is  certainly  not  necessary. 

W.  R.  Beavers  (Dallas,  Texas) :  These  remarks  about  the  con- 
text, or  metalanguage,  reminded  me  of  Bion  and  his  primitive 
group  concepts.  He  felt  that  in  working  with  groups,  he  saw  and 
began  to  communicate  with  them,  not  about  their  intrapsychic 
assumptions,  but  about  the  primitive  group  assumptions.  As  I 
recall,  there  was  an  assumption  of  the  fight-or-flight  and  of  the 
pairing  group.  This  sounds  very  much  like  your  ideas  concerning 
the  basic  mammalian  assumptions  underneath  that  which  is  con- 
ventional conversation. 

Bateson:  I  am  slightly  familiar  with  the  ideas,  but  have  not 
worked  with  them. 


PART  III— NEUROPHYSIOLOGICAL  ASPECTS  OF 
INFORMATION  STORAGE  AND  TRANSFER 

Moderator:  Hebbel  E.  Hoff,  M.D.,  Ph.D. 


CHAPTER 
IX 

INFORMATION  STORAGE  IN  NERVE  CELLS 

Frank  Morrell,  M.D. 


Bi 


►EHAVIORAL  observations  have  generally  supported  the 
notion  that  (aside  from  genetic  information)  there  are  two  quali- 
tatively different  forms  of  information  storage  in  the  nervous 
system.  So-called  "recent"  memory  is  made  of  particularly  labile 
stuff.  A  cerebral  concussion  produces  an  amnesia  not  only  for  the 
injury  itself  but  also  for  the  events  immediately  leading  up  to  the 
injury,  a  circumstance  about  which  many  lawyers  are  painfully 
aware.  The  impact  of  experience  requires  time  for  fixation.  If 
neural  activity  is  interfered  with  during  this  fixation  or  con- 
solidation period  by  electro-shock  (13,  14,  53,  54),  trauma  (50), 
severe  cold  (44),  or  rapid  induction  of  ether  or  barbiturate  anes- 
thesia (1),  subsequent  recall  of  the  experience  may  be  seriously 
compromised.  For  example,  Duncan  (13)  and  Gerard  (14,  44) 
have  shown  that  rats  or  hamsters  trained  in  an  avoidance  situation 
or  in  maze-learning  have  a  normal  learning  curve  if  a  maximal 
electro-shock  is  delivered  four  hours  after  each  training  session. 
If  the  shock  follows  the  training  by  one  hour  there  is  slight  de- 
terioration; at  a  fifteen  minute  interval  there  is  major  interference 
with  retention  and  at  five  minutes  or  less,  learning  is  completely 
prevented.  Acute  anoxia  introduced  at  similar  time  intervals  has 
the  same  effect  (53).  Since  all  of  the  agencies  known  to  produce 
amnesia  or  loss  of  recent  memory  are  also  known  to  alter  electrical 
activity  of  the  central  nervous  system,  the  mechanisms  subserving 
the  initial  stage  of  memory  recording  are  inferred  to  be  electrical 
in  nature.  Other  evidence  supporting  a  clear  distinction  between 
short-term  and  durable  memory  mechanisms  is  the  finding  that 

189 


1 90  Information  Storage  and  Neural  Control 

focal  epileptogenic  lesions  prevent  new  learning,  i.e.,  impair 
memory  recording,  but  do  not  disturb  behavior  learned  before 
establishment  of  the  epileptic  lesion   (31,  34,  35,  42,  51). 

A  limiting  case  in  the  requirement  for  a  finite  fixation  time  is 
the  classical  example  of  one-trial  learning.  However,  even  in  this 
instance,  it  has  been  supposed  (Hebb)  (24)  that  the  neural  con- 
sequences of  the  single  experience  persist  in  the  form  of  rever- 
berating impulses  for  a  considerable  time  after  the  environmental 
signal  has  ceased.  x\lthough  all  or  none  impulses  circulating  in 
closed  neuronal  chains  represent  one  possible  mechanism  for  the 
initial  imprinting  or  short-term  memory  the  actual  kind  or  kinds 
of  electrical  activity  involved  remain  unknown.  In  fact,  there  is 
nothing  in  the  experimental  evidence  concerned  with  manipula- 
tion of  the  consolidation  process  which  affords  compelling  proof 
that  consolidation  depends  upon  reverberating  impulses  of  the 
all  or  none  type  (41).  Other  kinds  of  electrical  activity,  that  is, 
other  than  the  classical  axon  spike  or  even  the  conventionally 
recorded  EEG  may  well  be  equally  important. 

I  should  like  to  present  some  evidence  which  suggests  that 
cortical  steady  potential  gradients  may  have  a  determining  in- 
fluence in  the  process  wherein  a  sequence  of  impinging  impulses 
is  transformed  into  structural  change  in  the  nervous  system.  This 
portion  of  the  paper,  therefore,  is  concerned  with  the  initial  or 
electro-sensitive  stage  of  memory  recording. 

Significant  shifts  of  the  cortical  steady  potential  have  been  shown 
to  occur  consequent  to  stimulation  of  peripheral  receptors  (19)  as 
well  as  when  stimulating  electrodes  are  applied  directly  to  brain 
substance  (2,  7,  8,  9,  16,  17,  18).  Some  years  ago,  we  found  (37) 
that  the  surface  negative  DC  shift  resulting  from  low  frequency 
stimulation  of  nucleus  centrum  medianum  in  the  thalamus  would 
appear  as  a  conditioned  response  to  a  pure  tone  after  thirty  to 
forty  paired  trials. 

Figure  1  illustrates  the  first  paired  trial.  The  tone  elicited  no 
response.  Upon  onset  of  the  thalamic  stimulus,  a  pronounced 
negative  shift  of  the  base  line  of  the  EEG  occurred,  which  was 
confined  to  the  hemisphere  ipsilateral  to  the  stimulated  site. 
After  about  forty  paired  trials  (Fig.  2)  a  similar  DC  shift  was 
regularly  induced  by  the  tone  alone.  Note  particularly  that  this 


Information  Storage  in  Nerve  Cells 


191 


RC 
LC 


Mm 


v/-^^^v^MVUWv^ 


Tone 


'^^^-h^^ 


t  Stim. 


■U  .J 


B 


I  stim.  off 

Fig.  1.  Initial  trial  in  which  a  low  intensity  500  cycle  per  second  tone  lasting 
ten  seconds  is  paired  with  four  per  second  shocks  (6  volts,  1  millisecond  duration) 
delivered  through  bipolar  stimulating  electrodes  in  the  left  centre  median.  The 
tracing  is  from  an  unesthetized  rabbit.  Electrodes  derived  from  the  somato- 
sensory regions  of  both  hemispheres  and  recorded  monopolarly  to  a  reference 
on  the  pinna.  A  and  B  are  a  continuous  sequence,  (A)  indicating  the  pronounced 
negative  shift  witli  slight  after-positivity  on  the  onset  of  thalamic  stimulation 
and  (B)  the  reversal  of  steady  potential  shift  at  the  cessation  of  thalamic  stimu- 
lation. Note  that  the  steady  potential  change  is  limited  to  the  ipsilateral  hemi- 
sphere. Calibration:  50  microvolts  and  one  second  (37). 

R   C  r 


L  C 


•Jr,    V^/.V 


^>^^^.-^/V■•f'^^^ '^V^  ^  .  , 


^YKr}r^f'r,^\-J^r^\j'.  --r ' 


Tone 


•/  V  ../I 


Fig.  2.  Same  experiment  as  in  Figure  1.  After  forty  trials  the  tone  alone  elicited 
the  same  ipsilateral  negative-positive  steady  potential  shift.  This  gradually  dis- 
appeared over  a  series  of  sLx  unreinforced  trials  but  was  restored  by  a  single 
subsequent  reinforcement  witli  thalamic  stimulation.  Calibration:  50  microvolts 

and  one  second  (37). 


1 92  Information  Storage  and  Neural  Control 

conditioned  DC  shift  was  also  restricted  to  the  previously  stimu- 
lated hemisphere  although  the  tone  was  presented  to  both  ears 
equally  in  open  field  conditions. 

In  addition  to  such  direct  conditioning  of  a  cerebral  electrical 
event,  the  increasing  availability  of  DC  amplifiers  made  it  possible 
for  Rusinov  (49)  and,  more  recently,  Rowland  (47)  to  identify 
steady  potential  shifts  occurring  regularly  in  the  course  of  classical 
behavioral  training.  Rusinov  (48)  also  discovered  a  most  intriguing 
behavioral  eff'ect  when  low-level  surface  positive  polarizing  cur- 
rents were  applied  to  a  part  of  the  motor  cortex.  The  current  levels 
employed  were  sub-threshold  with  respect  to  direct  production 
of  limb  movement.  But  during  the  period  of  current  flow  (and 
for  some  minutes  afterward)  any  ambient  sound,  light  or  touch 
would  produce  the  limb  movement  to  be  expected  from  adequate 
(supra-threshold)  stimulation  of  the  motor  area  to  which  the 
current  was  applied.  Rusinov  felt  that  the  anodal  polarization 
produced  a  "dominant  focus"  of  excitation  which  facilitated  the 
development  of  a  temporary  connection  between,  for  example, 
the  auditory  and  motor  systems. 

We  have  been  able  to  confirm  the  Rusinov  experiment  in  our 
own  laboratory  and,  in  addition,  have  made  some  observations 
on  the  activity  of  single  nerve  cells  in  such  polarized  regions  (40). 
Single  cells  in  motor  cortex  did  not  respond  to  acoustic  stimulation 
before  polarization  (Fig.  3A).  During  the  passage  of  anodal  cur- 
rent (10  microamps)  cells  of  several  different  types  (Figs.  3B,  C 
and  D)  were  easily  triggered  by  the  same  acoustic  signal.  Since 
we  were  interested  in  mechanisms  for  information  storage  we  per- 
formed the  experiment  in  a  slightly  diff'erent  way,  a  way  which 
allowed  observation  of  a  selective  sensitivity  with  respect  to  signals 
differing  in  their  history  of  exposure  to  polarizing  current.  A  group 
of  stimuli  was  chosen  and  all  members  of  the  group  were  presented 
repeatedly  to  the  animal  until  habituation  (as  judged  by  lack  of 
behavioral  or  EEG  response)  was  complete.  A  polarizing  electrode 
together  with  a  fine  microelectrodewas  placed  on  the  motor  cortex 
and  the  current  was  turned  on.  One  member  of  the  previously 
habituated  stimulus  group  was  selected  (in  this  case  a  200  cycle 
per  second  tone)  and  was  presented  to  the  animal  about  thirty 
times  in  the  course  of  forty-five  minutes.  A  burst  of  unit  activity 


Information  Storage  in  Nerve  Cells 


193 


B 


Fig.  3.  Patterns  of  response  in  single  units  to  an  acoustic  stimulus.  Duration  of 
the  tone  of  200  cycles  per  second  is  indicated  by  the  two  upward  deflections  in 
the  second  channel  of  the  oscilloscope.  Before  polarization  (A)  there  was  no 
effect  on  the  discharge  frequency  of  a  unit  in  motor  cortex.  During  polarization 
responses  to  sound  appeared  either  in  the  form  of  a  single  high  frequency  burst  (B), 
a  sudden  cessation  of  firing  (C),  or  high-frequency  bursts  at  the  "on"  and  at 
the  "oflf"  of  the  tone  (D).  Calibration:  5  millivolts  and  one  second  (40). 


194 


Information  Storage  and  Neural  Control 

A  B 


Fig.  4.  "Generalization  and  differentiation"  in  single  unit  responses.  During  the 
passage  of  anodal  current  (A  &  B)  the  critical  tone  (A)  and  a  single  presentation 
of  the  indifferent  tone  (B)  were  equally  effective  in  provoking  high  frequency 
bursts.  Twenty  minutes  after  cessation  of  current  flow  (C  &  D)  the  critical  tone 
(C)  continued  to  elicit  the  response  while  the  indifferent  tone  (D)  did  not.  Forty 
minutes  after  discontinuing  polarization  (E  &  F)  neither  signal  produced  any 
change  in  unit  discharge  frequency.  Calibration:  2  millivolts  and  500  milli- 
seconds (40). 


occurred  with  each  stimulation  (Fig.  4A).  Another  member  of 
the  habituated  stimulus  group  (500  cycle  per  second  tone)  was 
presented  once  in  the  polarization  period  and  also  elicited  unit 
discharge  (Fig.  4B).  The  current  was  then  discontinued  and  in 
the  following  twenty  minutes  the  200  cycle  per  second  tone  con- 
sistently  provoked  unit  activity  (Fig.   4G)    while   the   500  cycle 


Information  Storage  in  Nerve  Cells 


195 


Fig.  5.  Conditioning  of  a  rhythmic  burst  response  to  a  single  flash.  Anodal 
polarization  was  applied  to  the  visual  receiving  area.  Single  flash  elicited  a 
single  burst  in  a  quiescent  (A)  and  in  a  randomly  firing  cell  (B).  Three  per 
second  stroboscopic  stimulation  (C)  produced  driving  of  unit  discharge  at  tliat 
frequency.  A  single  flash  (D)  delivered  thirty  seconds  after  termination  of  the 
rhythmic  stimulus  resulted  in  repetitive  unit  discharge  at  about  three  per  second. 
Unit  potentials  are  seen  in  the  upper  channel  of  tlie  oscilloscope;  stimulus  artifacts 
in  the  lower  channel.  Amplitude  calibration:  2  millivolts.  Time  calibration: 
500  milliseconds  (A  &  B)  and  one  second  (C  &  D)  (40). 


per  second  tone  (Fig.  4D)  invariably  failed  to  induce  a  change  in 
the  pattern  of  unit  firing.  About  forty  minutes  after  cessation  of 
polarization  neither  signal  was  effective  (Figs.  4E  and  F).  Under 
these  circumstances  it  seemed  evident  that  the  polarized  cell 
population  had  retained  some  stipulation  of  signal  characteristics 
so  that  for  a  brief  period  in  the  post-polarization  interval  the  cells 
behaved  differentially  with  respect  to  the  two  signals. 

Short-term  storage  of  a  temporal  pattern  has  also  been  observed 
in  cells  of  the  visual  cortex.  Figure  5A  illustrates  the  response  of 
a  quiescent  cell  to  a  brief  flash  of  light,  (the  flash  artifact  is  recorded 
on  the  second  beam  of  the  oscilloscope).  Figure  5B  shows  a  similar 
burst  response  in  a  spontaneously  active  cell.  During  anodal 
polarization  it  was  extraordinarily  easy  to  "drive"  such  cells  with 


196 


Information  Storage  and  Neural  Contiol 


c 
en 


a; 
(/) 

c 
o 

Q. 

cr 

o 
E 

.d 
cr 


o 


00   r 


B      80    - 


60    - 


40    - 


20 


05     I  10  20  30 

Time  in  minutes  after  "conditioning" 
train  of    3 /sec  flicker 

Fig.   6.   Time  course  for  "decay"  of  conditioned  rhytJimic  response  of  single 

cortical  unit. 


low  frequency  intermittent  light  (Fig.  5C).  After  a  few  minutes 
of  stimulation  the  three  per  second  flash  was  discontinued  and 
thirty  seconds  later  a  single  flash  resulted  in  a  series  of  bursts 
having  a  three  per  second  frequency  (Fig.  5D).  Single  flashes 
delivered  at  intervals  longer  than  thirty  seconds  were  less  and 
less  likely  to  provoke  such  a  rhythmic  response  but  occasional 
rhythmic  responses  to  single  flash  were  noted  as  long  as  twenty 
minutes  after  the  end  of  the  conditioning  train  (Fig.  6).  This 
seems  a  particularly  clear  illustration  of  the  capacity  of  the  polar- 
ized cells  to  retain  some  representation  of  an  imposed  stimulus 
pattern  for  a  relatively  long  period  of  time.  Indeed  the  order  of 
magnitude  of  this  time  interval  is  itself  significant.  It  correlates 
well  with  the  data  of  Gerard  (14),  Duncan  (13)  and  others  (44, 
53,  54)  on  the  abolition  of  learned  responses  consequent  to  massive 
electro-shock  delivered  at  various  intervals  following  the  training 
session. 


Information  Storage  in  Nerve  Cells 


197 


CJ-2       -\^\^\^\^ 
A 

R»E. 


(((( 

Tet. 


-v^v 


Post. 


S   R   T 


S   R  T 


J 


/v^ 


^A^r- 


Pre.  Tet.  Post. 

Fig.  7.  Oscillographic  tracings  from  a  deeply  anesthetized  (A)  and  an  unanes- 
thetized  (B)  cat.  Derivations  are  from  implanted  bipolar  electrodes  (R)  arranged 
as  indicated  in  the  diagram.  Recording  electrodes  (R)  are  situated  between  the 
stimulating  electrodes  (S)  on  one  side  and  the  tetanizing  electrode  pair  (T)  on 
the  other.  Explanation  in  text.  Calibration:  100  microvolts  and  100  milliseconds. 
Negativity  at  the  recording  electrode  produces  a  downward  deflection  of  the 
beam  in  this  and  the  two  succeeding  figures.   (Chow,  K.  L.  and  Dewson,  J.: 

unpublished  data.) 


Short-term  storage  of  a  temporal  pattern  may  also  be  demon- 
strated in  another  way.  Following  a  technique  originally  described 
by  Roitbak  (45),  Doctors  K.  L.  Chow  and  James  Dewson  (10) 
have  used  tetanization  of  a  local  cortical  region  to  produce  an 
effect  similar  to  that  of  the  ''dominant  focus."  Three  pairs  of 
implanted  electrodes  were  arranged  as  shown  in  the  diagram  of 
Figure  7  so  that  the  stimulating  pair  was  at  one  end  of  the  array, 
the  tetanizing  pair  at  the  other  end,  and  the  recording  electrodes 
in  between. 

In  the  deeply  anesthetized  animal  (Fig.  7A)  nine  per  second 
shocks  delivered  before  tetanization  produced  only  a  small  direct 
cortical  response  (DCR)  arising  almost  immediately  out  of  the 
shock  artifact.  The  nine  per  second  stimulus  was  continued  through- 
out the  fifty  per  second  tetanus  and  into  the  post-tetanization 
period.  Immediately  following  the  tetanus  the  response  to  the 
shock  was  altered  and  could  be  distinguished  from  the  DCR  by 


198 


Information  Storage  and  Neural  Control 


■,'<f    " 


Pre. 


Tet. 


3cps 


[c  d 

Fig.  8.  Influence  of  tetanization  across  a  cortical  section.  Cat  is  unanesthetized, 
awake,  and  carries  implanted  bipolar  electrodes  arranged  so  that  recording  (R) 
and  stimulating  (S)  pairs  are  within  an  island  of  neuronally  isolated  cortical 
tissue  while  the  tetanizing  (T)  pair  is  on  the  intact  cortex  outside  the  isolated 
zone.  Pre-  and  post-tetanization  voltage  is  the  same  as  that  used  in  the  tracing 
labeled  "3  cps."  Stimulating  voltage  is  then  reduced  50%  (A)  and  70%  (B). 
The  pre-tetanization  voltage  is  then  resumed  and  then  stimulation  stopped  (D). 
Further  explanation  in  text.  Calibration:  100  microvolts  and  100  milliseconds. 
(Chow,  K.  L.  and  Dewson,  J.:  unpublished  data.) 


a  longer  latency  and  a  hump  on  the  descending  limb.  The  response 
had  never  occurred  prior  to  tetanization  and  was  clearly  locked 
to  the  stimulus  frequency.  However,  in  the  waking  animal  (Fig.  7B) 
where  spontaneous  background  rhythms  were  more  prominent, 
phase  locking  was  much  less  precise  and  it  was  difficult  to  be  sure 
of  a  response  truly  related  to  the  stimulus  frequency. 

In  order  to  determine  whether  the  effect  of  tetanization  on  the 
responsiveness  of  cells  at  the  recording  site  was  mediated  through 
synaptic  connections  rather  than  by  an  electrotonic  or  field  effect, 
a  neuronally  isolated  cortical  slab  was  prepared.  The  blood  supply 
to  the  slab  was  preserved.  The  electrodes  were  arranged  so  that 
the  stimulating  and  recording  pairs  were  within  the  isolated  zone 
but  the  tetanizing  pair  was  placed  outside  (Fig.  8). 


Information  Storage  in  Nerve  Cells  199 

Background  electrical  activity  in  such  neuronally  isolated  regions 
is  much  less  prominent  than  in  normal  cortex  even  when  the 
animal  is  awake.  The  pattern  more  closely  approximates  the 
deeply  anesthetized  state.  The  cells  perhaps  are  not  so  busy  doing 
other  things  and  are  therefore  more  a\^ailable  for  recruitment  by 
an  active  pacemaker.  Thus  it  is  not  surprising"  that  the  responses 
obtained  are  similar  to  those  of  Figure  7A.  Nine  per  second  shocks 
before  tetanization  produced  only  a  small  DCR.  After  tetanization 
a  new  response  emerged  which  was  tightly  locked  to  the  stimulus 
frequency.  Changing  the  stimulus  frequency  to  three  per  second 
perturbed  the  system  somewhat  but  clear  responses  did  appear 
at  the  new  frequency.  When  the  stimulus  voltage  was  reduced 
by  about  half  (Fig.  8A)  the  responses  were  also  reduced  and  were 
less  reliably  evoked.  Further  reduction  of  stimulus  voltage  abolished 
all  response  (Fig.  8B)  but  returning  to  the  pre-tetanization  voltage 
restored  the  stimulus-locked  response  in  full  (Fig.  8C).  When  the 
stiinulus  was  turned  off  (Fig.  8D)  there  was  no  trace  of  an  after- 
discharge. 

The  variations  in  stimulus  voltage  just  described  indicate  that 
the  response  arising  after  tetanization  depends  upon  the  specific 
stimulus.  It  cannot  be  explained  as  an  apparent  coherence  re- 
sulting from  emergence  of  a  background  rhythm,  paroxysmal  or 
not,  having  frecjuency  characteristics  close  to  those  used  to  stimu- 
late. Finally,  as  suggested  above,  this  experiment  supports  the 
notion  that  the  tetanization  effect  is  transmitted  to  the  cells  at 
the  recording  site  by  non-synaptic  means. 

Figure  9  illustrates  an  experiment  similar  to  that  just  described. 
Electrodes  were  arranged  exactly  as  in  the  diagram  of  Figure  8 
within  and  beside  a  neuronally  isolated  cortical  region.  Direct 
your  attention  particularly  to  Figure  9B.  In  this  experiinent  the 
pre-tetanization  stimulus  frecjuency  was  nine  per  second.  After  the 
tetanus  an  altered  but  stimulus-locked  response  was  evident  at 
nine  per  second.  At  that  point  during  a  period  coincident  with 
the  second  tracing  the  experimenter  slowly  changed  the  stimulus 
frequency  from  nine  to  three  per  second.  The  response  pattern 
reflects  the  changing  temporal  character  of  the  signal  and  locks 
in  at  three  per  second  (third  tracing).  But  something  else  has 
happened  as  well.  Between  the  major  three  per  second  deflections 


200 


BC5I 


Information  Storage  and  Neural  Control 


pre. 


5  s. 


i".-^\^ 


let. 


\/v/\Ay\,^r^iPw^/^\Ar. 


9cps 


V^/v 


3cps 

Fig.  9.  Same  preparation  and  same  electrode  arrangement  as  in  Figure  8.  The 
relevant  section  is  labeled  "B."  A  pre-  and  immediate  post-tetanization  stimulus 
frequency  of  nine  per  second  was  shifted  to  three  per  second.  As  the  response 
"locks"  in  at  three  per  second  (last  tracing)  a  series  of  inter-stimulus,  regularly 
spaced  waves  appear  which  seem  to  recapitulate  the  nine  per  second  rhythm 
For  discussion,  see  text.  Calibration:  100  microvolts  and  100  milliseconds. 
(Chow,  K.  L.  and  Dewson,  J.:  unpublished  data.) 


are  evenly  spaced  smaller  potentials  which  seem  to  recapitulate 
the  earlier  nine  per  second  sequence.  Is  this  the  electrical  expression 
of  the  neural  trace  of  nine  per  second?  Unfortunately  the  experi- 
ments are  still  too  few  to  afford  a  confident  answer. 

How,  it  may  be  asked,  can  these  experiments  with  tetanization 
be  linked  to  the  earlier  studies  on  polarizing"  currents?  Although 
measurements  of  the  steady  potential  were  not  made  in  these 
studies  the  inference  is  clear  from  the  fact  that  tetanization  is 
effective  across  a  solution  of  neural  continuity  that  an  alteration 
of  electrical  field  contours  has  occurred.  Furthermore  there  is 
abundant  data  from  many  laboratories  to  indicate  the  drastic 
alterations  of  steady  potential  gradients  accompanying  tetanic 
stimulation  of  this  kind  (7,  8,  9,  15,  16,  17,  18). 


Information  Storage  in  Nerve  Cells  201 

Now  it  may  be  legitimate  to  question  the  pertinence  of  these 
examples  of  cellular  memory  or  the  potential  gradients  with  which 
they  are  associated  to  anything  resembling"  a  functional  memory 
in  the  behaving  animal  or  man.  Nevertheless,  these  considerations 
have  led  to  a  direct  test  of  the  hypothesis  that  manipulation  of 
the  cortical  steady  potential  might  alter  a  learned  response. 
Particular  attention  was  paid  to  the  sign  and  orientation  of  the 
electrical  field  and  to  behavioral  evidence  bearing  upon  a  dis- 
tinction between  learning  and  memory. 

Twelve  mature  male  rabbits  weighing  between  two  and  three 
kilograms  were  used.  Prior  to  training,  the  animals  were  operated 
upon  under  nembutal  anesthesia  with  sterile  precautions  for  im- 
plantation of  six  stainless  steel  epidural  electrodes.  In  addition, 
nylon  plugs  were  inserted  bilaterally  over  motor  (both  fore  limb 
areas)  and  visual  cortex. 

Beginning  no  earlier  than  one  week  after  surgery,  the  animals 
were  trained  to  perform  a  conditioned  avoidance  response  which 
involved  lifting  of  the  right  forelimb.  The  unconditional  stimulus 
(UCS)  was  an  electric  shock  (sixty  per  second  square  waves)  of  a 
voltage  just  sufficient  to  cause  unconditional  limb  withdrawal  on 
all  stimulus  applications.  Stimuli  were  delivered  through  needle 
electrodes  in  the  foot  pad.  The  conditional  stimulus  (CS)  was  a 
flickering  light  at  eight  to  ten  flashes  per  second  from  a  Grass 
stroboscope  placed  ten  inches  in  front  of  the  animal  at  eye  level. 
Stimuli  were  presented  by  hand  and  the  UCS  was  omitted  if  the 
animal  responded  by  flexion  of  the  appropriate  limb  within  1.5 
seconds  after  flicker  was  turned  on.  The  animals  were  trained  to 
a  level  of  70-75  per  cent  correct  responses  in  twenty  trials  before 
polarization  was  begun.  The  low  level  of  performance  was  chosen 
intentionally  in  the  hope  that  we  might  be  able  to  observe  enhance- 
ment of  conditioning  as  well  as  depression  secondary  to  polarization. 

Capillary  pore  electrodes  (saline  bridge  and  silver-silver  chloride) 
were  placed  in  the  nylon  plugs  for  polarization.  A  constant  current 
stimulator  having  a  high  source  resistance  was  used.  Current  was 
continuously  monitored  and  maintained  at  about  10  microamps  per 
square  millimeter.  Pore  diameter  in  the  visual  cortex  was  5  milli- 
meters, thus  covering  more  than  half  of  the  visual  cortex  on  each 
side.  Pore  diameter  in  motor  cortex  was  2  millimeters.   Current 


202  Information  Storage  and  Neural  Control 

return  was  effected  through  a  saHne  soaked  pad  held  against  the 
palate  by  the  mouth  bar  of  the  head  holder.  Fine  needle  electrodes 
were  inserted  into  the  forelimb  flexors  to  record  the  electromyo- 
gram.  The  animals  were  restrained  with  the  head  fixed  and  the 
limbs  free.  Figure  10  illustrates  the  electrical  record  from  one  ani- 
mal at  the  beginning  of  the  paired  trials  (Fig.  1 OA)  when  there  was 
no  C;R  and  later  (Fig.  lOB)  when  the  ClR  was  present  allowing  the 
animal  to  avoid  the  shock.  Polarization  had  not  been  applied  at  all. 

These  animals  were  given  twenty  trials  a  day  every  day  except 
for  occasional  interruptions  which  may  be  seen  as  breaks  in  the 
graphs.  After  achieving  the  70-75  per  cent  criterion  polarization 
was  applied  (either  anodal  or  cathodal  to  either  motor  or  visual 
cortex)  throughout  the  training  session  on  a  given  day.  "Polariza- 
tion days"  and  "non-polarization  days''  as  well  as  type  and  site  of 
polarization  were  distributed  randomly  within  the  above  criterion 
limits.  As  an  additional  control  for  possible  sensory  effects  of  the 
constant  current,  training  sessions  were  carried  out  during  polariza- 
tion of  the  ear. 

Figure  11  shows  a  graph  of  the  learning  curve  in  a  typical 
animal.  Per  cent  correct  responses  in  each  day's  block  of  twenty 
trials  is  indicated  by  the  points  on  the  graph.  The  black  dots 
represent  days  on  which  no  polarization  was  given.  The  two  inter- 
ruptions in  the  graph  signify  that  training  sessions  were  omitted  for 
one  or  more  days.  The  legend  indicates  the  site  of  cathodal  polari- 
zation in  the  six  sessions  in  which  it  was  applied  in  this  animal. 

Cathodal  polarization  of  the  ear  and  of  the  motor  cortex  resulted 
in  no  apparent  change  in  the  "expected"  per  cent  response. 
Bilateral  cathodal  polarization  of  the  visual  area  produced  a 
striking  deficit  in  performance  for  the  entire  training  session 
during  which  the  current  was  applied.  Strangely  enough  the 
decreased  performance  was  also  apparent  the  following  day  when 
training  was  carried  out  without  polarization.  Note,  furthermore, 
that  a  break  in  the  sequence  of  daily  training  sessions  of  one  or 
more  days  was  almost  invariably  followed  by  a  prominent  deterior- 
ation in  performance  (Figs.  11,  12,  13,  14).  The  latter  was  charac- 
teristic of  all  of  these  animals  under  the  experimental  conditions 
outlined.  The  only  exception  is  illustrated  in  Figure  12A  where 
the  first  lapse  in  training  of  one  day  was  not  followed  by  deterior- 


Injormation  Storage  in  Nerve  Cells 


203 


-Xy^Av^^^-^l^yi.^Vlvjv  ~^-~'-"^~'~-^^-'j|yjW'tai1^"'''W\^.XUvvV~ 


.^-~v'A*'>».~'v»,(^llf(|lJ,YfV'*''''' 


'As//^/-^yv'A;-*A,Ay-^//,^;s/v,\,^^-,-v.yvv^^/* 


7    v*'V^'^''''^^''',*AVy.»^/ 


Fig.  10.  Unconditioned  (A)  and  conditioned  (B)  avoidance  response  in  the 
rabbit  as  recorded  by  electromyogram  of  the  right  forelimb.  The  upper  six 
channels  are  EEG  tracings  obtained  at  the  same  time.  Electrodes  1,  3,  and  5 
derive  respectively  from  left  motor,  somatic  sensory  and  visual  corte.x.  Electrodes 
2,  4,  and  6  are  corresponding  placements  on  the  right  hemisphere.  Calibration: 
50  microvolts  and  one  second.  Further  explanation  in  the  text. 


so 


b60 
-a 

*-50 


:30 


;?*io 


\  \  \ 

CoThodal  Vo\Q<r\zar\ox\(RiZc-bO) 


0  PoloriZQTion  ofmoTor  corres 
9 Polarizorion  of  visual  corr&x 
^Polarization  of  ear     , 


15 


Time  in  daijs 

Fig.  11.  Effect  of  cathodal  polarization  and  of  a  break  in  training  on  a  typical 
learning  curve.  In  this  and  subsequent  curves  only  the  later  portion  of  the  full 
curve  is  plotted,  i.e.,  beginning  when  the  animal  was  making  at  least  40  per  cent 
correct  responses.  In  most  animals  it  took  two  to  three  weeks  (or  280-420  trials) 
before  the  level  at  which  the  graphs  begin  was  reached.  See  text. 


204 

100 
90 


60 
50 
40 

30 

_^ 

0  20 

'l 

♦- 

3-10 

4- 

c 

(9 

t- 
.&00 

o 

s<?0 


880 


fj  70 
o 

1 60 
50 
40 
30 
20 


Information  Storage  and  Neural  Control 


a.  Cathodal  Polarization    CR2&C-0O) 


0  Polanz-ation  (-)  of  mofor  cortex 
9  Polarization  (->  of  visual  correx 
(§)  Polor'iz Qi-ion  (-) of  ear 


I  \  r \ r 

b.  Anodol  Polarization  (R290-0OJ 


O  Polarization  (■>-) of  motor  cortex 
0  Polariz-otion  (.-I-)  of  visual  cortex 
@  Polorizotiani.-*-)  of  ear 


10  - 


10 


15  20 

Titne  in  daqs 


25 


30 


Fig. 


12.  Effect  of  cathodal  (A)  and  anodal  (B)  polarization  on  learning  curves  in 
two  different  animals.  See  text. 


ation,  although  on  the  second  and  third  occasions  when  training 
was  omitted  for  four  and  three  days,  the  usual  decrease  did  occur. 
It  seems  unlikely  that  the  impaired  performance  on  the  day 
following  cathodal  polarization  was  due  to  a  long  persisting  effect 
of  the  cathodal  current.  Indeed  as  we  shall  see  later,  the  level  of 
decrement  on  the  day  after  cathodal  polarization  of  visual  cortex 


Information  Storage  in  Nerve  Cells 


205 


100 
90 
80 
70 

60 

50 

^0 
•/I 
"5  30 

'C 

•I- 
^20 

t- 
c 

? 

•I- 

olOO 

Cil 

i_ 

o90 
o 

c80 
o 
b  TO 

a 

60 
50 
^0 
30 
20 
10 


a.  Cathodal    Polari'zafion  ('/?/9c-6(3j 


0  Polar)  zat'i  on  (-)  of  motor  corTex 
9  PolariT-Qtioni-) of  viiual  correx- 
(§)  Polar  I  z  Qfion(-)ofear' 

\ I . I 


b.  Anodal   Polarixation  (Rloa-oO) 


O  PolariTiotionCf)  of  motor  cor  re.\ 
0  Polar} zotioti  (-t)  of-  v'S uol  cartas 
@  Polar/  zotion  C^)  of  ear 


15  10  15  20  25  30 

Time  in  claims 
Fig.  13.  Effect  of  cathodal  (A)  and  anodal  (B)  polarization  on  learning  curves 
in  two  additional  animals.  Explanation  in  text. 


corresponds  closely  to  that  occurring  after  a  lapse  in  training. 
It  would  appear  that  training  under  conditions  of  cathodal  polari- 
zation of  visual  cortex  did  not  result  in  registration  or  retention 
of  that  day's  experience.  The  animal  behaved  as  though  there 
had  been  no  training  at  all  on  that  day. 

Figures  12B  and  13B  present  curves  of  the  typical  response 
pattern  in  two  animals  subjected  to  bilateral  anodal  polarization 
of  motor  cortex,  visual  cortex  and  the  ear.  There  is  no  evidence 


206 


Information  Storage  and  Neural  Control 


Carhodoi  Polarization  Ancdal  PolorizQt-ion 

R3Z  Qc  -60 


O  0  Polarization  of  motor  cortex 
-0'  O  Polarisation  of  viiual  cortex 
©  ®  Polarization  of  ear 


25 


Tim<2  in  claims 

Fig.  14.  Effect  of  cathodal  and  anodal  polarization  applied  on  different  occasions 

in  the  same  animal.  Note  again  the  marked  decrease  in  performance  following 

a  six  day  break  in  the  training  schedule. 


that  performance  was  significantly  altered  on  the  day  on  which 
polarization  was  carried  out  at  any  of  the  three  sites.  It  is  interesting 
that  in  every  instance  (see  also  Fig.  14)  there  was  an  abrupt  rise 
in  the  response  per  cent  on  the  day  following  anodal  polarization 
of  visual  cortex. 

Figure  14  presents  the  data  on  one  of  the  two  animals  receiving 
both  anodal  and  cathodal  sequences.  Note  again  the  depression 
of  performance  during  passage  of  cathodal  current  in  visual  cortex 
and  the  maintenance  of  depression  on  the  day  following.  The  usual 
performance  decay  following  a  lapse  in  training  is  also  apparent. 

In  the  same  animal  the  application  of  anodal  current  to  the 
same  three  areas  resulted  in  no  significant  change.  Yet  on  the 
day  following  the  visual  anodal  polarization  performance  reaches 
its  all  time  peak  for  this  animal. 

Figure  15  illustrates  sample  records  of  conditioned  responses 
obtained  under  the  various  conditions  in  this  experiment.  Figure 
15A  represents  cathodal  polarization  of  the  motor  cortex.  Note 
the  characteristically  long  latency  of  the  CR  although  the  animal 
responded  correctly  as  many  times  per  session  as  it  would  without 


Information  Storage  in  Nerve  Cells  207 

A  B  c  D 

w«^-v-^.v-W'Av^\^/^WwV'^  '*^.-''>^'f^j(./v^*^'"j--»-^v"'^''^  ,^-.^.>^^^l'Wf^^•l||^'V^vV^^  "^"^wUff  1  ^'^-^■^-'^^^''V" 


Fig.  15.  EEG  tracings  and  conditioned  response  performed  during  various  con- 
ditions of  polarization.  A. — cathodal  polarization  of  motor  corte.x.  B. — anodal 
(motor).  C— cathodal  (visual).  D. — anodal  (visual).  Calibration:  50  microvolts 
and  one  second.  Explanation  in  text. 

polarization.  On  the  other  hand,  during  anodal  polarization  of 
the  motor  cortex  (Fig.  15B)  the  latency  of  CR  was  very  short  and 
the  amplitude  of  photic  "driving"  was  much  reduced.  A  similar 
EEG  pattern  was  observed  during  application  of  cathodal  current 
to  the  visual  areas  (Fig.  15C).  Reversing  the  direction  of  current 
flow  (Fig.  15D)  resulted  in  considerable  augmentation  of  photic 
"driving." 

A  summary  of  data  in  critical  training  sessions  is  given  in  Table  I. 
The  number  of  observations  and  the  median  and  range  for  the 
number  of  conditioned  responses  per  session  are  listed  for  the 
following  conditions: 

Polarization  of  ears;  motor  cortex,  anodal,  cathodal;  visual  cortex, 
anodal,  cathodal;  session  before  anodal  and  cathodal  (visual) 
polarization;  session  after  cathodal  (visual)  polarization;  session 
after  anodal  (visual)  polarization;  session  after  break  in  training. 

Statistical  analysis  of  these  findings  leads  to  the  following  con- 
clusions: 

1)  Performance  under  tlie  condition  of  visual  cathodal  polari- 
zation differs  from  that  of  visual  anodal,  motor  anodal 
and  cathodal  and  ear  (anodal  and  cathodal)  at  better 
than  the  1  per  cent  level  of  confidence. 


208  Information  Storage  and  Neural  Control 

2)  There  is  no  significant  difference  in  performance  between 
any  one  of  the  polarization  concUtions  (except  visual 
cathode)  and  any  other. 

3)  Performance  on  the  day  JoUowing  visual  cathodal  polari- 
zation differs  from  that  on  the  day  preceding  it  at  the  1 
per  cent  level. 

4)  There  was  no  significant  difference  between  performance 
on  the  day  following  visual  cathodal  polarization  and  the 
day  after  a  break  in  training. 

5)  Comparison  of  performance  on  the  day  after  visual  anodal 
polarization  with  that  on  the  day  preceding  (visual, 
anodal)  polarization  yields  a  difference  significant  at 
better  than  1  per  cent  level  of  confidence. 

With  respect  to  the  last  "conclusion''  listed,  one  must  hasten 
to  add  that  there  is  no  justification  for  attributing  the  improve- 
ment to  an  effect  of  anodal  polarization.  The  difference  may 
simply  reflect  the  rising  learning  curve  or  the  gain  normally 
expected  in  two  days  of  practice.  Only  if  the  experiment  were 
performed  at  a  point  on  the  learning  curve  where  the  gain  to  be 
expected  in  two  days  of  practice  was  negligible  could  one  attribute 
the  change  to  the  neurological  intervention.  Such  was  not  the 
case  in  these  experiments  and  therefore  the  influence  (if  any)  of 
anodal  polarization  of  the  cortical  receiving  area  for  the  conditional 
signal  remains  uncertain. 

In  summary,  it  is  clear  that  the  imposition  of  a  surface  negative 
potential  gradient,  along  the  axis  of  the  main  neural  elements 
of  the  cortical  receiving  area  for  the  conditional  signal,  interferes 
with  conditioned  performance  and  prevents  retention  of  the 
experience  acquired  during  such  polarization.  Although  the  evi- 
dence is  much  less  conclusive  it  seems  possible  that  surface  positive 
currents,  while  not  producing  any  improvement  in  performance, 
may  lead  to  increased  retention  of  the  information  transmitted 
during  the  period  of  current  flow.  These  last  experiments  lend 
some  support  to  the  notion  that  the  electrophysiological  changes 
secondary  to  imposed  potential  gradients,  illustrated  in  the  previous 
studies,  may  have  behavioral  significance  and  may  be  relevant 
to  the  manner  in  which  the  central  nervous  system  achieves  short- 


Information  Storage  in  Nerve  Cells 


209 


pa    oi 


^^r 


^  TT<  '^ 


^ 


(o 


o 


^  >") 


(U    ^»      G 


210  Information  Storage  and  Neural  Control 

term  information  storage.  Despite  these  brief  and  uncertain  insights, 
it  is  most  tantaUzing  to  realize  that  even  at  the  single  unit  level 
of  analysis  the  nature  of  the  neural  code  has  still  strangely  escaped 
detection.  Thus  in  Figure  5D  the  segment  of  record  preceding 
the  application  of  the  single  test  flash  revealed  no  trace  of  the 
information  which  the  subsequent  stimulus  demonstrated  had 
been  retained  in  that  particular  cell.  This  negative  evidence  argues 
against  the  notion  that  the  short-term  memory  trace  is  preserved 
by  means  of  nerve  impulses  continuously  circulating  in  more  or 
less  closed  neuronal  chains.  The  recording  systems  employed 
should  have  been  adequate  to  discern  such  activity  had  it  been 
present.  Perhaps  the  relevant  electrical  signs  are  more  likely  to 
be  found  in  tlie  slow  local  oscillations  of  synaptic  potential  or 
other  sources  of  slowly  varying  voltage.  Such  oscillation  would 
have  been  missed  by  the  short  time-constant  recording  system. 
It  is  also  possible,  of  course,  that  the  encoding  process  took  place 
in  cells  penultimate  to  the  one  monitored. 

I  should  now  like  to  leave  the  question  of  electrical  mechanisms 
for  information  storage  in  the  nervous  system  and  turn  to  some 
other  approaches  which  have  recently  gained  prominence. 

While  it  seems  reasonable  to  postulate  an  electrical  basis  for 
the  labile  short-term  storage  mechanism,  it  is  certainly  difficult 
to  assume  that  the  relatively  permanent  memory  trace  which 
remains  undisturbed  by  the  drastic  perturbations  of  cerebral 
function  produced  by  convulsions,  electro-shock,  concussion,  or 
anesthesia  so  deep  as  to  cause  electrical  silence,  can  be  based 
upon  continuously  circulating  nerve  impulses  (6,  41).  Most 
workers,  therefore,  have  tended  to  think  more  in  terms  of  morplio- 
logical  or  chemical  alterations.  The  essential  thesis  argues  that 
recurrent  impulse  impingement  or  synaptic  bombardment  results 
in  a  durable  morphological  or  chemical  change  which  renders 
tliat  particular  junction  or  cell  more  easily  susceptible  to  subse- 
quent activation  via  the  same  pathway. 

Recently  hypotheses  implicating  ribonucleic  acid  (RNA)  in  the 
molecular  organization  responsible  for  long-term  information 
storage  have  been  proposed  quite  independently  by  a  number  of 
workers  (14,  26,  27,  29,  30,  32,  39,  43).  To  my  knowledge  the  first 
statement  of  this  general  hypothesis  in  the  English  literature  was 


Information  Storage  in  Nerve  Cells  211 

that  of  Katz  and  Halstead  in  1950  (30).  Most  recently  in  a  series 
of  lectures  and  articles  Hyden  (26,  27)  has  forcefully  argued  for 
implication  of  RNA  in  the  molecular  mechanism  of  memory. 
Although  the  concrete  evidence  is  sparse,  some  is  now  available. 
Kreps  (32,  56)  in  the  Soviet  Union,  is  reported  to  have  demon- 
strated an  alteration  in  RNA  synthesis  in  regions  of  the  nervous 
system  related  to  the  conditioned  stimulus  after  establishment  of 
the  conditioned  response.  In  cats,  John,  Wenzel  and  Tschirgi  (29) 
noted  that  intraventricular  injection  of  ribonuclease  was  followed 
l^y  deterioration  of  pattern  discrimination  lasting"  about  four  days. 
Avoidance  CRs  in  the  same  animal  were  unaffected.  Unfortunately 
no  control  data  were  presented  with  regard  to  local  or  general 
changes  in  brain  RNA  content  or  turnover.  Therefore,  since  other 
substances  such  as  calcium  or  potassium  ion  also  reversibly  impair 
CR  performance,  it  is  not  certain  that  the  disturbance  was  related 
to  an  alteration  of  the  RNA  substrate  rather  than  to  another  more 
nonspecific  action  of  ribonuclease. 

Our  own  explorations  in  this  area  were  derived  from  an  inci- 
dental observation  made  in  the  course  of  investigation  of  an 
entirely  separate  problem.  We  had  been  studying  some  physio- 
logical properties  of  the  chronic  focal  epileptogenic  lesion  produced 
in  animals  by  local  freezing  of  a  small  area  of  the  cortical  surface 
(36).  Following  this  procedure,  the  gradual  establishment  of  an 
epileptogenic  lesion  may  be  verified  by  recording  the  paroxysmal 
electrical  activity  which  appears  in  the  cortical  tissue  immediately 
adjacent  to  the  frozen  zone.  We  were  interested  in  studying  the 
ontogenesis  of  an  epileptic  lesion  from  a  chemical  as  well  as  an 
electrical  point  of  view,  and  among  a  number  of  findings  was  the 
observation  that  nerve  cells  in  the  area  of  epileptic  discharge 
stained  densely  with  methyl  green  pyronin  (39).  Methyl  green 
pyronin  is  one  of  the  substances  generally  used  for  the  histochemical 
demonstration  of  RNA. 

It  was  not  particularly  surprising  to  find  increased  concentra- 
tions of  RNA  in  cells  discharging  at  abnormally  high  rates.  Hyden 
and  co-workers  (4,  5,  20,  21,  22,  23,  25)  using  much  more  elegant 
technicjues  had  already  demonstrated  increases  in  cellular  RNA 
consequent  to  prolonged  stimulation.  Recently  lizuka  et  al.  (28), 
in  Japan,  have  confirmed  our  own  observations  specifically  with 


212 


Information  Storage  and  Neural  Control 


1_3  A  R  459 


3-5 


'»A»jf\VvK/V^'^'*''V^Ai>^^ 


5-7 


[/Vf/VWVv^ 


v<^,^/JNv^VJ■ 


vJ^/~v^J\/v'AA^^vVV^--'^,A'vv^'VJV^''^-'^/VW^^ 


Fig.  16.  Electroencephalogram  of  an  unanesthetized  rabbit  twenty-four  hours  (A) 
and  three  days  (B)  after  production  of  an  ethyl  chloride  lesion.  The  site  and  ex- 
tent of  the  lesion  are  indicated  by  the  cross-hatched  area  on  the  diagram.  Deriva- 
tions are  bipolar  from  implanted  electrodes  over  the  indicated  regions.  Cali- 
brations: 50  microvolts  and  one  second  (39). 


respect  to  cells  undergoing  convulsive  discharge.  However,  there 
was  another  phenomenon  noted  in  the  animals  with  chronic 
experimental  epilepsy  which  made  it  possible  to  probe  more 
deeply  into  the  relationsliip  between  ribonucleic  acid  and  cellular 
memory  (43). 

Freezing  a  small  segment  of  the  surface  of  one  cerebral  hemi- 
sphere results  within  a  few  hours  in  the  appearance  of  high  voltage 
epileptiform  spikes  confined  to  the  site  of  the  primary  lesion.  These 
are  illustrated  in  the  upper  part  of  Figure  16.  Simultaneous 
recordings  from  the  opposite  hemisphere  and  from  other  portions 
of  the  same  hemisphere  did  not  reveal  any  abnormality.  After  a 
time,  varying  from  a  few  days  to  three  weeks,  one  may  observe 
(Fig.   16B)  the  development  of  similar  paroxysmal  activity  in  an 


Information  Storage  in  Nerve  Cells 


213 


Fig.  17.  Characteristics  of  the  dependent  mirror  focus.  In  the  ink-written  tracing 
the  upper  two  channels  record  the  primary  focus  and  the  lower  two  channels 
the  mirror  focus.  The  ethyl  chloride  lesion  is  indicated  by  cross  hatching.  Cali- 
bration: 100  microvolts  and  one  second.  In  the  oscillographic  tracing,  the  upper 
channel  records  the  primary  region  and  the  lower  one  the  secondary  region. 
Calibration:  100  milliseconds  (39). 


area  of  the  opposite  hemisphere  homotopic  with  that  of  the  primary 
lesion.  The  contralateral  hemisphere  had  not  been  exposed  or 
damaged  in  any  way  during  the  original  operative  intervention. 
The  paroxysmal  discharge  in  the  contralateral  hemisphere  results 
directly  from  massive  synaptic  bombardment  over  known  ana- 
tomical pathways  from  cells  of  a  primary  epileptogenic  lesion. 
Consequently  the  electrical  abnormality  in  the  contralateral  focus 
is  considered  to  represent  a  secondary  epileptogenic  lesion  (38). 
At  first  the  secondary  discharge  w^as  clearly  dependent  upon  the 
primary  in  the  sense  that  spikes  only  occurred  in  temporal  con- 
junction with  those  in  the  primary  lesion,  had  a  measureable 
latency  following  the  primary  spike  (Fig.  17)  and  disappeared 
altogether  after  excision  or  neuronal  isolation  of  the  original  focus. 
The  pattern  of  activity  in  the  secondary  area  looks  like  a  "reflec- 
tion" of  that  in  the  primary  and  thus  has  earned  the  colloquial 
name  of  "mirror  focus."  If  the  primary  lesion  was  not  excised 
or  isolated  the  mirror  focus  eventually  became  independent. 
Secondary  spikes  were  then  unrelated  in  time  to  those  in  the 
primary  focus  (Fig.  18)  and  did  not  subside  if  the  original  lesion 


214  Information  Storage  and  Neural  Control 

1-2 

3-4 
4-5 

6-7 
7-8 


Fig.  18.  Electrographic  characteristics  of  the  independent  mirror  focus.  Electro- 
encephalogram of  an  unanesthetized  rabbit  taken  three  weeks  after  production 
of  an  ethyl  chloride  lesion  in  the  area  designated  by  crosshatching.  Discharges 
originating  in  the  primary  lesion  (electrode  2)  and  in  the  secondary  region 
(electrode  7)  are  unrelated  in  time  of  occurrence.  Note  also  that  there  is  some 
depression  of  activity  in  electrodes  just  posterior  to  the  primary  lesion  while 
this  is  not  true  in  electrodes  posterior  to  the  mirror  focus.  Calibration:  50  micro- 
volts and  one  second  (39). 


was  subsequently  ablated.  We  have  demonstrated  that  the  func- 
tional characteristics  of  the  cell  network  within  the  mirror  focus 
are  more  or  less  permanently  altered  and  that  the  alteration  is 
manifested  both  by  the  spontaneous  behavior  of  these  cells  and 
by  their  response  to  stimulation  (38,  39,  43). 

The  sequence  just  described  may  be  prevented  by  section  of 
the  corpus  callosum  either  before  production  of  the  primary  lesion 
or  within  twenty-four  hours  afterward.  In  addition,  the  develop- 
ment of  independent  secondary  discharge  may  also  be  prevented 
if  the  callosal  connections  remain  intact,  but  a  sub-pial  partial 
isolation  of  the  contralateral  cortex  is  carried  out  within  the  same 
time  interval.  Figure  19  illustrates  such  a  preparation.  The 
isolation  deprives  the  cortex  of  all  of  its  subcortical  connections 
as  well  as  those  relating  it  to  other  cortical  areas  in  the  same 


Information  Storage  in  Nerve  Cells 


215 


Fig.  19.  Dissectionof  rabbit  brain  to  illustrate  features  of  the  extracallosal  isolation. 
A  slab  of  cerebral  cortex  in  the  hemisphere  opposite  the  ethyl  chloride  lesion  is 
dissected  so  that  the  cortex  is  separated  from  all  subcortical  connections  and 
from  the  surrounding  intracortical  regions  as  well.  The  callosal  pathway  remains 
intact  and  is  the  only  connection  through  which  input  is  available  to  the  dissected 
region.  For  photography,  the  cortex  was  lifted  to  demonstrate  the  underlying 
white  matter  but  the  operative  procedure,  of  course,  is  done  in  such  a  way  as 
to  preserve  the  pial  circulation  to  the  cortical  slab  (38). 


Fig.  20.  A  Weil  stained  cross  section  of  the  extracallosal  isolation.  The  integrity 
of  the  callosal  pathway  is  well  visualized  (38). 


hemisphere.  Dependent  secondary  discharge  does  occur  in  such  a 
preparation  as  do  electrically  evoked  trans-callosal  potentials, 
indicating  that  the  undercut  region  is  viable  and  the  callosal 
pathway  intact.   A  Weil-stained  section  is  shown  in  Figure  20. 


216  Information  Storage  and  Neural  Control 

It  appears  then  that  the  enduring"  changes  in  synaptic  function 
which  form  the  basis  of  the  independent  mirror  focus  require  that 
at  least  two  forms  of  input  be  available  to  the  cortical  region 
concerned. 

It  seemed  appropriate  to  inquire  whether  the  change  in  excita- 
bility or  irritability  of  the  mirror  focus  was  dependent  upon 
impulses  circulating  in  closed  chains  of  neurones  or  whether  it 
was  based  upon  structural  alterations  of  cells  within  the  network. 
As  a  first  step,  neuronal  isolation  of  the  region  of  primary  discharge 
was  carried  out  according  to  the  technique  of  Kristiansen  and 
Courtois  (33).  Figure  21 A  illustrates  persistent,  perhaps  even 
augmented  activity,  in  the  mirror  focus  after  isolation  of  the  pri- 
mary lesion.  There  was  cessation  of  paroxysmal  discharge  in  the 
isolated  primary  lesion.  The  mirror  region  was  then  similarly 
isolated  (Fig.  21 B  and  C).  Some  residual  spiking  sometimes 
persisted  for  several  minutes  in  the  isolated  mirror  region  (Fig.  2 IB) 
but  soon  disappeared  to  be  replaced  by  electrical  silence  (Fig.  21C). 
After  these  isolations  were  performed  the  calvarium  was  replaced 
and  the  animal  returned  to  its  cage  for  several  months.  Surface 
recording  during"  that  period  indicated  no  return  of  paroxysmal 
discharge.  The  lack  of  grossly  recordable  spontaneous  paroxysmal 
activity  was  associated  with  a  corresponding  absence  of  spon- 
taneous unit  discharge  when  at  a  later  date,  single  cells  of  the 
isolated  epileptic  zone  were  probed  with  microelectrodes.  The 
last  two  observations  afford  reasonably  compelling  proof  that  self- 
re-exciting  impulse  chains  do  not  persist  after  the  isolation  pro- 
cedure. If  the  increased  excitability  characteristic  of  the  epileptic 
focus  is  dependent  upon  continuous  self  re-excitation,  the  isolation 
procedure  should  abolish  the  abnormal  excitability.  A  direct  test  of 
this  prediction  was  then  undertaken. 

The  animals  which  had  been  subjected  to  complete  neuronal 
isolation  of  both  primary  and  secondary  epileptogenic  regions 
were  prepared  for  an  acute  experiment.  Several  non-epileptic 
animals  had  had  a  comparable  isolated  cortical  slab  prepared 
at  the  same  time  as  those  in  the  epileptic  group.  In  a  third  group 
of  animals  neuronal  isolation  of  normal  cortical  tissue  in  one 
hemisphere  was  accomplished  prior  to  the  introduction  of  an 
epileptogenic  lesion  in  the  opposite  hemisphere  at  a  point  exactly 


biformation  Storage  in  Nerve  Cells 


217 


3- 


s/Ia-u~^|vV''^'-^A'^'-^H^V^i 


7-! 


,/vw\-/V^wVV^'l^v^JV*^vv|v/ 


6-4 


I  -2 


3-5 


6-4 


Fig.  21.  Bilateral  cortical  isolation  in  an  animal  with  well  developed  independent 
mirror  focus.  This  is  the  same  animal  shown  in  Figure  18.  Recordings  were 
made  with  the  animal  under  40  milligrams  per  kilogram  of  nembutal  anesthesia. 
Derivations  are  from  the  electrodes  indicated  in  the  diagrams.  Isolation  of  the 
primary  lesion  is  first  carried  out  (21A)  and  demonstrates  loss  of  paroxysmal 
spike  discharge  in  the  primary  region  while  the  secondary  area  continues  to 
discharge  actively.  Isolation  of  the  secondary  region  is  then  undertaken  (21 B  & 
C).  For  a  few  moments  abnormal  discharge  persists  in  the  isolated  secondary 
region  (B)  but  soon  disappears  (C)  to  be  replaced  by  almost  complete  electrical 
silence.  Note  that  the  electrode  positions  have  been  changed  in  B  and  C.  Cali- 
bration: 50  microvolts  and  one  second  (39). 


218  Information  Storage  and  Neural  Control 

contralateral  to  the  center  of  the  isolated  slab.  Once  the  epileptic 
lesion  had  begun  to  discharge  actively  it  too  was  isolated  in  the 
same  way.  It  was  thus  possible  to  compare  the  properties  of  neurally 
isolated  non-epileptic  tissue  in  one  hemisphere  with  similarly 
isolated  but  epileptic  tissue  in  a  comparable  region  of  the  opposite 
hemisphere  in  the  same  animal. 

Although  many  different  test  situations  were  investigated,  only 
one  will  be  discu.ssed  at  this  time.  Approximately  three  months 
after  the  cortical  isolations  were  made  the  animals  were  prepared 
for  an  acute  experiment.  Wide  exposure  of  both  cerebral  hemi- 
spheres and  a  tracheotomy  were  performed  under  ether  anesthesia 
after  which  the  ether  was  allowed  to  dissipate  and  the  animals 
were  maintained  under  Flaxedil  and  artificial  respiration.  The 
pial  surface  was  covered  with  warm  mineral  oil  or  saline.  Epilep- 
tiform after-discharges  were  induced  in  the  intact  normal  cortex 
outside  the  isolated  zones  either  by  direct  electrical  stimulation 
or  by  placement  of  small  pledgets  of  filter  paper  soaked  in  Metrazol. 
Propagation  of  these  after-discharges  was  monitored  by  means  of 
recording  electrodes  distributed  throughout  the  intact  cortex  and 
within  the  isolated  area.  The  extent  to  which  high  voltage  dis- 
charge originating  externally  spreads  across  the  solution  of  neural 
continuity  to  excite  cells  within  the  isolated  zone  is  considered  to 
be  a  measure  of  the  excitability  of  those  cells. 

In  our  experience  it  was  rare  indeed  for  paroxysmal  discharge 
to  cross  the  neural  gap  and  excite  non-epileptic  isolated  cortex 
(Fig.  22).  As  may  be  seen  in  Figure  22  this  was  true  even  when 
the  epileptiform  activity  was  of  extremely  high  voltage,  long 
duration,  and  spread  quite  readily  to  the  opposite  hemisphere. 
On  the  other  hand  the  isolated  epileptic  tissue  of  the  mirror  focus 
was  quite  easily  invaded  by  epileptiform  activity  arising  ex- 
ternally (Fig.  23). 

In  the  experiment  illustrated  in  Figure  23,  tungsten  micro- 
electrodes  having  tip  diameters  of  1-5  micra  were  inserted  to 
a  depth  of  500-1000  micra  into  the  isolated  slab.  Since  a  search 
for  spontaneously  firing  units  was  rarely  successful  it  was  necessary 
to  rely  upon  multiple  placements  at  a  depth  where  unit  discharge 
might  reasonably  be  expected  in  connection  with  surface  electro- 
graphic  paroxysms.   Microelectrode  recording  was  employed   in 


Information  Storage  in  Nerve  Cells 


219 


2-4, 


7-M 


Fig.  22.  Failure  of  epileptiform  after-discharge  to  invade  non-epileptic  neuronally 
isolated  region.  Unanesthetized  rabbit.  Implanted  electrodes  at  sites  indicated 
on  diagram.  Channel  designations  refer  to  the  correspondingly  numbered  elec- 
trodes and  denote  grid  1  and  grid  2  respectively.  Electrical  stimulus  had  been 
applied  to  cortical  surface  at  site  of  electrode  4.  The  electrode  wiUiin  the  isolated 
region  (7)  is  connected  to  a  reference  (M)  sewed  into  the  cervical  muscles. 
Calibration:  50  microvolts  and  one  second. 


Normol  corlex  (1-21 

Normol  cortex  (2-3)  ^ 

Isolated  cortex  (4-M) 


Normol  cortex 


Microelectrode 


Fig.  23.  Propagation  of  epileptiform  discharge  into  an  isolated  epileptic  region. 
Two  examples  with  both  surface  and  simultaneous  microelectrode  recording. 
A  pledget  of  filter-paper  soaked  in  Metrazol  was  placed  on  normal  cortex  out- 
side the  isolated  slab  at  2  cm.  distance.  The  electrographic  discharge  so  induced 
spread  slowly  across  the  cortex  and  after  some  delay  invaded  the  isolated  zone. 
Single  units  within  the  slab  were  recorded  through  tungsten  microwires  having 
a  tip  resistance  of  10-40  megohms.  Calibration:  50  microvolts  and  one  second 
for  the  ink-writer  tracings  and  one  second  for  the  cathode  ray  oscilloscope  (39). 


220  Information  Storage  and  Neural  Control 

order  to  avoid  the  ambiguity  engendered  when  high  amplitude 
potentials  arising  externally  are  conducted  in  volume  to  the  large 
electrodes  resting  upon  the  surface  of  the  isolated  segment.  Thus 
in  the  first  part  of  the  tracings  in  the  two  experiments  illustrated 
in  Figures  23A  and  B  the  large  electrodes  on  the  surface  of  isolated 
cortex  (Channel  3,  Fig.  23A  and  Channel  2,  Fig.  23B)  record 
potential  variations  precisely  concordant  in  time  with  those  in 
the  surrounding  normal  cortex  where  the  seizure  was  initiated. 
Not  until  seconds  later  did  the  microelectrode  tracing  reveal  that 
single  elements  within  the  slab  had  developed  high  frequency 
self-sustained  discharge.  Since  the  high  impedance  of  the  micro- 
electrode  tip  precludes  recording  at  any  distance,  we  cannot 
escape  the  conclusion  that  nonsynaptic  activation  of  ganglionic 
elements  within  the  isolated  region  had  occurred. 

Although  only  negative  evidence  can  be  presented  for  the  case 
of  non-epileptic  isolated  cortex,  the  contrast  between  that  and 
the  ease  with  which  invasion  of  epileptic  zones  can  be  demon- 
strated has  led  to  the  conclusion  that  abnormal  excitability  per- 
sists in  the  secondary  epileptogenic  focus  for  several  months  after 
an  isolation  procedure  which  eliminated  a  self-reexcitation  mech- 
anism. Presumably,  therefore,  the  persistence  of  abnormal  behavior 
in  these  cells  depends  upon  structural  or  biochemical  alterations 
rather  than  upon  continuing  electrical  input. 

On  the  basis  of  the  reasoning  discussed  earlier  the  ribonucleic 
acid  distribution  in  the  mirror  focus  was  examined  histochemically, 
first  with  the  methyl  green  pyronin  method  and  subsequently 
(with  concordant  results)  with  Azure  B  and  Gallocyanin  at 
acid  pH.  After  preliminary  electrical  studies  had  clearly  indicated 
the  extent  and  distribution  of  both  primary  and  secondary  dis- 
charging areas  the  animals  were  sacrificed  and  brains  perfused 
in  situ.  Serial  sections  were  prepared  and  those  from  primary  and 
secondary  foci  were  compared  with  those  from  electrically  un- 
involved  areas  of  brain.  Figure  24  demonstrates  a  small  nest  of 
darkly  stained  cells  in  a  section  taken  from  the  electrically  defined 
mirror  focus.  The  border  of  the  densely  stained  region  is  fairly 
sharp,  and  to  the  left  is  the  adjacent  normal  cortex,  so  that  one 
may  compare  the  dye-binding  property  of  normal  cortical  tissue 
with  that  of  the  electrically  abnormal  zone  on  the  right.  A  slightly 


Information  Storage  in  Nerve  Cells  221 


1    .       ^ 


.  i"  ^' 


Fig.  24.  Sections  Lhruugh  ilic  region  ul  ilit-  uiinur  lucus.  Note  the  collection 
of  densely  stained  cells  to  the  right  of  the  photomicrograph  compared  with  the 
characteristic  staining  of  normal  cortex  to  the  left.  Methyl  green  pyronin  stain. 

Magnification  x75  (39). 


higher  power  photomicrograph  (Fig.  25)  illustrates  the  pene- 
tration of  the  pyronin-positive  material  into  the  dendrite  and 
also  indicates  the  wedgelike  distribution  of  the  stained  cell  system. 
Pigmented  cells  extend  throughout  the  depth  of  the  cortex.  At 
still  higher  magnification  the  extent  of  penetration  into  the  dendrite 
is  clearer  (Fig.  26)  and  one  may  observe  a  concentration  of  the 
pyronin-positive  material  in  a  dense  layer  along  the  inner  surface 
of  the  cell  membrane.  The  altered  tinctorial  properties  of  these 
cells  were  abolished  by  pre-treatment  of  the  slide  with  ribonuclease 
and  were  unaffected  by  similar  treatment  with  deoxyribonuclease 
and  other  enzymes.  Although  the  histochemical  picture  was  some- 
what obscured  by  surgical  artifacts  the  cells  in  the  isolated  mirror 
focus  exhibited  the  same  pyronin-dense-pattern  as  had  the  intact 
secondary  region.  Further  controls  may  be  found  in  the  original 
report   (39). 


222  Information  Storage  and  Neural  Control 


~  f    ^      '  'y 


/ 


./ 


Fig.  25.  Slightly  higher  power  photomicrograph  through  region  of  mirror  focus. 
The  appearance  of  normal  cortical  cells  with  this  method  is  seen  in  the  lower 
right  and   upper  left  hand   corners.    Methyl  green   pyronin   stain.    Magnifica- 
tion x85  (39). 


Interpretation  of  the  histochemical  results  is  still  an  entirely 
open  question.  The  evidence  is  not  sufficient  to  conclude  that  the 
alteration  in  RNA  is  specifically  related  to  afferent  bombardment 
since,  although  the  general  areas  coincide,  there  is  no  way  to 
know  whether  a  given  pyronin-dense  cell  has  participated  in  the 
epileptiform  activity.  Furthermore  the  nature  of  the  nucleotide- 
dye  mole  interrelation  is  still  incompletely  understood  (52). 
Increased  staining  with  basic  dyes  does  not  necessarily  indicate 
an  increase  in  absolute  amount  of  RNA.  It  is  also  possible  that 
changes  in  polymerization  and  possibly  submolecular  factors 
affecting  charge  distribution  may  influence  dye-binding. 

Despite  many  areas  of  uncertainty  the  bulk  of  experimental 
evidence  is  consistent  with  the  notion  that  except  for  certain 
plant  viruses  the  nucleotide  sequence  in  RNA  is  specified  by 
genetic    information    in    DNA.    If   the    DNA-RNA    specification 


hiformation  Storage  in  Nerve  Cells  223 


^ 


Fig.  26.  Higher  power  photomicrograph  demonstrating  the  characteristic  con- 
centration of  pyronin-positive  material  along  the  inner  surface  of  the  membrane. 
The  stained  material  extends  far  into  the  dendrite.  Note  also  the  appearance  of 
a  bilobed  nucleus.  Methyl  green  pyronin  stain.  Magnification  x840   (39). 


system  were  susceptible  in  a  random  way  to  ionic  fluxes  induced 
by  nerve  impulses  the  ordinary  metabolic  machinery  of  the  cell 
would  be  rapidly  undone.  This  could  be  avoided  only  if  the  nerve 
cell  represented  a  special  case  of  uncoupling  of  the  DNA-RNA 
specification  system,  thus  allowing  a  degree  of  freedom  for  the 
nucleotide  sequence  in  RNA  to  be  influenced  by  environmental 
factors.  Or  alternatively  one  might  assume  that  only  certain  pre- 
selected molecules  are  available  to  influence  by  ionic  flux.  One 
may  entertain  the  view  that  all  possible  RNA  nucleotide  sequences 
and  their  correspondingly  coded  proteins  are  already  available 
within  the  cell.  An  incoming  pattern  of  electrical  impulses  might 
select  or  re-orient  some  of  these  molecules  at  the  expense  of  others. 
Availability  of  the  stored  information  might  be  based  upon  a 
cellular  "recognition"  of  the  same  pattern  of  impulse  impingement 


224  Information  Storage  and  Neural  Control 

or  synaptic  activation  which  established  the  original  alteration. 
Such  "recognition"  may  be  similar  in  mechanism  (still  known) 
to  those  occurring  in  morphogenesis  and  in  antigen-antibody 
reactions.  However,  it  is  well  to  be  aware  that  when  we  substitute 
electric  currents  (whether  or  not  generated  by  chemical  trans- 
mitters) for  the  "antigen"  we  enter  a  realm  of  biological  phe- 
nomena not  based  upon  the  classical  chemistry  of  atoms  and 
molecules — one  in  which  electron  or  charge  transfer  reactions 
afford  the  more  crucial  energizing  mechanisms.  It  is  also  apparent 
that  those  who  would  consider  a  role  for  the  nucleic  acids  in  the 
molecular  basis  of  memory  must  also  explain  how  an  electrical 
current  could  induce  a  molecular  rearrangement  which  is  there- 
after irreversible  and  immune  to  further  perturbations  of  its 
electrical  surround.  Perhaps  the  binding  of  an  appropriately 
modified  RNA  protein  complex  to  phospholipid  would  not  only 
protect  it  from  further  electrical  influence  but  also  fix  it  to  the 
cell  membrane  where  the  function  of  "recognition"  is  most  likely 
to  take  place. 

Finally  I  should  like  to  return  to  the  beginning  and  add  one 
more  note  of  complexity  to  an  already  complex  story.  We  have 
mentioned  the  retrograde  amnesia  produced  by  a  cerebral  con- 
cussion. Clinical  experience  gives  clear  evidence  that  immediately 
following  an  injury  the  memory  loss  may  extend  backwards  in 
time  for  weeks,  months,  or  even  years,  so  that  the  patient  reports 
his  age  as  several  years  younger  than  is  actually  the  case.  During 
recovery  the  memory  gap  decreases  gradually  with  recall  of  more 
distant  events  first  and  recent  events  last.  Russell  and  Nathan, 
in  an  extensive  review  (50),  have  emphasized  that  the  pattern  of 
recovery  shows  no  relationship  to  the  importance  of  the  events 
remembered.  Thus  one  patient  remained  amnesic  for  his  marriage, 
which  had  occurred  three  weeks  prior  to  the  injury,  but  recalled 
perfectly  reading  a  trivial  newspaper  story  six  weeks  earlier.  It  is 
clear  that  memory  returns  not  in  order  of  importance  but  only 
in  order  of  time.  To  be  sure,  even  under  the  best  of  circumstances 
recovery  is  never  complete;  it  is  almost  always  possible  to  demon- 
strate a  complete  and  permanent  loss  of  memory  for  the  events 
immediately  preceding  an  injury.   Perhaps  it  is  this  last,   brief, 


Injormation  Storage  in  Nerve  Cells  225 

blank  interval  which  is  relevant  to  the  electrical  aspects  of  the 
memory  mechanism  discussed  earlier.  Yet  the  total  recovery  pat- 
tern in  retrograde  amnesia  stresses  the  lability  and  vulnerability 
of  the  most  recently  acquired  experience  and  suggests  that  it 
is  not  only  the  electrical  or  short-term  aspects  of  memory  which 
consolidate;  some  form  of  consolidation  must  also  occur  in  the 
structural  or  "permanent"  stage  of  information  storage  (12). 

A  molecular  mechanism  for  information  storage  must  embrace 
all  these  features.  As  a  provisional  target  we  might  envision  a 
molecular  species  which  may  be  altered  by  ionic  flux  but  once 
altered  is  immune  to  other  electrical  interventions,  which  has 
nothing  to  do  with  basic  metabolic  processes,  which  can  replicate 
itself  within  a  cell  and  which  can  alter  the  output  of  that  cell  so 
as  to  disseminate  its  "spoor"  to  the  next  cell  along  the  pathway. 
But  to  envision  is  not  to  identify.  The  target  promises  to  be  elusive. 
The  analogy  of  the  inirror  focus  may  be  rough  indeed  but  it  is 
pertinent  to  recall  that  the  alterations  observed  in  electrical  and 
chemical  properties  are  brought  about  through  the  same  neural 
pathways  available  to  physiological  stimulations.  No  quantitative 
relationship  between  these  data  and  the  events  responsible  for 
behavior  is  implied.  Perhaps  there  is  no  relationship  at  all.  Never- 
theless, used  as  an  experimental  tool  this  model  and  the  observa- 
tions it  has  yielded  so  far  indicate  that  we  have  in  hand,  to  see 
and  to  investigate,  clear-cut  and  permanent  changes  in  cellular 
and  synaptic  properties  related  to  the  past  history  of  that  cell 
or  synapse. 

ACKNOWLEDGMENTS 

These  studies  were  supported  by  U.S.P.H.S.  grant  B-3543.  I 
wish  to  express  my  gratitude  to  the  many  individuals  who  have 
helped  in  various  aspects  of  these  investigations.  Special  appre- 
ciation is  due  to  Dr.  K.  L.  Clhow  and  Mr.  Paul  Naitoh  for  help 
in  some  of  the  experiments  and  to  Professor  Lincoln  Moses  for  the 
statistical  analysis.  Gratitude  is  hardly  the  word  to  express  the 
indebtedness  to  my  wife,  Dr.  Lenore  Morrell,  whose  forebearance 
with  dinners  grown  cold  and  evenings  in  the  laboratory  made 
this  work  possible. 


226  Information  Storage  and  Neural  Control 

REFERENCES 

1.  Abt,  J.  P.,  Essman,  VV.  B.  and  Jarvik,  M.  E.:  Ether-induced  retro- 

grade amnesia  for  one-trial  conditioning  in  mice.  Science,  733: 
1477-1478,  1961. 

2.  Arduini,   A.:    Enduring   potential   changes  evoked   in   the  cerebral 

cortex  by  stimulation  of  brain  stem  reticular  formation  and 
thalamus.  In:  Reticular  Formation  of  the  Brain,  edited  by,  Jasper, 
H.  H.,  Proctor,  L.  D.,  Knighton,  R.  S.,  Noshay,  W.  C,  and 
Costello,  R.  T.  Boston,  Little,  Brown  and  Co.,  1958,  333  pp. 

3.  Bishop,  G.  H.,  and  O'Leary,  J.  L.:  The  effect  of  polarizing  currents 

on  cell  potentials  and  their  significance  in  the  interpretation  of 
central  nervous  system  activity.  Electroenceph.  Clin.  Neurophysiol.,  2: 
401-416,  1950. 

4.  Brattgard,    S.:    The    importance    of   adequate    stimulation    for    the 

chemical  composition  of  retinal  ganglion  cells  during  early  post- 
natal development.  Acta  Radiologica  Supp.  96.,  1952. 

5.  Brattgard,  S.  and  Hyden,  H.:  Mass,  lipids,  pentose  nucleoproteins 

and  proteins  determined  in  nerve  cells  by  x-ray  microradiography. 
Acta  Radiologica  Supp.  94,  1952. 

6.  Brazier,  M.  A.  B.:  Long-persisting  electrical  traces  in  the  brain  of 

man  and  their  possible  relationship  to  higher  nervous  activity. 
In:  The  Moscow  Colloquium  on  Electroencephalography  of  Higher 
Nervous  Activity,  edited  Ijy  Jasper,  H.  H.  and  Smirnov,  G.  D. 
Electroenceph.  Clin.  NeurophysioL,  Supp.  13,  1960. 

7.  Burns,  B.  D.:  Some  properties  of  the  cat's  isolated  cerebral  cortex. 

J.  Physiol.,  777.-50-68,  1951. 

8.  Burns,  B.  D.:  The  production  of  after-burts  in  isolated  unanesthetized 

cerebral  cortex.  J.  Physiol.,  725.- 427-446,  1954. 

9.  Burns,  B.  D.:  The  mechanism  of  after-bursts  in  cerebral  cortex.  J. 

Physiol.,  727.- 168-1 88,  1955. 

10.  Chow,  K.  L.  and  Dewson,  James:  Unpublished  data. 

11.  Chow,  K.  L.:  Brain  Functions.  Ann.  Rev.  Psychol.,  72.-28 1-310,  1961. 

12.  Deutsch,  J.  A.:  Higher  nervous  function:  the  physiological  bases  of 

memory.  Ann.  Rev.  Physiol.,  24:259-286,   1962. 

13.  Duncan,  C.  P.:  The  retroactive  effect  of  electroshock  on  learning. 

J.  Camp.  Physiol.  Psychol.,  42:32-44,  1949. 

14.  Gerard,  R.  W.:  The  fixation  of  experience.  In:  Cioms  Symposium  on 

Brain  Mechanisms  and  Learning,  edited  by  Delafresnaye,  J.  F., 
Fessard,  A.,  and  Konorski,  J.  Oxford,  Blackwell  Scientific  Pub., 
1961. 


Information  Storage  in  Nerve  Cells  227 

15.  Gerard,  R.  W.  and  Libet,  B.:  The  control  of  normal  and  convulsive 

brain  potentials.  Amer.  J.  Psychiat..  96.- 1 125-1 152,   1940. 

16.  Goldring,  S.,  and  O'Leary,  J.  L.:  Experimentally  derived  correlates 

between  EGG  and  steady  cortical  potential.  J.  Neurophysiol.,  14: 
275-288,  1951. 

17.  Goldring,  S.,  and  O'Leary,  J.  L.:  Summation  of  certain  enduring 

sequelae  of  cortical  activation  in  the  rabbit.  Electroenceph.  Clin. 
Neurophysiol.,  J.-329-340,   1951. 

18.  Goldring,  S.  and  OTeary,  J.  L.:  Cortical  D.C.  changes  incident  to 

midline  thalamic  stimulation.  Electroenceph.  Clin.  Neurophysiol.,  9: 
577,  1957. 

19.  Gumnit,  R.  J.:  The  distribution  of  direct  current  responses  evoked 

by  sounds  in  the  auditory  cortex  of  the  cat.  Electroenceph.  Clin. 
Neurophysiol.,  13:889-895,  1961. 

20.  Hamberger,  C.  A.:  Gytochemical  investigations  on  N.  vestibularis. 

Acta  Otolaryng.  Supp.,  78:53-62,  1949. 

21.  Hamberger,  C.  A.  and  Hyden,  H.:  Production  of  nucleoproteins  in 

the  vestibular  ganglion.  Acta  Otolaryng.  Supp.  75:53,   1949. 

22.  Hamberger,  C.  A.  and  Hyden,  H.:  Transneuronal  chemical  changes 

in  Deiter's  nucleus.  Acta  Otolaryng.  Supp.  75:82,  1949. 

23.  Hammer,  G.:  On  quantitative  cytochemical  changes  in  spiral  gang- 

lion cells  after  acoustic  trauma.  Acta  Otolaryng.  Supp.  740.T37-144, 
1958. 

24.  Hebb,  D.  O.:   The  Organization  of  Behavior.  New  York,  Wiley,  1949. 

25.  Hyden,   H.:   Determination   of  mass  of  nerve-cell  components.  J. 

Embryol.  Exp.  Morph.,    7;315-317,   1953. 

26.  Hyden,   H.:  Biochemical  changes  in  glial  cells  and  nerve  cells  at 

varying  activity.  In:  Proc.  Fourth  International  Congress  of  Bio- 
chemistry. Vol.  Ill  Biochemistry  of  the  Central  Nervous  System.  London, 
Pergamon  Press,  1960. 

27.  Hyden,  H.  and  Pigon,  A.:  A  cytophysiological  study  of  the  functional 

relationship  between  oligodendroglial  cells  and  nerve  cells  of 
Deiter's  nucleus.  J.  Neurochem.,  6:57-12,  1960. 

28.  lizuka,  R.,  Takeda,  T.,  Tanabe,  M.,  Takahata,  N.  and  Suwa,  N.: 

Histochemical  studies  on  the  nucleic  acid  and  thiamin  of  nerve 
cells  under  electroconvulsion,  insulin  coma  and  malononitrile 
injection.  Folia  Psychiat.  Neurol.  Jap.,  13:\-\A,  1959. 

29.  John,  E.  R.:  High  nervous  functions  (Brain  functions  and  learning). 

Ann.  Rev.  Physiol,  2J;451-484,  1961. 

30.  Katz,  J.  J.  and  Halstead,  W.  C:  Protein  organization  and  mental 

function.  Comp.  Psychol.  Monog.,  20:\-58,  1950. 


228  Infort?iation  Storage  and  Neural  Control 

31.  Kraft,  M.  S.,  Oljrist,  W.  D.  and  Pribram,  K.  H.:    The  effect  of 

irritative  lesions  of  the  striate  cortex  on  learning  of  visual  dis- 
criminations in  monkeys.  J.  Cornp.  Physiol.  Psychol.,  53:\7-22,  1960. 

32.  Kreps,  E.:  Cited  in  the  use  of  radioactive  isotopes  in  the  study  of 

functional  biochemistry  of  the  brain  by  Palladin,  A.  V.  and 
Vladimirov,  G.  E.  Proc.  International  Conference  on  the  Peaceful 
Uses  of  Atomic  Energy.  Vol.  12,  United  Nations,  New  York,  1955. 

33.  Kristiansen,  K.  and  Courtois,  G.:  Rhythmic  electric  activity  from 

isolated  cerebral  cortex.  Electroenceph.  Clin.  JVeurophysiol.,  7;265-272, 
1949. 

34.  Morrell,   F.:   Interseizure  disturbances  in  focal  epilepsy.  Neurology, 

(5;  327-334,  1956. 

35.  Morrell,  F.:   Effects  of  experimental  epilepsy  on  conditioned  elec- 

trical potentials.   Univ.  oj  Minn.  Med.  Bu'l.,  29:82-102,  1957. 

36.  Morrell,   F.:    Experimental  focal   epilepsy  in   animals.   AM  A  Arch. 

Neurol.,   7:141-147,   1959. 

37.  Morrell,  F.:  Microelectrode  and  steady  potential  studies  suggesting 

a  dendritic  locus  of  closure.  In:  The  Moscow  Colloquium  on 
Electroencephalography  of  Higher  Nervous  Activity,  edited  by 
Jasper,  H.  H.  and  Smirnov,  G.  D.  Electroenceph.  Clin.  Neurophysiol., 
Supp.  13,  1960. 

38.  Morrell,   F.:    Secondary  epileptogenic  lesions.   Epilepsia    7;538-560, 

1960. 

39.  Morrell,  F.:  Lasting  changes  in  synaptic  organization  produced  by 

continuous  neuronal  bombardment.  In  CIOMS  Symposium  on  Brain 
Mechanisms  and  Learning,  edited  by  Delafresnaye,  J.  F.,  Fessard,  A., 
Konorski,  J.  Oxford,  Blackwell  Scientif.  Pub.,  1961. 

40.  Morrell,  F.:   Effect  of  anodal   polarization  on  the  firing  pattern  of 

single  cortical  cells.  Ann.  New  Tor k  Acad,  of  Sci.,  92:860-87 6,  1961. 

41.  Morrell,  F.:   Electrophysiological  contributions  to  the  neural  basis 

of  learning.  Physiol.  Rev.,  47.-443-494,  1961. 

42.  Morrell,  F.,  Roberts,  L.  and  Jasper,  H.:  Effect  of  focal  epileptogenic 

lesions  and  their  ablation  upon  conditioned  electrical  responses  of 
the  brain  in  the  monkey.  Electroenceph.  Clin.  Neurophysiol.,  8:2X1- 
236,  1956. 

43.  Morrell,  F.,  Sandler,  B.  and  Ross,  G.:  The  "Mirror  Focus"  as  a 

Model  of  Neural  Learning.  Proc.  XXI  Internat.  Physiol.  Cong., 
Buenos  Aires,  1959,  193  pp. 

44.  Ransmeier,  R.  E.  and  Gerard,  R.  W. :  Effects  of  temperature,  con- 

vulsion and  metabolic  factors  on  rodent  memory  and  EEG.  Amer. 
J.  Physiol.,  779;663-664,  1954. 


Information  Storage  in  Nerve  Cells  229 

45.  Roitbak,  A.  I.:  Bioelectric  Phenomena  in  the  Cortex  of  the  Cerebral 

Hemispheres.  Part  1  Acad.  Sci.  Georgian  SSR,  Tiflis,  1955. 

46.  Roitbak,  A.  I.:  Electrical  phenomena  in  the  cerebral  corte-x  during 

the  extinction  of  orientation  and  conditioned  reflexes.  In:  The 
Mo.scow  Colloquium  on  Electroencephalography  of  Higher  Ner- 
vous Activity,  edited  by  Jasper,  H.  H.,  and  Smirnov,  G.  D.  Elec- 
troenceph.  Clin.  NeurophysioL,  Supp.  13,  1960. 

47.  Rowland,  V.:   Second  Conference  on  Brain  and  Behavior,  edited 

by  Brazier,  M.  A.  B.  In  press. 

48.  Rusinov,  V.  S.:  An  electrophysiological  analysis  of  the  connecting 

function  in  the  cerebral  cortex  in  the  presence  of  a  dominant 
area.  Communications  at  the  XIX  International  Physiological 
Congress.,  Montreal,   1953. 

49.  Rusinov,  V.   S.:  General  and  localized  alterations  in  the  electro- 

encephalogram during  the  formation  of  conditioned  reflexes  in 
man.  In:  The  Moscow  Colloquium  on  Electroencephalography  of 
Higher  Nervous  Activity,  edited  by  Jasper,  H.  H.  and  Smirnov, 
G.  D.  Electroenceph.  Clin.  NeurophysioL,  Supp.  13,  1960. 

50.  Russell,  W.  R.  and  Nathan,  P.  W.:  Traumatic  Amnesia.  Brain,  69: 

280-300,  1946. 

51.  Stamm,  J.  S.  and  Pribram,  K.  H.:  EfTects  of  epileptogenic  lesions 

in  frontal  cortex  on  learning  and  retention  in  monkeys.  J .  Neuro- 
physioL,  2J;552-563,  1960. 

52.  Steiner,  R.  F.  and  Beers,  R.  F.,  Jr.:  Polynucleotides.  Natural  and 

Synthetic  Nucleic  Acids.  Amsterdam,  Elsevier  Pub.  Co.,  1961. 

53.  Thompson,   R.:   The  comparative  effects  of  ECS   and   anoxia   on 

memory.  J.  Comp.  Physiol.  Psychol.,  50.-397-400,  1957. 

54.  Thompson,  R.  and  Dean,  W.:  A  further  study  on  the  retroactive 

eflfect  of  ECS.  J.  Comp.  Physiol.  Psychol.,  45.-488-491,  1955. 

55.  Tobias,   Julian    M.:    Experimentally    altered    structure    related    to 

function  in  the  lobster  axon  with  an  extrapolation  to  molecular 
mechanisms  in  excitation.  J.  Cell.  Comp.  Physiol.,  52.-89-126,  1958. 

56.  Vladimirov,  G.  E.:  The  influence  of  excitation  of  the  central  nervous 

system  on  some  aspects  of  metabolism  in  the  cerebral  hemispheres 
of  animals.  Proc.  XIX  Internat.  Physiol.  Cong.  Montreal,  pp. 
856-857,  1953. 


CHAPTER 
X 

HOW  CAN  MODELS  FROM  INFORMATION 
THEORY  BE  USED  IN  NEUROPHYSIOLOGY?* 

Mary  A.  B.  Brazier 


w. 


HY  is  it  that  information  theory  has  had  such  an  attraction  for 
neurophysiologists?  From  the  earliest  dissemination  among  scien- 
tists of  Shannon's  information  theory  (9),  developed  in  the  context 
of  communications  technology,  and  of  Wiener's  communication 
theory  (14)  which  expanded  its  frontiers,  neurophysiologists  have 
been  prominent  among  those  who  wished  to  explore  the  potenti- 
alities for  their  field. 

There  were  many  reasons  for  this,  but  I  would  suggest  that  there 
were  three  major  ones,  namely: 

1)  de-emphasis  on  energy-coupling  within  systems  and  emphasis 
on  informational  coupling; 

2)  the  formulation  of  models  for  dealing  with  signals-in-noise; 
and 

3)  the  exploration  of  piobabilistic  models  rather  than  determin- 
istic ones. 

These  are,  of  course,  all  interrelated.  I  would  like  to  discuss  the 
first  two  items  briefly,  together  with  some  other  facets  of  information 
theory  that  impinge  on  neurophysiology,  and  then  give  more 
detailed  attention  to  the  subject  of  probabilistic  models  of  nervous 
system  activity. 


*The  work  reported  here  was  supported  by  USPHS  Grant  NB-03160  and  Contract 
Nonr  233(69)  from  the  Office  of  Naval  Research. 

230 


How  Can  Models Jrom  Information  Theory  he  Used  in  Neurophysiology)?     231 

Let  us  look  first  at  the  difference  between  energy-based  concepts 
and  information-based  concepts.  All  down  the  ages  the  nerves  have 
been  recognized  as  message-carriers  and  as  late  as  the  last  century 
the  most  distinguished  physiologist  of  the  time,  Johannes  Miiller, 
was  using  the  term  "nerve  energy.''  "We  are,"  he  wrote,  "com- 
pelled to  ascribe,  with  Aristotle,  peculiar  energies  to  each  nerve, 
energies  which  are  vital  qualities  of  the  nerve."  Even  the  later  19th 
century  neurophysiology,  dominated  by  Du  Bois  Reymond,  was 
primarily  focussed  on  the  concept  of  the  conservation  of  energy. 

You  will  remember  that  it  was  because  of  his  adherence  to 
energy  concepts  that  Sherrington  (10)  found  himself  unable  to 
envisage  a  physiological  basis  for  mental  processes.  In  Man  on  His 
Nature  he  wrote: 

"No  attributes  of  'energy'  seem  findable  in  the  process  of  mind. 
That  absence  hampers  explanation  of  the  tie  between  cerebral  and 
mental." 

He  goes  on  to  write  of  the  brain  being  "a  physiological  entity  held 
together  by  energy-relations"  and  expresses  his  despair  of  being 
able  to  correlate  such  a  physiological  entity  with  a  mental  experi- 
ence. "The  two  for  all  I  can  do,"  he  wrote  "remain  refractorily 
apart.  They  seem  to  me  disparate;  not  mutually  convertible;  un- 
translatable the  one  into  the  other." 

Coming  to  our  own  times,  we  have  seen  a  great  deal  of  investiga- 
tive effort  go  into  a  search  for  energy  correlates  of  brain  function. 
An  example  is  the  search  for  a  metabolic  change  underlying  the 
sleep  state.  Anesthesiology  is  another  field  in  which  one  finds  many 
studies  centering  around  alterations  of  brain  metabolism  as  the 
major  factor  of  importance  in  the  changing  levels  of  consciousness. 

It  is  only  with  recent  years  that  we  find  attention  being  diverted 
from  the  question  "What  is  the  level  of  activity  in  the  brain  as  a 
whole?"  to  "Which  system  within  the  brain  is  now  dominantly 
active?"  The  latter  question  contains  the  implication  that  it  is  a 
re-routing  of  nerve  impulses,  a  change  in  the  informational  coupling 
rather  than  in  the  general  metabolic  level  of  the  brain's  activity 
that  may  yield  the  clue  to  functional  changes.  In  order  to  effect  a 
coupling  of  parts  within  the  nervous  system  there  does  not  have 
to  be  a  great  interchange  of  energy— only  the  infinitesimal  transfer 
concomitant  with  the  passage  of  the  nerve  impulse. 


232  Information  Storage  and  Neural  Control 

Taking"  the  examples  just  given,  neurophysiologists  are  finding 
closer  correlates  with  the  states  of  sleep  and  anesthesia  from  studies 
of  the  coupling  between  the  cortex,  the  thalamus  and  the  brain 
stem  than  they  have  found  in  their  measurements  of  energy-transfer 
reflected  in  arterio-venous  differences  between  carotid  and  jugular 
blood.  In  the  limbic  system  there  is  now  evidence  for  re-routing 
taking  place  during  the  learning  process  in  animals  being  trained 
in  a  T-box  (1).  Many  other  examples  could  be  quoted. 

The  second  attraction  that  I  mentioned  was  the  way  information 
theory  handles  the  problem  of  signals-in-noise;  but  here,  neuro- 
physiologists generally  use  this  term  in  the  vernacular  rather  than 
in  its  critically  defined  sense.  This  is  because  we  do  not  usually 
apply  the  criteria  for  randomness  when  speaking  of  biological  noise. 
As  a  matter  of  fact,  many  use  the  term  'noise'  in  quite  the  opposite 
sense  from  that  defined  by  mathematical  theory.  In  the  neuro- 
physiological  journals  we  frequently  find  'noise'  used  to  describe 
disorderly,  unpredictable  activity  in  which  no  regularity  can  be 
detected. 

On  the  other  hand,  the  mathematical  approach  (5,  11)  has  a 
very  clear-cut  set  of  criteria  for  random  processes — criteria  based  on 
probability  distributions  that  effectively  result  in  statistical  regu- 
larity, statistical  orderliness  and  statistical  predictability. 

The  whole  gamut  of  criteria  for  a  mathematical  model  of  random 
processes  would  be  very  difficult  to  apply  to  the  nervous  system,  but 
already  some  consideration  has  been  given  to  this  problem  (6). 
The  probability  functions  that  have  seemed  to  be  the  least  difficult 
to  carry  from  the  mathematical  model  into  the  'real'  nervous 
system  have  been  those  of  means,  spectra  and  correlation  functions. 
These  comparatively  simple  factors  bring  us  only  to  a  limited  and 
fractional  descriptive  usefulness,  and  hence  an  increasing  number 
of  neurophysiologists  are  exploring  this  approach. 

The  statistical  regularity  of  a  random  process  bears  considerable 
interest  for  the  neurophysiologist  because  of  his  familiarity  with 
the  concept  of  the  statistically  steady  state  that  has  earned  itself  the 
name  of  homeostasis.  The  fact  that  the  brain,  in  its  evolution,  has 
reached  a  stage  in  man  where  the  neuronal  mechanisms  for  homeo- 
static  control  of  his  milieu  interieur  are  handled  by  his  medullary 
brain  stem,  frees  the  cortex  from  these  concerns  and  reserves  it  for 


How  Can  Models  from  Injormation  Theory  be  Used  in  Neurophysiology?    233 

higher  functions,  thus  giving  man  what  Claude  Bernard,  in  his 
famous  phrase,  called  "la  condition  de  la  vie  libre." 

This  concept  of  a  statistically  regular,  predictable  randomness 
of  'noise'  against  which  the  neurophysiologist  emphasizes  his 
'signal'  when  averaging  by  computers,  has  a  close  relationship  to 
one  of  the  most  basic  principles  of  information  theory.  This  prin- 
ciple is  that  information  is  carried  by  departure  from  orderliness  or, 
in  other  words,  by  departure  from  the  predictable.  Even  the 
intuitive  concept  of  information  is  a  change  from  what  you  already 
know  and  can  predict. 

Several  neurophysiologists  have  now  invoked  this  principle  to 
explain  such  phenomena  as  "attention"  and  "habituation"  and 
"the  orienting  reflex,"  together  with  their  attendant  electrical 
concomitants.  One  such  example  is  the  model  proposed  by  Sokolov, 
(12)  which  envisages  novelty,  i.e.,  departure  from  the  statistically 
expected  state,  as  being  the  factor  that  evokes  activity  in  the  brain 
stem  and  the  resultant  orienting  reflex.  The  concept  of  "attention" 
being  related  to  matching  the  probability  of  a  neuronal  event 
against  the  expected  distribution  of  possible  events,  will  be  found 
in  the  work  of  many  neurophysiologists.  * 

This  brings  us  to  the  third  major  attraction  of  information  theory 
for  the  neurophysiologist:  the  use  of  a  probabilistic  model  for  the 
nervous  system  rather  than  a  deterministic  one.  I  would  like  to 
approach  this  from  the  neurophysiologist's  angle. 

I  have  spoken  earlier  about  Johannes  Miiller  and  you  will  re- 
member his  famous  "Law  of  Specific  Nerve  Energies"  by  which 
each  of  the  myriad  facets  of  sensation  was  assigned  its  special  nerve 
— how  uneconomical,  but  how  simple.  A  single  output  would  be 
obtained  from  a  single  input.  Nothing  could  be  more  deterministic. 
One  could  design  no  simpler  code.  But  it  was  too  good  to  be  true. 

In  the  earlier  part  of  this  century,  the  belief  in  a  ubiquitous 
all-or-nothing  law  for  the  nervous  system  and  the  demonstrations  of 
the  coding  of  intensity  by  frequency  of  discharge  in  single  fibers 
of  the  peripheral  nervous  system,  led  eventually  to  exploration  of 
single  cell  discharges  within  the  brain  itself. 

Iminediately,  it  became  clear  that  coding  was  no  simple  problem. 
Miiller  would  have  been  chagrined  to  see  how  many  difl'erent 


"For  early  examples  see  references  (2)  and  (7). 


234  Information  Storage  and  Neural  Control 

peripheral  loci  could  fire  an  individual  cell  in  the  brain.  Many 
examples  drawn  from  the  somatic  and  other  sensory  systems  have 
been  published,  but  Miiller  would  have  been  even  more  dismayed 
had  he  been  shown  that  there  are  cells  within  the  brain  that  are 
no  respecters  of  sense  modalities.  Proof  has  been  given  of  con- 
vergence of  sensory  modalities  onto  individual  neurons  of  the 
midbrain  and  thalamic  brain  stem  and  onto  units  in  the  limbic 
system.  Even  cortical  neurons  are  not  simon-pure. 

The  complexities  do  not  cease  there,  for  such  coding  as  can  be 
established  for  individual  presynaptic  fibers  is  found  to  be  trans- 
formed at  the  synapse  and  to  send  on  a  different  pattern  of  discharge 
in  the  postsynaptic  output.  Some  of  the  exquisite  response  patterns 
that  investigators  have  been  able  to  identify  in  the  primary  neurons 
from  the  receptors  are  therefore  only  one  link  in  a  chain  of  se- 
quential codes.  Moreover,  the  evidence  that  recoding  in  post- 
synaptic fibers  varies  in  different  neuronal  aggregates  is  over- 
whelming, for  there  is  great  variation  in  the  degree  of  convergence 
and  divergence. 

The  biologist  has  long  known  how  rare  is  a  one-to-one  relation- 
ship between  input  and  output  of  a  synapse  but  is  now  beginning 
to  realize  that  the  relationships  may  not  even  be  linear.  He  may 
well  have  to  wait  for  the  mathematicians  to  progress  further  with 
their  analyses  of  nonlinear  systems  before  he  himself  can  master  the 
transformation  characteristics  of  the  code  as  it  passes  seriatim 
through  a  chain  of  synaptic  relays. 

As  an  example  of  changing  code,  the  findings  of  Whitfield  (13)  in 
the  auditory  system  may  be  quoted.  Whitfield  found  that  the  rate 
of  firing  becomes  progressively  less  as  the  impulses  proceed  through 
the  serial  synaptic  stations  on  their  way  to  the  cortex.  Moreover, 
the  rate  of  unit  firing  becomes  less  and  less  dependent  on  the 
strength  of  the  stimulus  at  each  successive  relay  station.  In  other 
words,  the  intensity  of  the  stimulus  is  no  longer  being  signalled 
simply  by  frequency  of  discharge.  The  coding  has  changed  and 
some  clues  to  its  nature  are  already  known.  These  point  to  the 
distribution  of  excitation  and  inhibition  among  the  fibers  of  the 
pathways  as  being  a  crucial  factor. 

Although  in  this  outline  which  poses  the  problem,  all  the  facets 
that  must  enter  into  any  consideration  of  neural  coding  cannot  be 


How  Can  Models  from  Information  Theory  be  Used  in  Neurophysiology?    235 

enumerated,  the  phenomenon  of  lateral  interaction  among  mem- 
bers of  a  neuronal  population  should  not  escape  attention.  This  is 
seen,  perhaps  most  strikingly,  in  the  zone  of  inhibition  that  develops 
around  a  locus  of  excitation.  The  "inhibitory  surround"  has  now 
been  demonstrated  for  the  visual,  auditory,  and  somatic  afferent 
systems  and  emphasizes  complex  patterns  of  interaction  rather 
than  conduction  over  isolated  paths.  There  is  also  the  interference 
with  direct  routing  of  impulses  from  the  receptor  to  the  cortex, 
mediated  by  recurrent  collaterals  and  centrifugal  feed-back  control 
over  afferent  pathways.  Of  two  more  contributions  to  knowledge 
which  have  added  to  the  neurophysiologist's  task,  one  is  the 
realization  that  the  all-or-nothing  discharge  is  a  comparatively  rare 
event  in  the  central  nervous  system,  graded  responses  (which  may 
or  may  not  lead  to  cell  discharges)  being  all  important.  Wliat  of 
these  graded  changes?  How  do  they  affect  the  code? 

One  aspect  of  the  problem  has  been  approached  by  the  analysis 
of  the  intervals  between  nerve  discharges.  Although  the  action 
potential  of  an  axon  is  all-or-nothing  and  hence  digital,  the  graded, 
analog  character  of  the  receptor's  action  can  be  preserved  in  the 
code  by  the  intervals  between  discharges,  for  intervals  between 
spikes  are  continuously  variable  and  therefore  can  transmit  graded 
input.  In  fact,  a  great  deal  of  work  in  many  laboratories  is  cur- 
rently being  devoted  to  pulse-interval  analysis  of  the  message  set  up 
by  stimulation  of  receptors. 

More  difficult  is  the  problem  of  graded  delivery  of  the  inessage 
at  the  higher  cerebral  level  where  its  result  may  be  effector  cell 
discharge,  passage  into  association  areas,  passage  into  storage 
neurons  changing  their  cellular  function,  modulation  of  other 
currently  incoming  messages,  or  dissipation  of  a  type  about  which 
we  have,  as  yet,  no  knowledge.  Graded  responses  in  dendrites  do 
not  necessarily  induce  discharges  of  their  cell  bodies;  nevertheless, 
their  influence  as  modulators  inay  be  critical  for  the  "meaning''  of 
the  message. 

Last,  but  not  least  in  importance,  is  the  evidence  for  a  great 
deal  of  endogenous  discharge  of  many  neurons  of  the  central 
nervous  system  in  the  absence  of  overt  external  stimuli.  How  is  the 
brain  to  select  those  discharges  that  are  evoked  by  messages  initiated 
in  its  environment  from  those  that  form  its  background  activity? 


236  Information  Storage  and  Neural  Control 

If  this  selection  can  be  made,  on  what  grounds  is  a  resukant  efferent 
discharge  determined? 

No  one  yet  knows  what  the  mechanism  effecting  these  decisions 
may  be,  but  it  has  occurred  to  many  that  discriminations  may  be 
made  by  the  brain  on  a  statistical  basis,  i.e.,  on  the  probability  that 
the  afferent  patterns  are  significantly  different  from  those  which 
are  currently  taking  place  in  the  brain  or  which  its  past  experience 
has  set  its  neurons  to  "expect"  by  a  change  in  their  cellular 
function. 

This  statistical  viewpoint  may  be  defined  as  the  "probabilistic" 
model  in  contrast  to  a  "deterministic"  one  in  which  a  given  stimulus 
elicits  a  stereotyped  response  irrespective  of  the  likelihood  of  its 
occurrence. 

The  probabilistic  approach  recognizes  the  need  for  the  brain  to 
assign  iinportance  to  those  signals  which  require  effector  action 
and  suggests  that  this  assay  of  importance  must  be  on  a  basis  of  the 
probability  of  the  signal  not  being  a  chance  variation.  With  all 
the  on-going  discharges  of  cerebral  neurons  that  workers  with 
microelectrodes  have  so  convincingly  demonstrated,  some  pro- 
cedure must  surely  take  place  before  a  'meaningful'  signal  can  be 
selected  from  this  incessant  activity. 

No  assessment  of  probability  can  be  made  without  averaging. 
Therefore,  those  who  have  begun  to  explore  a  statistical  model  for 
coding  in  the  nervous  system  have  turned  to  techniques  for  averag- 
ing neuroelectric  activity  over  the  passage  of  time  as  well  as  over 
space  as  represented  by  neuronal  aggregates.  To  aid  in  this  task 
many  have  adopted  a  prosthesis  Just  as  the  microanatomist  has 
adopted  the  microscope  as  a  prosthesis  to  enrich  his  visual  ability, 
so  has  the  neurophysiologist  begun  to  use  the  computer  as  a  pros- 
thesis for  his  calculating  ability. 

The  statistical  characteristics  of  spontaneously  discharging 
neurons  must  be  known  to  the  brain  before  it  can  react  appropri- 
ately to  an  unexpected,  meaningful  signal  requiring  action.  One 
might  even  speculate  that  the  nonresponding  but  spontaneously 
discharging  neurons  that  so  many  observers  have  found  with  their 
microelectrodes,  are  "comparison"  generators  and  the  responding 
neurons  "information"  generators.  If  this  were  so,  only  when  the 
normally  expected  difference  between  the  two  categories  of  genera- 


How  Can  Models  from  InJormatio7i  Theory  be  Used  in  Neurophysiology?     237 

tor  activity  was  exceeded  by  a  statistically  significant  amount 
would  the  incoming"  signal  be  meaningful. 

To  analyze  the  myriad  complexities  of  the  brain's  function  by 
nonstatistical  description  of  unit  discharges  is  too  gigantic  a  task 
to  be  conceived,  but  exploration  in  terms  of  probability  theory  is 
both  practical  and  rational. 

In  characterizing  nervous  activity,  therefore,  one  would  not 
attempt  the  precise  definition  that  arithmetic  demands  but  would 
seek  the  statistical  characteristics  of  the  phenomena  that  appear 
to  be  relevant.  The  margin  of  safety  that  the  brain  has  for  appropri- 
ate reaction  is  thus  much  greater  than  a  deterministic,  arithmetic- 
ally precise  operation  would  impose.  Chaos  would  result  from  the 
least  slip-up  of  the  latter,  whereas  only  a  major  divergence  from 
the  mean  would  disturb  a  system  working  on  a  probabilistic  basis. 
The  rigidity  of  arithmetic  is  not  for  the  brain,  and  a  search  for  a 
deterministic  code  based  on  arithmetical  precision  is  surely  doomed 
to  disappointment. 

Turning  now  to  the  scanty  data  which  are  all  that  today's  neuro- 
physiologist  has  as  yet.  In  terms  of  actual  data  culled  from  the  brain 
I  propose  to  mention  only  two  categories  here: 

1 )  The  averaging,  over  time,  of  intervals  between  unit  discharges 
in  the  brain; 

2)  The  averaging,  over  time  and  space,  of  activity  in  neuronal 
aggregates. 

An  example  of  averaging  units  over  time,  the  first  category,  is 
the  work  of  Mountcastle  (8)  in  which  he  has  been  designing  experi- 
ments to  test  the  hypothesis  that  an  intracortical  mechanism 
exists  which  integrates  frequency  over  short  periods  of  time  and 
responds  only  when  intervals  of  sufficient  brevity  occur.  These 
experiments  have  revealed  a  striking  change  in  pulse-interval 
distribution  in  circumstances  that  give  support  to  this  hypothesis. 

This  investigation  is  alluded  to  so  briefly  at  this  time  because 
it  is  being  quoted  solely  as  an  example  of  the  first  category  of 
statistical  approach,  i.e.,  averaging  over  time  only.  But  the  central 
nervous  system  must  have  some  mechanism  for  dealing  with 
multiple  complex  inflow,  and  it  would  seem  more  profitable  to 
expand  this  approach  to  the  second  category  of  study  that  I  men- 


238  Information  Storage  and  Neural  Control 

tioned;  namely,  averaging"  not  only  over  time  but  over  neural 
aggregates,  in  order  to  get  the  profile  of  a  population  of  neurons. 
This  is  of  particular  importance  in  the  brain  because  of  the  demon- 
strated interaction  of  units  within  populations.  This  second  ap- 
proach necessitates  the  use  of  electrodes  large  enough  to  record 
from  populations  of  neurons  and  thus  able  to  average  over  space 
as  well  as  time. 

I  will  illustrate  this  approach  by  brief  mention  of  some  examples 
drawn  from  our  laboratory.  Suppose  we  take  the  response  of  an 
unanesthetized  cat  to  a  flash  of  light  that  is  repeated  monoton- 
ously without  any  change  in  timing,  or  intensity,  or  in  any  other 
of  its  parameters. 

At  the  beginning  of  a  train  of  such  stimuli,  the  message  the  brain 
will  receive  will  contain  at  least  three  major  components; 

1)  the  stimulus  is  visual, 

2)  the  stimulus  is  repetitive, 

3)  the  stimulus  is  novel. 

On  prolonged  repetition,  however,  the  third  of  these  messages 
(that  the  stimulus  is  novel)  is  no  longer  being  sent.  The  probability 
of  its  arrival  is  now  very  high. 

If  the  hypothesis  is  to  be  regarded  as  tenable,  one  of  the  tests 
the  neurophysiologist  must  make  is  a  demonstration  that  the 
response  of  the  brain  to  a  novel  stimulus  is  difTerent  on  the  average 
from  its  reponse  to  a  familiar  one. 

What  would  be  demanded  by  the  hypothesis  under  discussion? 
Averages  of  a  sample  of  responses  late  in  the  series  would  be  ex- 
pected to  carry  two  of  the  same  components  of  the  message  as  are 
carried  by  the  first  set  of  flashes;  namely,  that  the  stimulus  is  visual 
and  that  it  is  repetitive,  but  the  third  component,  i.e.,  that  the 
stimulus  is  novel,  would  need  some  change  of  signal. 

When  the  responses  to  a  repetitive  train  of  flashes  are  recorded 
from  the  visual  cortex  of  an  unanesthetized  cat  with  permanently 
implanted  electrodes,  one  finds  that  the  short  latency  responses 
that  have  been  identified  with  transmission  in  the  specific  aflferent 
systems  persist  for  the  whole  duration  of  the  train.  They  apparently 
carry  the  first  two  components  of  the  message  (that  the  stimulus  is 
visual  and  that  it  is  being  repeated). 


How  Can  Adodels from  Information  Theory  be  Used  in  Neurophysiology?     239 

At  the  beginning  of  such  a  train  of  flashes  there  are  also  long 
latency  responses  in  the  visual  cortex.  These  have  been  shown  to 
reach  the  cortex  by  the  nonspecific  afferent  systems  of  the  midline 
brain  stem  and  thalamus.  It  would  seem  possible  that  tlie  third 
major  component  of  the  message,  the  one  signalling  'novelty'  in 
the  stimulus,  may  be  carried  by  these  nonspecific  afferents,  for  as 
repetition  continues,  this  sequence  of  later  waves  fades  out.  Averag- 
ing of  the  first  sixty  to  arrive,  then  the  second  sixty,  the  third  sixty, 
and  so  on,  shows  this  late  component  of  the  multiple  response  to 
be  dropping  out  as  the  novelty  wears  off. 

The  effect  can  be  fractionated  even  farther  in  the  nonspecific 
system  by  actually  recording  in  a  nucleus  of  this  midline  nonspecific 
system  (the  centre  mechan)  where  one  of  the  most  prominent  of  its 
average  electrical  responses  to  flash  (the  late  wave)  can  be  seen  to 
fail  with  repetition  of  the  unchanging  stimulus,  while  the  earlier 
components  persist. 

The  serial  change  in  the  late  component  of  the  multiple  response 
is  very  marked.  Whatever  the  mechanism  for  this  depressed  respon- 
siveness may  prove  to  be,  it  is  tempting  to  propose  that  this  forms 
part  of  the  mechanism  that  conveys  presence  or  absence  of  novelty. 

This  work  has  been  described  in  detail  and  illustrated  elsewhere, 
(3,  4)  so  it  will  not  be  given  more  space  here.  However,  lest  these 
examples  appear  to  suggest  too  simple  a  picture  of  the  brain's 
message-receiving  systems,  let  me  add  that  not  only  does  one  find 
presence  or  absence  of  a  component  of  the  response,  as  novelty 
wears  off,  as  in  the  foregoing  sample,  but  one  also  finds  situations 
in  which  the  time-relationships  of  the  components  of  the  brain's 
electrical  responses  may  change.  Possibly  it  is  in  the  time  domain 
that  the  neurophysiologist  will  find  the  most  clues  for  the  solution 
of  this  problem. 

I  make  only  a  brief  allusion  here  to  the  laboratory  work.  This 
is  not  intended  as  the  report  of  a  research,  but  as  an  example  of 
work  carried  out  with  a  probabilistic  model  in  mind,  and  to  illus- 
trate the  point  that  the  response  probabilities  of  the  nervous  system 
are  influenced  by  the  past  events  it  has  experienced. 

Returning  now  to  the  more  general  topic  of  the  utilization  by 
neurophysiology  of  information  theory,  let  us  not  forget  that  one 
of  the  innovations  of  information  theory  was  the  axiom  that  in- 


240  Information  Storage  and  Neural  Control 

formation  is  measurable,  and  tliat,  in  fact,  Shannon  in  his  classical 
paper  gave  a  precise  mathematical  definition  of  information.  It  is 
so  difficult  to  define  information  measures  for  ensembles  in  biology 
that  most  biologists  who  use  information  theory  usually  do  not 
attempt  to  do  so  in  a  quantitative  way.  Generally,  they  do  not 
actually  measure  the  information;  and  hence,  they  fail  to  exploit 
the  full  potentialities  of  the  theory.  Yet  many  feel  that  someday, 
somehow,  more  exactly  defined  information  measures  may  be 
brought  into  neurophysiology.  I  need  only  mention  as  an  example 
Shannon's  formulation  of  the  problem  of  channel  capacity  and  his 
solution  for  dealing  with  equivocation.  C^hannel  capacity  is  surely 
a  basic  factor  in  the  communication  functions  of  the  nervous  system. 

It  is  so  tempting  to  think  of  information  transfer  in  the  brain 
as  being  simply  a  matter  of  transniission  in  specific  nerve  tracts. 
If  this  simple-minded  concept  could  for  one  moment  be  defended, 
one  would  then  begin  to  study  such  communication  channels  in 
terms  of  the  finite  set  of  signals  that  can  be  initiated  in  the 
channel,  the  set  of  signals  that  arrives  and  the  probability  of  the 
reception  of  any  given  signal.  If  only  the  brain  worked  like  a  simple 
telegraph  system  we  would  immediately  be  able  to  make  precise 
statements  about  such  things  as  channel  noise  and  would  be  able 
to  calculate  channel  capacity. 

In  contrast,  all  the  work  that  the  neurophysiologists  have 
pursued  has  revealed  to  us  the  enormity  of  interaction  within  the 
brain — the  correlations,  couplings,  linkages,  and  statistically  inter- 
dependent elements  that  contribute  to  its  organization  and  make 
any  measurement  of  its  interacting  ensembles  or  any  mathematical 
statement  of  its  entropy  conditions  formidable  in  the  extreme. 

In  closing,  let  me  say  that  the  application  of  quantitative  in- 
formation theory  to  neurophysiology  lies  largely  in  the  future. 
Possibly  a  partial  answer  to  the  question  in  the  title  of  this  paper 
is  that  if  information  theory  has  not  led  to  the  uncovering  of  many 
new  facts  in  neurophysiology,  it  may  have  led  to  many  new  ideas. 

REFERENCES 

1.  Adey,  W.  R.,  Dunlop,  C.  W.,  Hendrix,  C.  E.:  Hippocampal  slow 
waves;  distribution  and  phase  relations  in  the  course  of  approach 
learning.  AM  A  Arch.  Neurol.,  3.-74-90,  1960. 


How  Can  Models  from  Information  Theory  be  Used  in  Neurophysiology?    241 

2.  Bates,  J.  A.  V.:  Significance  of  information  theory  to  neui-ophysiology. 

In:  Information   Theory  Symposium.  London,   1950,  p.   137. 

3.  Bi'azier,  M.  A.  B.:  Responses  in  non-specific  systems  as  studied  by 

averaging    teclmiques.    In:    Specific    and    Unspecific    Mechanisms    of 
Sensory-Motor  Integration,  Ed.  G.  Moruzzi  (in  press). 

4.  Brazier,   M.   A.   B.:    Information   carrying  characteristics   of  brain 

responses.    In:    The  Physiological  Basis  of  Mental  Activity,   Ed.    R. 
Hernandez-Peon  (in  press). 

5.  Davenport,  W.  F.,  Root,  W.  L.:  An  Introduction  to  the  Theory  of  Ran- 

dom Signals  and  Noise.  New  York,  McGraw-Hill,  1958. 

6.  Goldstein,  M.  H.:  Averaging  techniques  applied  to  evoked  responses. 

In:   Computer   Techniques  in  EEC  Analysis,   Ed.   M.   A.   B.   Brazier, 
EEG.  Clin.  Neurophysiol.,  Supp.  20,  1962,  p.  59. 

7.  Grey  Walter,  W.:  In:  Brain  Mechanisms  and  Consciousness,  Ed.  J.  F. 

Delafresnaye,    Oxford,    Blackwell    Scientific    Publications,    1954, 
p.   372. 

8.  Mountcastle,   V.   B.:   Duality   of  function   in   the   somatic   aflferent 

system.  In:  Brain  and  Behavior.  Ed.  M.  A.  B.  Brazier,  Washington, 
D.  C.,  American  Institute  of  Biological  Sciences,  1961,  p.  67. 

9.  Shannon,   C.  W.:  A  mathematical  theory  of  communication.  Bell 

System  Tech.  J.,  27.-379-424;  623-657,  1948. 

10.  Sherrington,  C.  S.:  Man  on  His  Nature.  London,  Cambridge  University 

Press,  1951. 

11.  Siebert,  W.  M.:  The  description  of  random  processes.  In:  Processing 

of  Neuroelectric   Data,    Tech.    Report    351,    Communications   Bio- 
physics, RLE  MIT,  1959,  p.  66. 

12.  Sokolov,  E.  N.:  Neuronal  model  and  the  orienting  reflex.  In:  Central 

.Kervous  System  and  Behavior,   Ed.   M.   A.   B.   Brazier,   New  York, 
Josiah  Macy,  Jr.  Foundation,  1960,  p.  187. 

13.  Whitfield,  I.  C:  The  physiology  of  hearing.  In:  Progress  in  Biophysics 

and  Biophysical  Chemistry  8:\-Al ,  1957. 

14.  Wiener,  N.:  Cybernetics,  New  York,  John  Wiley,  1948. 

DISCUSSION  OF  CHAPTER  X 

Harold  W.  Shipton  (Iowa  City,  Iowa):  In  the  averaged 
record  that  you  showed,  were  the  stimuli  being  delivered  ran- 
domly, or  were  they  as  regular  as  you  could  make  diem?  Would 
you  care  to  comment  on  whether  the  time  characteristics  of  the 
external  drive  signal  appear  to  you  to  be  important  in  the  con- 
struction of  the  model? 


242  Information  Storage  and  Neural  Control 

Mary  A.  B.  Brazier  (Los  Angeles,  California) :  They  were 
certainly  not  absolutely  random.  On  the  contrary,  the  intervals 
between  stimuli  were  as  constant  as  we  could  make  tliem.  In  an 
experiment  such  as  tliis,  there  is  very  great  difficulty  for  the 
neurophysiologist  because  the  responses  depend  so  much  on  the 
state  of  the  animal.  Although  one  would  like  to  have  a  longer 
interval  between  stimuli,  it  is,  in  my  experience,  almost  impossible 
to  hold  an  animal  in  the  same  stage  of  the  sleep-wakefulness 
continuum  for  as  many  stimuli  as  we  use,  if  the  interval  between 
flashes  is  longer  than  one  second. 

L.  M.  N.  Bach  (New  Orleans,  Louisiana) :  I  am  curious  about 
the  disappearance  of  the  second  component  in  tlie  centrum 
medianum  response  with  repetitive  stimulation  as  a  possible 
inverted  index  of  post-tetanic  potentiation.  Do  you  consider  it 
a  testable  proposition  that  the  disappearance  of  the  second  com- 
ponent could  be  correlated  with  post-tetanic  potentiation,  or  do 
you  consider  tliat  there  is  no  relationship  at  all? 

Brazier:  It  should  be  testable,  but  it  is  rather  difficult  to  design 
an  experiment  in  wliich  to  test  this. 

Gregory  Bateson  (Palo  Alto,  California):  Would  you  have 
expected  the  part  of  the  signal  which  denotes  novelty  to  follow 
the  other  two  components?  Would  it  not  have  been  a  better 
arrangement  to  have  the  system,  when  it  had  diagnosed  novelty, 
transmit  the  information  ahead  of  the  other  components  of  the 
signal? 

Brazier:  I  had  no  "expectation,"  tliough  now  that  you  raise 
the  question,  would  you  not  expect  the  brain  to  need  to  receive 
the  signal  before  it  could  assess  its  novelty?  What  you  have  sug- 
gested would  make  a  very  good  design  for  a  communication 
system,  although  the  nervous  system  does  not  appear  to  be  de- 
signed in  this  manner. 


T. 


CHAPTER 
XI 

NEURAL  MECHANISMS  OF 
DECISION  MAKING* 

E.  Roy  John,  Ph.D.** 
GENERAL  CONSIDERATIONS  ABOUT  MEMORY 


HIS  paper  is  largely  concerned  with  the  rather  specialized 
decision-making  involved  when  a  cat  decides  which  of  two  previ- 
ously experienced  frequencies  of  flickering  light  is  being  pre- 
sented. Since  the  constituent  flashes  of  the  two  flicker  frec[uencies 
are  identical,  such  decision-making,  or  differential  discrimination, 
would  seem  difficult  to  perform  on  the  basis  of  the  instantaneous 
quality  of  the  stimulus.  In  contrast  to  existential  discriminations, 
based  on  the  presence  or  absence  of  a  stimulus,  differential  dis- 
crimination of  this  sort  logically  would  seem  to  require  the  nervous 
system  to  analyze  the  temporal  sequence,  or  pattern,  of  stimulation. 
Although  one  can  conceive  of  possible  alternate  niechanisms 
for  the  mediation  of  such  behavior  by  time-measuring  devices  or 
filter  networks,  a  plausible  mechanism  for  the  analysis  of  sequential 
stimuli  would  be  a  coincidence  detector  which  compared  patterns 
of  incoming  activity  with  patterns  generated  by  a  stored  representa- 
tion of  previously  experienced  sequences — a  memory.  This  hy- 
pothesis, with  some  relevant  electrophysiological  evidence,  has  been 
presented  in  detail  elsewhere  (7,  9).  My  purpose  here  is  to  review 


*The  work  described  in  this  paper  was  supported  in  part  by  Research  Grant 
MY-2972  from  the  National  Institute  of  Mental  Health,  and  Grant  G21831  from 
the  National  Science  Foundation. 

**The  author  takes  pleasure  in  acknowledging  the  kindness  of  Marc  Weiss  for 
making  available  the  data  shown  in  Figures  6,  7.  8,  and  9,  and  the  assistance  of  Arnold 
L.  Leiman  and  Anthony  L.  F.  Gorman  in  acquisition  of  portions  of  the  data  here 
reported. 

243 


244  Information  Storage  and  Neural  Control 

that  evidence  and  to  supplement  it  with  a  number  of  recent  findings 
in  our  laboratories  which  will  also  serve  to  illustrate  some  technical 
innovations  we  were  utilizing  for  these  purposes. 

Before  I  undertake  this  task,  I  wish  to  emphasize  that  the  hy- 
pothesis stated  does  not  imply  the  mediation  of  memory  by 
regenerative  electrical  activity.  The  large  literature  on  the  con- 
solidation process  reviewed  recently  (4)  shows  that  there  are  at 
least  two  phases  of  memory  storage:  1)  An  early,  labile  consolida- 
tion phase,  in  which  the  representation  of  a  recent  experience  is 
susceptible  to  severe  interference  or  destruction  by  numerous 
chemical  or  electrical  perturbations,  and  during  which  memory 
may  well  consist  of  persisting  electrical  patterns  of  a  reverberatory 
sort;  and,  2)  a  later  stable  phase  in  which  such  perturbations  have 
no  effect,  and  during  which  memory  is  stored  in  some  other  fashion, 
perhaps  as  a  structural  modification.  This  necessitates  a  coupling 
mechanism  whereby  the  reverberatory  electrical  activity  main- 
tained during  the  consolidation  phase  gradually  stipulates  the 
structural  change  which  will  serve  to  represent  it.  A  number  of 
workers  have  discussed  the  possibility  that  such  structural  changes 
might  be  the  specification  of  macromolecular  configurations 
(5,  6,  25);  and,  as  Dr.  Morrell  has  told  you,  a  number  of  labora- 
tories, including  his  and  mine,  have  presented  data  suggesting  that 
ribonucleic  acid  (RNA)  may  play  a  role  in  this  function  (1,  2,  3, 
11,  18).  Whether  or  not  RNA  does  participate  in  long-terin 
memory  storage,  it  seems  reasonable  at  present  to  assume  that 
some  form  of  long-term  structurally  mediated  storage  does  exist. 
Various  data  seem  to  require,  further,  that  the  postulated  coupling 
between  electrical  patterns  and  the  long-term  storage  device  be 
reversible — that  the  pattern  of  iterated  or  sustained  electrical 
activity  stipulate  some  representational  structural  modification, 
and  that  this  structural  modification  be  able  to  generate  an  elec- 
trical pattern  identical  to  the  one  which  established  it. 

Time  does  not  permit  detailed  review  here  of  the  evidence  which 
I  believe  is  relevant  to  the  dynamics  by  which  such  a  representa- 
tional system  is  built,  but  such  a  detailed  discussion  has  been  pre- 
sented elsewhere  (7).  For  our  present  purposes,  I  hope  it  will  suflSce 
to  summarize  what  I  consider  to  be  the  salient  characteristics  of 
these  representational  systems:  1)  The  repeated  occurrence  of  as- 


Neural  Mechanisms  of  Decision  Making  245 

sociated  neural  activity  in  anatomically  extensive  regions  of  the 
nervous  system  causes  a  functional  relationship  to  become  estab- 
lished between  these  regions.  Subsequent  to  such  association, 
stimulation  of  one  region  causes  a  response  to  occur  in  other  regions, 
although  this  response  did  not  occur  before  the  associated  activity; 
2)  such  altered  response  relationship  cannot  be  interpreted  as 
merely  a  reflection  of  altered  threshold,  since  Morrell  has  shown 
that  the  new  response  is  differential,  and  is  displayed  only  to  the 
stimulus  to  which  the  association  was  established  and  not  to  closely 
similar  stimuli;  and  3)  discharge  can  occur  from  such  a  representa- 
tional system  with  a  temporal  pattern  which  reflects  the  pattern  of 
stimulation  while  it  was  established. 


TRACER  STIMULI,  LABELED  POTENTIALS, 
AND  INFORMATION 

The  bulk  of  the  data  which  I  wish  to  present  here  was  obtained 
in  studies  of  changes  in  the  electrophysiological  response  to  inter- 
mittent stimuli  during  the  establishment  of  conditioned  responses. 
The  technique,  used  most  profitably  before  us  by  Livanov  and 
Polyakov  (14),  was  applied  by  Killam  and  me  in  our  studies  of 
conditioned  avoidance  and  approach  responses  in  cats  (8,  9).  We 
reasoned  that,  whatever  the  nature  of  the  new  responses  established 
in  the  brain  during  conditioning,  such  responses  should  appear 
fairly  reliably  whenever  the  stimulus  was  presented.  We  selected 
an  intermittent  light  flash,  which  we  called  a  "tracer  conditioned 
stimulus"  (TCS),  and  searched  the  electrical  activity  of  the  brain 
for  the  appearance  of  waveforms  at  the  frequency  of  the  TCS,  which 
were  called  "labeled  potentials."  Such  a  procedure  greatly  en- 
hances one's  ability  to  detect  stimulus-bound  signal  in  the  midst  of 
the  tremendous  amount  of  ongoing  business  in  the  nervous  system. 
The  appearance  of  labeled  potentials  in  a  structure  during  the 
presentation  of  a  TCS  is  sufficient  evidence  to  conclude  that  in- 
formation about  the  TCS  is  reaching  that  structure.  It  is  clear  that 
such  labeled  potentials  are  not  necessary  for  a  structure  to  be  in- 
fluenced. A  structure  which  shows  no  labeled  potentials  can  be 
receiving  information  about  a  TCS. 


246  Itiformation  Storage  and  Neural  Control 

In  recent  reviews,  both  Morrell  (19)  and  I  (6)  have  summarized 
the  large  amount  of  data  obtained  in  many  laboratories  from  many 
different  species  of  experimental  animals,  showing  that  striking- 
changes  in  the  amplitude  and  distribution  of  labeled  potentials 
take  place  during  the  establishment  of  conditioned  responses  to 
intermittent  stimuli.  Although  the  appearance  of  labeled  potentials 
in  a  structure  justifies  the  conclusion  that  information  about  the 
presentation  of  a  TCS  is  reaching  that  structure,  one  cannot  assume 
that  such  labeled  potentials  actually  are  the  neural  coding  of 
information  about  stimulus  frequency.  Labeled  potentials  may 
simply  be  nonfunctional  correlates  of  the  actual  processing  by 
nerve  cells  of  otherwise  coded  information  about  the  TCS.  Con- 
versely, one  cannot  prove  on  the  basis  of  present  evidence  that 
labeled  potentials  are  not  the  effective  neural  code  for  stimulus 
frequency. 

ASSIMILATION  AND  MEMORY  TRACE 

Many  phenomena  observed  in  earlier  work  directed  my  at- 
tention to  this  problem  because  they  suggested  a  functional  role 
for  labeled  potentials.  The  first  of  these  phenomena  was  called 
"assimilation  of  the  rhythm"  by  Livanov  (14),  who  first  observed  it. 
It  has  since  been  described  by  many  workers  utilizing  various 
species  in  diverse  experimental  situations  (6,  19).  If  one  studies 
the  resting  electrical  activity  of  various  brain  regions  in  an  animal 
learning  a  conditioned  response  to  an  intermittent  stimulus,  one 
observes  that  during  the  intertrial  intervals  a  marked  hypersyn- 
chrony  appears  at  the  stimulus  frequency,  or  at  a  harmonic  thereof. 
This  spontaneous,  frequency-specific  activity  can  dominate  the 
resting  electrical  activity  in  early  stages  of  learning.  In  our  experi- 
ence, it  tends  to  diminish  and  disappear  as  the  conditioned  response 
becomes  well  established,  but  will  return  briefly  following  per- 
formance of  an  erroneous  response.  Figure  1  illustrates  assimilation 
and  is  taken  from  a  paper  by  Killam  and  me  (8).  Note  that  the  slow 
hypersynchrony,  in  this  case  at  one-half  the  stimulus  fr-equency, 
appears  in  the  reticular  formation,  fornix,  and  septum  in  close 
relationship.  Assimilated  rhythms,  in  our  experience,  appear 
earliest,    are    most    marked,    and    persist    longest    in    nonspecific 


Neural  Mechanisms  of  Decision  Making 

SPOHTAWOUS    ACTIvmr   AT  WFFEREMT    STAGES      OF    TNAININO 


247 


...     ^  I. 


M    TNAINM    Mr    10%) 


KMh    TMUNMa  DAY   (24%) 


I 

I 

20t«   TKAININS    Mr   (*9%l 

Fig.  1. 

CON — Bipolar  transcortical  (visual)  derivation 

IPSI — Bipolar  derivation  from  the  same  optic  gyrus 

RF — Midbrain  reticular  formation 

SUP  COLL — Superior  colliculus 

FX — Fornix 

SEP — Septum 

AUD — Auditory  cortex 

AMYG — Lateral  amygdaloid  complex 

POST  HIPP — Dorsal  hippocampus 
"Assimilation  of  rhythm"  during  avoidance  training  using  ten  per  second  flicker 
as  conditioned  stimulus.   (From  John,  E.  R.  and  Killam,  K.   F.:  J.  Pharmacol. 
Exp.  Therap.,  725:252-274,  1959.) 


248  Information  Storage  and  Neural  Control 

regions.  Other  workers  have  reported  that  assimilation  of  the 
rhythm  appears  only  in  the  training  situation  and  is  not  observed 
when  the  animal  is  in  his  home  cage.  Such  observations  demonstrate 
that  the  brain  has  the  capacity  to  generate  temporal  patterns  of 
potentials  which  are  substantially  the  same  as  those  elicited  by  a 
previously  experienced  stimulus.  Although  obtained  under  different 
experimental  conditions,  the  phenomena  described  by  Morrell  and 
his  colleagues  in  studies  of  cortical  conditioning  (17,  20,  21)  and 
by  Stern  et  al.  (24)  in  their  studies  of  trace  conditioning  provide 
additional  evidence  of  this  capacity.  It  is  of  interest  that  one  can  see 
these  assimilated  rhythms  appearing  with  apparent  simultaneity  in 
regions  relatively  distant  from  each  other,  as  if  an  anatomically 
extensive  system  were  activated.  One  might  reasonably  ask  whether 
such  sustained  patterns  in  the  absence  of  a  previously  experienced 
stimulus  do  not  reflect  neural  processes  which  represent  that  ex- 
perience. These  endogenously  generated  potentials  may  be  a  mani- 
festation of  the  elusive  "memory  trace." 

ASSIMILATED  RHYTHMS  AND  GENERALIZATION 

Some  evidence  presented  earlier  by  Killam  and  me  is  com- 
patible with  a  functional  role  for  such  endogenously  generated 
frequency-specific  patterns.  We  observed  that  an  animal  trained 
to  perform  a  conditioned  avoidance  response  to  a  ten  per  second 
flickering  light  characteristically  displayed  twenty  per  second  po- 
tentials in  the  visual  cortex,  as  seen  in  Figure  2.  On  presentation  of 
a  seven  per  second  flicker  after  the  animal  reached  criterion  to  the 
ten  per  second  flicker,  the  animal  showed  evidence  of  generaliza- 
tion by  repeatedly  performing  the  conditioned  response  to  the 
new  stimulus  frequency.  Examination  of  the  electrical  records 
showed  that  the  response  of  visual  cortex  to  the  seven  per  second 
flicker  was  a  twenty  per  second  potential,  as  is  visible  in  Figure  3A. 
The  arrow  denotes  the  beginning  of  the  behavioral  response. 
After  repeated  presentation  of  the  seven  per  second  flicker,  the 
animal  no  longer  performed  the  generalized  response  but  sat 
quietly.  At  this  time,  the  seven  per  second  flicker  elicited  pre- 
dominantly seven  per  second  activity  in  the  visual  cortex.  Presenta- 
tion of  the  original  ten  per  second  conditioned  stimulus  at  this 
point  failed  to  elicit  performance  of  the  conditioned  response  for 


Neural  Mechanisms  of  Decision  Making  249 

SIGNAL ^AAA/VV^AAA-/V^AAAAJ^AAAAAAAAAAAAAA 


ipsi  vis  ^ 

SEP 

AUD  I 

VA  f-       ^ 

AMYG  I 

Fig.  2. 

LG — Lateral  geniculate 

VA — Nucleus  ventralis  anterior.  Abbreviations  otherwise  as  in  Figure  1. 
Characteristic  electrical  response  on  presentation  of  the  ten  per  second  flicker 
conditioned  stimulus  to  the  fully  trained  animal  (100%  performance).   (From 
John,  E.  R.  and  Killam.  K.  F.:  J.  Pharmacol.  Exp.  Therap.,  125:2S2-274,  1959.) 

several  trials,  during  which  a  slow  wave  at  about  seven  per  second 
could  be  observed  in  cortex,  as  shown  in  Figure  3B.  When  per- 
formance reappeared  to  the  ten  per  second  stimulus,  twenty-per- 
second  potentials  again  were  elicited  in  cortex. 

Another  example  of  this  is  provided  by  Figure  4,  which  illustrates 
recordings  obtained  during  generalization  to  a  ten  per  second 
flicker  by  a  cat  previously  trained  to  perform  a  conditioned  avoid- 
ance response  to  a  four  per  second  flicker.  Note  that  upon  presenta- 


250 


Information  Storage  and  Neural  Control 


C 
O 


s  ^ 


it:     „ 

TD      O 


-a 

c 
o 


V 


«J 


c 
o  .i^ 

he  ^ 
C     C 


2     3 

II 

o  •-: 


V 


-5    aj  |j-l    a. 


o  ^ 


m 


s    c  .o; 


03      lU 

be  — . 
bO 


C    .i3    ^ 


3 

C 
O 


i3     o 


<^1      en 


o   -S^  r^ 


^ 


C 


o    S 


< 
bb  I-  ' 


Neural  Mechanisms  of  Decision  Making 


251 


> 


:^ 


-^.  4 


J 


'■%- 


.^ 


>. 


^, 


*. 

^ 

f^ 

'j; 

% 

1 

V 

'    IM 

.5 

t 

•*>. 

"5 

r 

s. 

\ 

,y 

_>- 

^ 

'-4 

v. 

J. 

r" 

V 

"V 

i. 

-<■' 

'"'^-^ 

> 

"^4 

'<. 

?• 

y' 

^ 


> 


J3     ■"    'i^ 


.2   =5  -B 


3 
"3    "O 


o  ;5 

'T3  .a    ^ 

:5  '5  'a 

.-c  Qh      U 

-owe 

t)  o    c 

c^  V  .^ 

O  J  pri 

Q  I     I 

Q  "^    ^ 


MOO 


™      O      1- 


O 
< 


CLqn 


?3    a 


!^  i-i  pd  J 


1^ 


5    5P 

3  -^ 


-t 


^-.       O 

u     > 

o    i! 


^ 


Vi    '^' 


J> 


;r~      ^ 


? 

V- 


"¥: 


1^     !^     C    O 


o    c    S    ^  ^    5^ 

^  .^  -?. «  ^  - 


C/J 

i5 

0; 

C3 

0 
c 

1    < 

< 

u 

Q 

CO 

u 

1 

0 

P.  h 

H 

S 

0 

^§ 

^ 
W 

h 

■fi 

>  0 

u 

^ 

w 
u 

0 
0 

c 
0 
u 

0 

5P 


252  Information  Storage  and  Neural  Control 

tion  of  the  ten  per  second  flicker,  potentials  at  the  same  frequency 
clearly  appear  in  visual  cortex  and  lateral  geniculate,  with  less 
marked  evidence  of  response  in  the  intralaminar  nuclei  and  the 
reticular  formation.  The  cortical  response  suddenly  shifts  to  a 
hypersynchronous  slow  wave  at  a  frequency  between  four  and  five 
cycles  per  second,  while  the  animal  shows  a  startled  movement 
and  four  seconds  later  performs  the  conditioned  lever  press  estab- 
lished to  a  four  per  second  flicker.  Notice  that  the  lateral  geniculate 
maintains  ten  per  second  potentials  during  this  period,  although 
potentials  at  lower  frequency  are  visible  in  the  reticular  formation 
and  occasionally  in  the  intralaminar  nuclei. 

A  comparable  observation  has  been  reported  by  Majkowski  (16). 
After  a  rabbit  was  trained  using  a  three  per  second  light,  generaliza- 
tion was  obtained  upon  presentation  of  a  five  per  second  light.  As 
can  be  seen  in  Figure  5,  during  such  a  generalization  a  three  per 
second  wave  can  be  observed  in  motor  cortex,  although  the  re- 
sponse of  visual  cortex  is  at  five  per  second.  Related  findings  have 
been  described  by  other  workers  (6). 

Data  of  this  sort  suggest  that  during  generalization  a  neural 
system,  which  has  become  established  as  a  consequence  of  experi- 
ence with  the  intermittent  conditioned  stimulus  (CS),  is  somehow 
released  by  the  new  stimulus,  but  discharges  with  the  character- 
istic temporal  pattern  of  the  original  conditioned  stimulus.  This 
system  seems  to  include  the  mesencephalic  reticular  formation  and 
the  intralaminar  nuclei  in  association  with  the  visual  cortex.  It  is 
interesting  that  during  generalization  phenomena  of  the  sort 
described,  regions  of  cortex  other  than  the  region  of  the  conditioned 
stimulus,  such  as  ectosylvian  or  medial  suprasylvian,  tend  to  display 
potentials  at  the  frequency  of  the  peripheral  stimulus. 

Additional  data  on  this  phenomenon  have  recently  been  obtained 
in  our  laboratory  by  Marc  Weiss  (26)  who  trained  a  cat  to  perform 
a  conditioned  avoidance  response  to  a  four  per  second  flickering 
light.  After  establishment  of  this  response,  the  cat  generalized 
readily  to  a  ten  per  second  flicker.  Figure  6  (Top)  shows  the 
EEC's  obtained  from  various  brain  regions  during  such  generaliza- 
tion. Note  the  irregular  slow  activity  in  the  visual  cortex  contrasted 
with  the  regular  ten  per  second  response  in  the  lateral  geniculate. 
Figure  6  (Bottom)  shows  records  obtained  after  differentiation  of 


Neural  Mechanisms  of  Decision  Making 


253 


*-       ■-    i_A_     *-   — *— 


-\>^^- 


R.  MOTOR 


L.  VISUAL 


SEC. 


50>iV 


I    I    i    I    1    »   1    >   I    I    I    I 


L.  EMG 


^\r  -~\ 


\: 


R.  EMG 


^'»^- 


Fig.  5. 

L.  MOTOR — Left  motor  cortex 

R.  MOTOR — Right  motor  cortex 

L.  VISUAL — Left  visual  cortex 

L.  EMG— EMG  of  left  hind  limb 

R.  EMG— EMG  of  right  hind  limb 
Electrical  responses  to  five  per  second  flicker  during  generalization  of  right  hind 
leg  flexion  response  after  training  with  a  three  per  second  flicker  (rabbit).  (From 
Majkowski,  J.:  Acta  Physiologica  Polomca,  /A'( 5): 565-581,  1958.) 


the  conditioned  response,  during  which  the  animal  was  taught  to 
discriminate  between  ten  and  four  per  second  flicker.  Note  that 
the  visual  cortex  now  displays  markedly  increased  regularity  of 
ten  per  second  potentials  during  the  ten  per  second  flicker. 

As  is  evident  from  the  stimulus  trace,  these  two  records  were 
obtained  using  a  "limp  circuit"  which  periodically  deleted  a  flash 
from  the  flicker  train.  The  purpose  of  this  technique  was  to  attempt 
to  evaluate  the  extent  to  which  endogenously  generated  potentials 
would  "fill  in"  the  period  of  the  deleted  flash.  Sufficient  data  have 
not  yet  been  obtained  to  warrant  discussion  of  this  aspect  of  the 
records,  and  it  is  not  central  to  our  present  purpose. 


254  Information  Storage  and  Neural  Control 

GENERALIZATION    TO    10   cp»    FLICKER  AFTER    TRAINING   TO    4  cos  SOIIV 

R  VIS  ex.  '    'J    t     J    i     ■. 

L  VISClt.1*^'>,l'-'-  '  . 

"  ^'  '■'■''''    ■,  '  '     .        ,     ' 

L.IAT  «t»(»^^'   .'','"''.'■•",    /'•»."'•'     ■''■''■■-' 


IS.  WW  .      J 
L.  HP-VW^yiJVnV.'AV 


.KIM-——'-' — ,  ,,    -, ,  „.^,^,_^„^ ,  ,  .1  J  J     ,■■,,,  J...     ..,...,..     ,,,,f,ff       '•■'"" 

'       I  sec       ' 
RESPONSE   TO    10    cps    FLICKER    AFTER  DIFFERENTIATION  lOpv 

L  OORS  ner  .■**''VV>'-■^V^^''■'v'''"»--•-^v'•■^-A/•%'"v■A'J^«^A<;v■-•r--^'^^"^v.■'■^^ 

Li«s  «M^*,,^_,w\V'^W"f-^V^VvV''''A'^vv\WVi.>A^^^^^^ 

Fig.  6. 

R.  NUC.  RET. — Right  nucleus  reticularis 
L.  MSS  CX — Left  medial  suprasylvian  cortex 
R.  VIS  CX— Right  visual  cortex 
L.  VIS  CX — Left  visual  cortex 
L.  LAT.  GEN. — Left  lateral  geniculate 
L.  DORS.  HIPP.— Left  dorsal  hippocampus 
L.  RF — Left  mesencephalic  reticular  formation 
L.  CM — Left  centre  median 
(Histological  verification  not  yet  available.)    {Top)   Electrical  responses  to   ten 
per  second  flicker  during  generalization,  after  avoidance  training  using  a  four 
per  second  flicker  tracer  conditioned  stimulus.   (Bottom)  Electrical  responses  to 
ten  per  second  flicker  following  differentiation  of  avoidance  response.    (4  per 
second — S^,    10  per  second — S'^).    (From  Weiss,   Marc:   Unpublished  master's 
thesis,  University  of  Rochester,  1962.) 

Figure  7  illustrates  an  average  response  waveform  obtained  from 
the  lateral  geniculate  body  of  this  animal  during  generalized  per- 
formance of  the  conditioned  response  to  the  ten  per  second  flicker. 
This  computation  was  obtained  using  a  Mnemotron  average  re- 
sponse computer  and  is  based  on  100  periods  of  ten  per  second 


Neural  Mechanisms  oj  Decision  Making  255 

AVERAGE    RESPONSE    OF  LATERAL  GENICULATE    DURING 
GENERALIZATION    TO    lOcps    AFTER  TRAINING    TO    4cps 

100  SWEEPS 


lOcps  on 
Fig.  7.  Average  response  computed  from  lateral  geniculate  during  generalization 
to  ten  per  second  flicker,  after  avoidance  training  using  a  four  per  second  flicker 
tracer  conditioned  stimulus.  (From  Weiss,  Marc:  Unpublished  data.) 


flicker,  each  period  beginning  at  a  deleted  flash  and  lasting  for 
625  milliseconds.  Note  the  regularity  of  the  computed  waveform. 
Similar  regularities  were  observed  in  average  responses  computed 
during  generalization  to  ten  per  second  flicker  from  dorsal  hippo- 
campus, centre  median,  nucleus  reticularis,  and  medial  supra- 
sylvian  cortex.  * 

Figure  8A  shows  the  average  response  waveform  computed  from 
the  visual  cortex  at  this  stage  of  training  during  correct  performance 
to  a  four  per  second  flicker. 

Figure  8B  shows  a  comparable  average  response  waveform  com- 
puted from  potentials  recorded  from  the  visual  cortex  during 
generalized  performance  to  a  ten  per  second  flicker.  Note  the 
complex,  irregular  waveform. 

Figure  8C  shows  the  average  response  waveform  computed  from 
the  visual  cortex  during  correct  performance  to  the  ten  per  second 
flicker  after  diff'erentiation.  In  contrast  to  Figure  SB,  note  the 
markedly  increased  simplicity  and  regularity  of  the  waveform. 

I  was  impressed  by  the  fact  that  these  data  might  provide  the 
basis  to  test  the  hypothesis  that  the  waveforms  observed  during 
the  generalization  represented  the  interaction  between  an  endogen- 
ously  generated  representation,  or  memory,  of  the  stimulus  fre- 


*Histological  verification  of  electrode  placements  has  not  yet  been  obtained. 


256  Information  Storage  and  Neural  Control 

AVERAGE      RESPONSE     OF      VISUAL     CORTEX 


4cps  AFTER    AVOIDANCE 
TRAINING 


DARK     PERIOD 


B. 

DURING    GENERALIZATION 
lOcps     AFTER    TRAINING    TO 
4cps 


lOcps    ON 


lOOms 


1/ 


DARK     PERIOD 


c. 

lOcps    AFTER    DIFFERENTIATION 


Ocps  ON 

V 


,     lOOms 

I — 1 

DARK    PERIOD 


100  SWEEPS 


iOO   SWEEPS 


100   SWEEPS 


CALCULATION    OF  B.  FROM 
C+A     AND  C-A 

26  ms  ^   ., 

GENERALIZATION    WAVEFORM 

o- — oCALCULATED    WAVEFORM 

0=10  +  4      •=10-4 

Fig.  8.  Average  response  computed  from  visual  cortex:  (A)  In  response  to  four 
per  second  flicker  after  avoidance  training  using  a  four  per  second  flicker  tracer 
conditioned  stimulus.  (B)During  generalization  to  ten  per  second  flicker.  (C) 
In  response  to  ten  per  second  flicker  after  difTerentiation  training.  (D)  Com- 
parison of  generalization  waveform  with  calculated  interference  pattern.  (A,  B,  C 
from  Marc  Weiss:  Unpublished  data.) 


quency  used  during  training"  and  the  exogenously  derived  neural 
response  to  the  new  stimulus  eliciting  generalization.  Therefore,  I 
explored  the  interference  patterns  which  could  be  constructed  by 
algebraic  addition  or  subtraction  of  the  waveforms  (Figs.  8A  and 


Neural  Adechanisms  of  Decision  Making 


257 


8C)  obtained  from  the  visual  cortex  during  behaviorally  appropri- 
ate response  to  ten  per  second  and  four  per  second  flicker. 

Figure  8D  shows  the  approximation  to  the  generalization  wave- 
form which  can  be  produced  by  these  simple  algebraic  manipula- 
tions. At  each  point  of  the  curve,  the  manipulation  which  gave  the 
better  approximation  (10+4  or  10—  4)  was  selected.  It  is  not  clear 
what  the  physiological  basis  might  be  for  the  particular  sequence 
of  algebraic  operations  used  to  achieve  this  approximation. 


AVERAGE      RESPONSE 


4cps  AFTER    AVOIDANCE 
TRAINING 


RETICULAR       FORMATION 


100  SWEEPS 


250ms 


DARK    PERIOD 


100   SWEEPS 


B. 
DURING    GENERALIZATION 
lOcps     AFTER     TRAINING 
TO    4cps 


C. 

lOcps    AFTER     DIFFERENTIATION 


lOOms 
DARK    PERIOD 


CALCULATION    OF     B     FROM 
C  +  A    AND    C-A 


26  ms 


- —  GENERALIZATION    WAVEFORM 
°—o  CALCULATED    WAVEFORM 


0=10  +  4  •=10-4 

Fig.  9.  As  Figure  8,  but  data  derived  from  mesencephalic  reticular  formation. 


258  Information  Storage  and  Neural  Control 

Figure  9A  shows  the  average  response  waveform  obtained  from 
the  mesencephaHc  reticular  formation  at  this  stage  of  training 
during  correct  performance  to  a  four  per  second  flicker. 

Figure  9B  sliows  the  average  response  waveform  obtained  from  the 
mesencephalic  reticular  formation  during  generalization  to  the  ten 
per  second  flicker.  Note  the  highly  complex  and  irregular  waveform. 

Figure  9C  shows  the  average  response  waveform  obtained  from 
the  mesencephalic  reticular  formation  during  correct  performance 
to  the  ten  per  second  flicker  following  differentiation.  In  contrast 
to  Figure  9B,  note  the  increased  simplicity  and  regularity  of  the 
waveform. 

Figure  9D  shows  the  fit  to  the  generalization  waveform  of  the 
interference  pattern  which  can  be  obtained  by  arbitrary  algebraic 
addition  or  subtraction  of  the  two  waveforms  elicited  from  the 
reticular  formation  during  behaviorally  appropriate  performance 
to  four  per  second  and  ten  per  second  flicker,  as  shown  in  Figures 
9A  and  9C.  Again,  that  manipulation  (10  +4  or  10  —  4)  which 
gave  the  better  fit  was  selected. 

Thus,  one  can  synthesize  interference  patterns  from  average 
response  waveforms  computed  during  behaviorally  appropriate 
response  to  two  different  stimuli  and  can  approximate  closely  the 
actual  average  response  waveform  obtained  when  an  animal  re- 
sponds to  one  of  these  stimuli  by  a  previously  learned  behavior 
appropriate  to  the  other.  This  demonstration  provides  striking 
evidence  in  support  of  the  suggestion  that  the  neural  response  to 
the  ten  per  second  stimulus  actually  presented  was  modified  during 
generalization  by  an  electrical  influence  identical  with  the  response 
to  the  four  per  second  conditioned  stimulus  repeatedly  experienced 
during  the  earlier  establishment  of  the  conditioned  response.  At 
the  moment  I  see  no  way  to  evade  the  conclusion  that  the  conse- 
quence of  experience  with  the  four  per  second  flicker  during  learn- 
ing somehow  caused  a  modification  of  neural  structure  which  there- 
by gained  the  capacity  to  generate  electrical  activity  like  that  which 
established  it.  These  data  support  the  interpretation  that  such 
patterns  of  potentials  are  of  functional  significance  and  are  closely 
related  to  the  actual  processing  of  information. 

At  our  present  stage  of  knowledge,  no  mechanisms  come  to  mind 
which  might  serve  to  generate  and  mediate  an  interaction  of  the 


Neural  Mechanisms  of  Decision  Making  259 

sort  described.  Yet  some  insight  may  be  offered  from  the  fact  that 
only  visual  cortex  and  reticular  formation,  among  the  structures 
studied  in  this  animal,  displayed  these  peculiar  waveforms  during 
generalization.  Lateral  geniculate  was  notably  regular  in  its 
response.  This  configuration  suggests  that  somehow  an  interaction 
between  visual  cortex  and  reticular  formation  may  be  central  in 
the  mediation  of  phenomena  of  this  sort.  Further  work  is  obviously 
necessary  before  the  interpretations  offered  here  can  be  accepted 
as  accurate. 

CHARACTERISTICS  OF  MISTAKES 
DURING  DIFFERENTIATION 

Although  the  data  presented  in  the  preceding  section  are  of  a 
different  sort  from  those  which  Killam  and  I  described  previously, 
they  are  in  accordance  with  observations  we  made  while  studying 
the  difference  in  electrical  recordings  obtained  during  correct  and 
erroneous  performance  of  flicker  discriminations  in  a  differential 
approach-avoidance  situation  (9).  In  those  animals,  we  observed 
that  among  the  most  marked  changes  in  labeled  potentials  during 
differential  conditioning  were  those  which  occurred  in  the  reticular 
formation,  intralaminar  nuclei,  and  hippocampus.  A  particular 
relationship  between  the  configuration  of  potentials  in  these  struc- 
tures and  in  visual  cortex  seemed  to  be  closely  related  to  appropriate 
performance.  During  signal  presentation,  potentials  in  the  non- 
specific structures  could  often  be  observed  at  either  of  the  two 
flicker  frequencies  between  which  differential  response  had  been 
established.  Wlien  behavioral  performance  was  appropriate  to  the 
peripheral  OS,  the  frequencies  of  potentials  in  visual  cortex  and 
in  nonspecific  regions  were  in  good  correspondence  to  the  OS. 
However,  when  behavioral  performance  was  inappropriate,  the  cor- 
respondence of  labeled  potentials  to  tlie  CIS  diininished  and  periods 
of  hypersynchrony  appeared  at  the  frequency  of  the  stimulus 
appropriate  to  the  behavior  actually  performed,  particularly 
in  centralis  lateralis,  dorsal  hippocampus  and  reticular  forma- 
tion. In  Figure  10  are  presented  recordings  obtained  from  a  cat 
trained  to  perform  a  lever  press  to  obtain  milk  during  a  ten  per 
second  flicker  without  reinforceinent  during  six  per  second  flicker. 


260  Information  Storage  and  Neural  Control 

After  Operant  Conditioning  to  IO/5  "$<=',    6/s  -5^ 

LG  AA«'^,VWvVw''^*^i/^^  1 

Fx  ^''.A^^vv^^\AAV^\^4^f*^(^^M  i 

10*  ; , 

SIG  W/WVVWVVWWVWWWWVWVWVVW/WWWWWVWV'JVW^ 

VH  ^>/\^ht^\f^^/^j>f^^  I 

Correct 

FX  \\\hY¥4^'^'*^^^^ 

s  iG  /vwwwv  ;;;;M\/ww\/'//w'^vw/w\/'yw^AWM/vv;MWAVvvwwvMVWvwv^^ 

Error 

Fig.  10. 

MG — Medial  geniculate 

VC— Visual  cortex 

LG — Lateral  geniculate 

FX — Fornix 

VH — Ventral  hippocampus 

CL — Centralis  lateralis 

MSS — Medial  suprasylvian  cortex 

Records  obtained  during  differential  approach  conditioning  (10  per  second — S    , 

6  per  second — S^).  (Top)  Correct  response  to  ten  per  second  flicker.   (Bottom) 

Error  of  omission  to  ten  per  second  flicker.  (From  John,  E.  R.  and  Killam,K.  F.: 

J.  Nerv.  Merit.  Dis.,  73/.-183-201,  1960.) 

The  top  records  were  taken  during  correct  performance  to  ten 
per  second  flicker.  Note  in  particular  the  marked  frequency- 
specific  response  in  fornix  and  centraHs  lateralis.  The  bottom 
records  were  obtained  during  an  error  of  omission  when  the  cat 
failed  to  press  the  lever  in  response  to  the  ten  per  second  signal.  Note 
the  diminished  ten  per  second  labeled  potentials  and,  in  particular, 


Neural  Mechanisms  of  Decision  Making  261 

After  Operant  Conditioning  to  ICVs-S*^,  6/i-S^ 

s'G  — ^ s;f-o^.^^^vA\\^^^^^\J\J^^^\^^J\,\^J\^ 

Correct 

siQ — ^v:vVvV^.\\Vv\\\^,■"^^^^^^^^J\\ 

tnror 

Fig.  11. 

MG — Medial  geniculate 

VC — Visual  cortex 

LG — Lateral  geniculate 

FX — Fornix 

VH — Ventral  hippocampus 

CL — Centralis  lateralis 

MSS — Medial  suprasylvian  cortex 

Records  obtained  during  differential  conditioning  (10  per  second — S^,  6  per 

second — S   ).  (Top)  Correct  response  to  six  per  second  flicker.  (Bottom)  Error  of 

commission  to  six  per  second  flicker.  (From  John,  E.  R.  and  Killam,  K.  F.: 

J.  Nerv.  Merit.  Dis.,  737.- 183-201,  1960.) 

the  slow  potential  at  about  six  per  second  seen  most  clearly  in 
centralis  lateralis. 

Figure  11  shows  the  converse  phenomenon  in  the  same  cat.  The 
top  record  shows  correct  performance  to  the  non-reinforced  six 
per  second  flicker.  The  bottom  record  shows  an  error  of  commission 
to  the  six  per  second  flicker.  Note  the  lessened  frequency  specificity 


262  Information  Storage  and  Neural  Control 

After  CAR  to  6/S 

Fig.  12.  Records  obtained  during  lever  press  to  10  per  second  flicker  after  avoid- 
ance training  to  the  6  per  second  flicker.  Arrow  indicates  conditioned  response. 
(From  John,  E.  R.  and  Killam,  K.  F.:  J.  Mrv.  Merit.  Dis.,  7J7.-183-201,  1960.) 

of  potentials  in  the  lower  record  as  contrasted  with  the  upper;  in 
particular,  observe  the  period  of  approximately  ten  per  second  po- 
tentials in  centralis  lateralis. 

Figure  12  shows  the  potential  configuration  reliably  obtained  in 
this  cat  in  response  to  ten  per  second  flicker  following  the  establish- 
ment of  a  conditioned  avoidance  response  to  the  six  per  second 
flicker,  while  the  conditioned  lever  pressing"  response  to  ten  per 
second  flicker  was  maintained.  At  this  stage  in  this  animal,  presenta- 
tion of  the  ten  per  second  flicker  elicited  clear  labeled  potentials 
in  visual  cortex  and  several  other  structures,  while  an  initial  slow 
wave  at  about  six  per  second  appeared  in  centralis  lateralis  and 
fornix.  Superimposed  on  this  slow  potential,  almost  as  a  modulation, 
is  a  ten  per  second  potential  which  gradually  becomes  clearer  and 
eventually  dominates  the  record.  Wlien  lever  press  occurred  to 
the  ten  per  second  flicker,  it  almost  invariably  took  place  during 
a  period  when  the  ten  per  second  labeled  potential  dominated  the 
activity  of  centralis  lateralis.  Characteristically,  as  this  correspond- 
ence between  the  frequency  of  the  dominant  activity  in  the  non- 
specific structures  and   in   the  visual  cortex  occurred,   a  change 


Neural  Mechanisms  of  Decision  Makirrg  263 

During  Blockc3dc  of  CAR  After  Rcserpinc 

vc  -*v^A^AM^^MMM'V^^A^^MAA/W\A^  -^o^/v  i 

LG      ^H^aT'''^^^  I 

Fig.  13.  Records  obtained  in  response  to  ten  per  second  flicker  after  performance 
of  the  avoidance  response  to  six  per  second  flicker  was  blocked  by  injection  of 
reserpine  (1007/kg).  (From  John,  E.  R.  and  Killam,  K.  F.:  J.  Nerv.  Ment.  Dis., 

737.- 183-201,  1960.) 

was  observed  in  the  recorded  waveforms.  This  change  was  a  shift 
from  rounded  "waves"  to  more  sharply  peaked  spikes  and  was 
foUowed  one  or  two  seconds  later  by  performance  of  the  conditioned 
response. 

Some  indication  of  the  possible  functional  relevance  of  the  slow 
six  per  second  centralis  lateralis  waves  seen  during  the  approach 
signal  after  avoidance  training  is  provided  by  the  data  in  Figure  13. 
When  performance  of  the  avoidance  response  to  the  six  per  second 
TCS  was  completely  blocked  after  administration  of  100  7/kg.  of 
reserpine,  presentation  of  the  ten  per  second  TCS  no  longer  elicited 
the  previously  marked  slow  potentials  in  centralis  lateralis  and 
elsewhere,  but  instead  resulted  in  the  appearance  of  massive 
labeled  responses  at  ten  per  second  frequency.  Wlien  0.5mg/kg.  of 
amphetamine  was  administered  to  this  cat,  the  reserpine  blockade 
of  the  conditioned  avoidance  response  performance  to  the  six 
per  second  TCS  was  completely  reversed  in  a  few  minutes.  Presenta- 
tion of  the  ten  per  second  TCS  for  the  lever  pressing  response  to 


264  Information  Storage  and  Neural  Control 

obtain  milk  once  again  elicited  the  same  slow  potentials  in  centralis 
lateralis  and  elsewhere  as  seen  previously  in  Figure  12. 

These  observations  seemed  to  support  the  interpretation  that 
the  labeled  potentials  reflected  some  aspect  of  information  process- 
ing and  might  be  of  functional  significance.  Such  an  interpretation 
would  also  be  in  agreeinent  with  the  findings  of  Livanov  et  al.  (13) 
and  Liberson  et  al.  (12)  who  have  reported  that  direct  electrical 
stimulation  of  various  brain  structures  at  frequencies  like  those  of 
the  intermittent  conditioned  stimuli  used  in  establishing  a  condi- 
tioned response  resulted  in  performance  of  the  learned  behavior. 

Nonspecific  structures  seem  to  play  a  central  role  in  the  processing 
of  information  during  differentiation.  Evidence  of  differential 
suppression  of  potentials  after  habituation,  of  the  major  signs  of 
assimilation,  of  the  inost  marked  increinents  in  labeled  potentials 
during  differential  training,  and  of  shifts  in  the  frequency  of 
labeled  potentials  during  behaviorally  inappropriate  response  have 
all  been  observed  in  these  structures.  The  particular  configuration 
of  potential  patterns  during  differential  response  suggested  several 
hypotheses:  1)  The  role  of  specific  sensory  systems  may  be  con- 
ceived of  as  the  central  propagation  of  information  representing 
the  present  state  of  the  environment  to  a  particular  cortical  region; 
2)  this  information  may  be  compared,  via  the  diffuse  projection 
system,  with  a  representation  of  past  experiences  activated  in  the 
rhinencephalon  and  the  reticular  formation  by  the  similarity  be- 
tween past  and  present  environment,  modified  by  the  state  of 
the  organism  in  terms  of  effect  and  drive  level;  and,  3)  the 
appropriate  selective  performance  of  adaptive  behavioral  responses 
may  depend  upon  achievement  of  a  sufficient  congruence,  via  some 
unknown  coincidence  detection  mechanism,  of  the  potentials 
reflecting  present  and  past  experience. 

CONCURRENT  PERIPHERAL  AND 
CENTRAL  STIMULATION 

These  various  considerations  led  our  group  to  investigate  further 
the  question  of  whether  temporal  patterns  of  potentials  might  be 
information.  When  animals  are  trained  to  perform  a  differential 
discrimination  between  two  flicker  stimuli  differing  in  frequency, 


Neural  Mechanisms  of  Decision  Making  265 

are  the  observed  frequency-specific  potentials  a  reflection  of  the 
coding"  and  processing  of  information  causafly  related  to  the 
behavioral  performance,  or  do  they  merely  reflect  generalized 
processes  of  local  excitation  and  inhibition  that  are  not  specifically 
informational  and  bear  only  a  relationship  of  concomitance  to 
the  behavioral  performance? 

In  the  initial  studies  which  we  undertook  to  resolve  these  ques- 
tions (10),  an  attempt  was  made  to  evaluate  directly  the  functional 
significance  of  labeled  potentials  observed  in  various  brain  struc- 
tures in  cats  fully  trained  to  perform  diff'erential  avoidance  re- 
sponses to  two  flicker  conditioned  stimuli  of  different  frequencies. 
We  studied  the  behavioral  effects  of  direct  electrical  stimulation 
of  the  brain  at  frequencies  concordant  or  discordant  with  the 
frequency  of  the  peripheral  conditioned  stimuli  presented  simul- 
taneously. After  pilot  studies  showed  that  low  frequency  electrical 
stimulation  was  not  effective,  a  modulation  technique  was  devised. 
A  standard  "carrier"  waveform,  consisting  of  a  100  cycle  per 
second  biphasic  square  wave  with  a  2  millisecond  pulse  duration, 
was  modulated  at  the  frequency  of  the  peripheral  TCS.  This  pro- 
duced trains  of  bursts  of  100  cycle  per  second  square  waves,  with 
the  burst  frequency  identical  with  the  flicker  frequencies  to  which 
the  animals  were  conditioned.  Trains  at  different  frequencies  could 
be  manipulated  to  achieve  equal  duration  of  constituent  bursts 
or  to  equate  total  electrical  energy  by  selection  of  appropriate 
burst  durations. 

Most  structures  were  explored  both  unilaterally  and  bilaterally. 
For  each  structure,  we  determined  the  current  level  at  which 
central  stimulation  at  both  the  reinforced  (S  )  and  the  non- 
reinforced  (S^)  frequency  blocked  performance  to  concurrent 
photic  stimulation  at  the  S  frequency.  This  current  level  was 
defined  as  the  occlusion  threshold,  or  cut-off.  The  current  intensity 
at  which  conditioned  response  perfoimance  returned  to  concurrent 
photic  and  central  stimulation  at  one  central  frequency  but  not 
the  other  was  defined  as  the  differential  threshold.  If  a  differential 
threshold  was  observed,  a  series  of  trials  was  carried  out  to  de- 
termine the  reliability  of  such  an  effect.  Throughout  such  stimula- 
tion sessions,  central  stimuli  were  presented  in  counterbalanced 
frequency  sequence,  and  each  sequence  was  bracketed  by  trials 


266  Information  Storage  and  Neural  Control 

using  only  the  peripheral  conditioned  stimuli.  Only  central  se- 
quences bracketed  by  correct  performance  to  the  peripheral 
stimulus  alone  were  considered  acceptable. 

Intensive  studies  of  the  effects  of  concurrent  central  and  peri- 
pheral stimulation  have  been  carried  out  in  two  cats.  One  of  these 
animals  (Cat  4)  was  conditioned  to  press  a  lever  to  avoid  shock 
within  fifteen  seconds  after  the  onset  of  a  four  per  second  flicker, 
but  was  punished  if  lever  press  was  performed  during"  a  ten  per 
second  flicker.  The  other  animal  (Cat  10)  was  trained  to  the 
opposite  significance  of  flicker  frequency,  pressing"  the  lever  to 
ten  per  second  flicker  but  not  to  four  per  second.  Results  of  the 
concurrent  stimulation  studies  on  these  two  animals  are  sum- 
marized in  Table  I. 

Note  that  the  data  show,  at  a  very  high  significance  level,  that  a 
four  per  second  electrical  stimulation  of  the  visual  cortex  is  much 
more  effective  than  a  ten  per  second  input  in  achieving  inhibition  of 
conditioned  avoidance  response  performance  to  a  simultaneously 
presented  TCS  in  both  Cat  4  and  Cat  10,  although  the  meaning 
of  a  four  per  second  flicker  was  opposite  for  these  two  animals. 
Since  this  was  true  both  for  central  stimuli  of  equal  burst  duration 
and  for  those  of  equal  energy,  the  severe  disruption  can  be  at- 
tributed to  the  frequency  of  the  simulated  input.  Four  per  second 
central  stimulation  was  much  more  inhibitory  than  ten  per  second. 
This  effect  was  not  observed  in  auditory  or  medial  suprasylvian 
cortex,  but  appeared  to  be  rather  specific  for  the  cortex  of  the 
CS  modality.  This  suggests  that  the  input  in  some  way  interferes 
with  activity  in  the  visual  system  and  that  the  visual  cortex  or 
regions  to  which  it  projects  are  involved  in  the  mediation  of  the 
conditioned  response.  Such  conclusions  would  be  consonant  with 
those  of  Zuckermann  (27),  who  observed  interference  with  per- 
formance of  conditioned  responses  to  visual  stimuli  during  after- 
discharge  following  stimulation  of  visual  cortex  but  not  of  motor 
cortex  or  reticular  formation.  Such  a  conclusion  is  difficult  to 
reconcile  with  the  remarkable  ability  of  Cat  4  to  sustain  appropriate 
behavioral  response  to  a  four  per  second  flicker  when  2.5  times 
more  electrical  energy  was  applied  to  the  same  visual  cortex  at 
ten  per  second.  In  contrast  to  the  cortical  current  values  for  dis- 
ruption, note  the  exceedingly  low  current  required  in  subcortical 


Neural  Mechanisms  of  Decismi  Making 


267 


;^ 


o  o  o 
VVV 


o  o  o  o  o 


A  A  A  A  A 
-^  -*  ^  -^  •* 


o  o 
VV 


V      V 


AAII  AAAAAAAAAII  All  A 

■*0        TfOO-^OOOOO        o        o 


i  O  (M  O 


I      I      I      I      I      I      I      I      I      I      I      . 
Tj-oO-^O<NOLnOr0cq\0n 

to         •         •         ■   CN   T-.         ■   r-c 

'-I  o  o       o 


OOO'^OOOO 

„  „  i    I    I    I    I    I    I    I 

00C^h00CN1i-O(N1O\000 


fe    -54, 

(.1    ^ 

+  ^ 

=o 

O 

+ 

<^l  OI  ^  Ol  ^  T-c        ^_  <N]  ro  .^  tv]  fn  so        to 


+ 


^-c  ■rt  CMt-it-c-t-.  (NJcOt-c  T-.OJcOr-i'^i-.Tf  ( 


'*0'*OtJ-0-*0-^0^0'^0-^0-^0^0-D-' 


3 

„ 

o 

^ 

o 

o 

o 

(« 

4J 

C/3 

■£-G 

■a 

3 

.  ^ 

i;   G 

1^ 

lue  repr 
th  powe 
d  found 

t  va 
t  wi 
shol 

'c5 

.3^ 

ren 
can 
hre 

"o 

^Oi 

3S- 

Crrl 

^.§^•2 

h 

'So 

u  3  " 

ti 

;;  s 

^  cr  o 

o 

n  o 

hMZ22£ 

*       1        \    \-XK 

fe=^ 

268  Information  Storage  and  Neural  Control 

stimulation,  particularly  in  the  reticular  formation,  to  achieve 
similar  results. 

These  experiments  showed  that  low  frequency  stimulation  seemed 
to  have  an  intrinsically  inhibitory  effect.  Perhaps  the  fact  that 
differential  inhibition  occurred  in  the  same  direction  in  both  of 
these  animals  may  be  understood  on  this  basis.  However,  this 
study  did  not  provide  evidence  in  support  of  the  hypothesis  that 
the  configuration  of  labeled  potentials  during  performance  of 
a  difTerentiated  conditioned  response  to  a  TCS  represents  coded 
information  about  the  peripheral  stimulus  being  processed  by  the 
nervous  system.  Work  must  be  pursued  with  more  animals,  using 
other  frequencies  and  additional  anatomical  placements,  in  order 
to  clarify  these  questions. 

Yet  one  can  conceive  of  a  number  of  possible  reasons  for  the 
results  obtained:  1)  The  brain  does  not  code  or  process  information 
in  a  manner  which  is  related  to  the  observed  configurations  of 
labeled  potentials;  2)  powerful  intrinsic  "resonance''  to  low  fre- 
quency input  resulted  in  an  uncoded  inhibitory  effect  masking  our 
ability  to  find  differential  effects  based  on  the  significance  of  a 
particular  frequency  for  an  animal;  or  3)  the  organization  of  the 
coding  and  processing  of  this  sort  of  information  in  differential 
response  might  proceed  exactly  as  we  would  conjecture  on  the 
basis  of  configurations  of  labeled  potentials.  Our  inability  to 
demonstrate  differential  effects  based  on  central  stimulation  at 
presumably  informational  frequencies  might  simply  be  due  to 
the  fact  that  the  response  of  neural  tissue  to  our  artificial  waveforms 
was  inappropriate  for  functional  interaction  with  "brain  language." 

DIFFERENTIAL  CONDITIONING  TO  STIMULATION 
OF  A  CENTRAL  SITE 

A  number  of  experimental  strategies  have  been  devised  to  explore 
these  alternative  explanations.  One  relatively  straightforward 
approach  was  to  attempt  to  make  our  artificial  input  functionally 
equivalent  with  brain  language.  We  were  struck  by  the  fact  that  in 
hundreds  of  trials  we  had  failed  to  get  any  indication  of  behavioral 
response  to  direct  central  stimulation  alone.  This  seemed  to  contra- 
dict the  work  of  Livanov  et  al.  (13),  Liberson  et  al.  (12),  and  Neff  ^/ 


Neural  Mechanisms  of  Decision  Making 


269 


al.  (22,  23),  who  reported  response  to  central  stimulation  after  con- 
ditioning to  a  peripheral  stimulus.  We  came  to  the  conclusion  that 
this  difference  might  be  due  to  the  fact  that  our  animals  had  been 
highly  overtrained  to  differential  response,  with  punishment  for 
error.  Subsequent  pilot  work  by  Karl  Corley  in  our  laboratories  has 
confirmed  that  central  stimulation  will  elicit  responses  previously 
established  to  peripheral  stimuli  when  erroneous  performance  has 
not  been  punished.  Since  our  central  stimuli  had  never  been  coupled 
with  primary  reinforcement,  we  attempted  to  train  these  animals 
to  differentiate  between  the  pulse  trains  which  had  been  used  in 
the  previous  work.  Frequency  significance  for  each  animal  remained 
the  same  as  for  the  peripheral  flicker.  Thus,  one  can  consider  this 
to  be  an  attempt  to  transfer  the  differential  response  from  inter- 
mittent photic  stimulation  to  intermittent  central  stimulation. 

Figure  14  shows  the  learning"  curve  for  Cat  4.  Differential  training 
was  instituted  at  the  arrow  after  reliable  performance  of  the  avoid- 
ance response  had  been  established  to  four  per  second  bursts 
delivered  to  the  electrodes  on  left  and  right  visual  cortex.  Trials 


CAT    4 
CENTRAL    CONDITIONING    USING    BILATERAL   STIMULATION 
OF   VISUAL   CORTEX 


25    I    58    I 
50       83 


8     9     10 
SESSIONS 

CUMULATIVE    TRIALS 

-T— 
101 


I     121    I    171     I 
112        146       183 


16 


— I— 
30 


— r- 

40 


— r- 
63 


8  13  17        25 

Fig.  14.  Learning  curve  for  avoidance  response  to  direct  central  stimulation  in 
cat  previously  differentially  trained  to  flicker  conditioned  stimulus.  At  arrow 
central  diff"erentiation  training  began  (4  per  second — S    ,  10  per  second — S'^). 


270 


Information  Storage  and  Neural  Control 


resulting  in  failure  to  perform  in  response  to  the  S^  were  scored 
as  correct  only  when  bracketed  by  correct  performance  of  the 
conditioned  response  to  the  S  .  Four  per  second  bursts  were 
usually  25  milliseconds  in  duration,  although  response  could  be 
elicited  by  shorter  bursts.  Ten  per  second  bursts  were  usually  10 
milliseconds  in  duration.  Central  pulse  trains  at  different  modula- 
tion frequencies  were  equated  for  total  electrical  energy.  "Carrier" 
frequency  was  usually  200  cycles  per  second,  but  response  could  also 
be  elicited  reliably  at  100  cycles  per  second.  Carrier  pulse  width 
was  2  milliseconds.  Threshold  current  for  performance  was  found 
to  be  between  1.8  and  2.1  milliamperes.  Note  that  the  occlusion 
threshold  at  the  stimulation  site  for  concurrent  peripheral  and 
central  stimulation  had  been  4.0  milliamperes. 
^^Figuie  15  shows  the  learning  curve  for  Cat  10.  Stimulus  par- 
ameters were  as  for  Cat  4,  but  the  opposite  significance  was  at- 
tached to  frequency.  Threshold  current  was  around  1.8  milli- 
amperes. 


CAT   10 
CENTRAL    CONDITIONING    USING    BILATERAL     STIMULATION 
OF     VISUAL    CORTEX 


8      9     10     II 
SESSIONS 

CUMULATIVE    TRIALS 

I 1 1 1 1 1 : 1 1 1 1 — 

12     '     59    '    96    '     117    I     167  '    203 
37         71  107        142        183 


26 


48 


64 


— I — 
84 


15 


21 


~~i — 
27 


90 


35 


Fig.  15.  As  Plgure  14,  but  in  tliis  animal  the  significance  of  the  stimulus  fre- 
quencies was  reversed.  (10  per  second — S  ,  4  per  second — S^).  Note  apparent 
transfer  of  previously  established  peripheral  frequency  discrimination  to  central 

stimuli. 


Neural  Mechanisms  of  Decision  Making  271 

It  is  of  interest  that  in  this  animal,  once  the  central  stimulus  had 
been  established  as  informationally  adequate  by  conditioning,  the 
peripherally  established  c/ifferrntiated  response  appeared  to  generalize 
to  the  central  stimuli.  Differential  response  to  central  ten  per 
second  and  four  per  second  stimuli  was  perfect  on  the  twelfth 
session,  which  was  the  first  occasion  on  which  the  four  per  second 
stimulus  was  presented  following  conditioning  to  central  ten  per 
second  stimulus.  The  animal  seemed  to  benefit  from  the  previous 
differential  experience  with  peripheral  stimuli  at  these  frequencies. 
Whether  or  not  such  generalization  will  take  place  reliably,  it  is 
obvious  that  these  two  animals  have  been  trained  to  diff"erentiate 
between  two  sequences  of  events  of  identical  energy  occuring 
at  the  same  central  site.  Thus,  the  temporal  pattern  of  events  at  a 
place  in  the  brain  can  serve  as  information.  Further,  since  the  con- 
stituent pulses  of  all  central  stimuli  are  but  two  milliseconds 
wide,  the  temporal  pattern  of  significance  here  is  the  slow  modula- 
tion frequency  characterizing  the  stimulus  trains.  Such  evidence 
does  not  demonstrate  that  the  patterns  of  slow  labeled  potentials 
which  appear  in  certain  brain  structures  are  information,  but 
it  does  establish  that  slow  patterns  at  a  place  can  be  information. 

Clearly,  it  is  desirable  to  explore  the  propagation  of  such  centrally 
delivered  difTerential  stimuli  to  other  brain  regions  from  the  input 
site  before  and  after  they  are  established  as  adequate  conditioned 
stimuli.  The  interaction  of  such  pulsed  central  stimuli  with  con- 
current photic  stimuli  must  be  investigated,  and  the  behavioral 
as  well  as  electrophysiological  consequences  of  concordant  and 
discordant  central  and  peripheral  tracer  stimuli  are  presently  being 
studied  in  our  laboratories.  Such  studies  should  provide  additional 
insight  into  the  functional  role  of  labeled  potentials  in  performance. 

Some  additional  information  of  interest  has  been  obtained  from 
Cat  10.  Table  II  shows  the  consequences  of  a  number  of  trials  in 
which  one  or  the  other  of  the  visual  cortex  electrodes  was  stimu- 
lated together  with  some  other  cortex  placement.  Note  that 
differentiated  conditioned  response  was  obtained  fairly  consistently 
when  the  stimulated  electrode  pair  included  the  right  visual  cortex 
electrode  but  not  when  it  included  the  left  visual  cortex  electrode. 
It  is  pertinent  to  recall  that  the  central  pulses  were  biphasic.  These 
data  suggest  that  whatever  the  nature  of  the  neural  mechanism 


272  Information  Storage  and  Neural  Control 

TABLE  II 

Generalization  of  Differentiated  Avoidance  Response  to  Central 

Stimulation  of  Other  Electrode  Placements  Following  Establishment  of 

Differential  Response  to  Electrical  Stimulation  of 

Left  vs.  Right  Visual  Cortex 


10/Sec 

4/Sec. 

CR 

NR 

CR           NR 

Right  Visual  + 
other  cortical  sites 

10 

7 

0             11 

Left  Visual  + 
other  cortical  sites 

1 

18 

2               4 

Sub-Cortical  Sites 

0 

15 

mediating  the  differential  performance  to  the  central  stimulation, 
this  relationship  has  been  preferentially  established  to  only  one 
of  the  two  stimulated  regions.  This  finding  is  consonant  with 
the  analogous  report  of  Loucks  (15),  but  extends  his  observation 
to  differentiated  responses. 

It  is  also  of  interest  that  bipolar  subcortical  stimulation,  without 
reinforcement,  did  not  elicit  generalization  of  performance.  In 
subsequent  training,  we  obtained  some  indication  of  performance 
of  conditioned  responses  to  stimulation  of  the  reticular  formation  in 
this  animal.  A  total  of  175  conditioned  responses  were  obtained  in 
460  trials.  Performance  fluctuated  between  70  per  cent  and  0  per 
cent  and  did  not  stabilize.  This  may  indicate  changes  in  neural 
threshold  since  all  trials  were  conducted  in  the  same  current  range, 
or  may  reflect  artifact  due  to  an  increase  of  operant  level  which  took 
place.  Systematic  studies  of  generalization  and  transfer  of  differ- 
ential conditioned  response  from  one  site  of  central  stimulation  to 
another,  in  conjunction  with  electrophysiological  studies,  may  help 
elucidate  the  neuial  mechanisms  mediating  such  perfoimance. 

DIFFERENTIAL  EFFECTS  OF  LATENCY  ON 
DISRUPTION  DURING  CONCURRENT  STIMULATION 

Some  months  after  the  preceding  experiments  were  concluded,  a 
final  study  was  carried  out  on  Cat  4.  Under  extinction  conditions,  it 
was  observed  that  stimulation  of  visual  cortex  at  previously  effective 
parameters  no  longer  elicited  the  conditioned  response.  No  attempt 
was  made  to  ascertain  whether  the  introduction  of  reinforcement 


Neural  Mechanisms  of  Decision  Making  273 

would  restore  performance.  Instead,  we  measured  the  occlusion 
threshold  for  concurrent  four  per  second  flicker  and  four  per  second 
visual  cortex  stimulation  and  observed  it  to  be  around  3.0  milli- 
amperes.  An  experiment  was  then  devised  to  explore  whether  the 
disruptive  eff'ect  of  cortical  stimulation  varied  as  a  function  of  the 
time  that  such  stimulation  occurred  with  respect  to  the  instant 
when  the  flash  of  light  was  presented. 

Specifically,  we  investigated  the  consequences  of  using  our 
stimulus  generators  (Tektronix)  to  delay  direct  cortical  stimuli  so 
as  to  cause  them  to  coincide  either  with  the  early  or  late  phase  of 
the  cortical  response  to  the  peripheral  conditioned  flash.  Stimuli 
were  arranged  in  a  counterbalanced  sequence,  thus: 

4  per  second  flicker  alone 

4  per  second  flicker  +  4  per  second  visual  cortex  stimuli  (early) 

4  per  second  flicker  +  4  per  second  visual  cortex  stimuli  (late) 

4  per  second  flicker  alone 

4  per  second  flicker  +  4  per  second  visual  cortex  stimuli  (late) 

4  per  second  flicker  +  4  per  second  visual  cortex  stimuli  (early) 

4  per  second  flicker  alone 

All  cortical  stimuli  were  at  2.8  milliamperes.  Carrier  frequency 
was  100  cycles  per  second  with  two  millisecond  pulse  width.  Bursts 
consisted  of  five  biphasic  pulses  (25  millisecond  burst  width). 
"Early"  stimuli  were  so  phased  as  to  reach  visual  cortex  15  milli- 
seconds after  each  flash  of  the  four  per  second  flicker.  "Late" 
stimuli  were  timed  to  reach  the  cortex  either  80  milliseconds  or 
110  milliseconds  after  each  flash  of  the  four  per  second  flicker. 
Shock  to  the  feet  was  delivered  if  the  avoidance  response  was 
not  elicited  within  fifteen  seconds  when  the  four  per  second  flicker 
was  presented  alone.  All  trials  involving  central  stimulation  were 
under  extinction  conditions,  i.e.,  no  shock  was  delivered.  Central 
sequences  were  scored  only  if  bracked  by  correct  performance 
of  the  conditioned  response  in  less  than  fifteen  seconds  to  four  per 
second  flicker  alone,  without  punishment. 

Table  III  summarizes  the  results  of  these  experiments.  As  can  be 
seen,  central  stimuli  arriving  'iate"  were  very  much  more  dis- 
ruptive than  identical  perturbations  arriving  early.  This  suggests 


274 


Information  Storage  and  Neural  Control 


TABLE  III 

Effects  of  Electrical  Stimulation  of  Visual  Cortex  at 

Various  Delays  After  Presentation  of  Four  per  Second  Flash 

From  Peripheral  Tracer  Conditioned  Stimulus 

{2.8  mA,  100/cps.  Biphasic,  2  mS  Pulse  Width,  25  mS  Duration) 

CR  No  CR 


Delay 


Delay 


15 

mS 

80 

mS 

15 

mS 

10 

mS 

Delay 


15  mS 
80  +  no  mS 


34 

22 

15 

39 

Cl< 

No  CR 

16 
2 

12 
26 

ro  7 

CR 

AL 

No  CR 

50 

17 

34 
65 

X2  =11.7 
p  <  .001 

X2  =  15.1 
p  <  .001 

X2  =25.7 
p  <  .001 


that  the  processing  of  information  about  the  peripheral  conditioned 
signal  is  at  a  more  crucial  stage  in  the  visual  cortex,  or  at  the  site 
to  which  the  central  stimuli  propagate,  during  the  late  phase  of 
the  cortical  evoked  potential  than  during  the  early  phase.  Examina- 
tion of  average  response  computations  from  various  brain  structures 
in  this  animal  suggests  that  the  late  phase  of  the  cortical  average 
response  waveform  varies  in  form  and  latency  with  the  average 
response  seen  in  reticular  formation  and  centralis  lateralis. 


SUMMARY  AND  CONCLUSIONS 

Diverse  kinds  of  evidence  have  been  presented  here  both  to 
illustrate  the  nature  of  research  in  progress  and  to  evaluate  a  body 
of  data.  Although  the  number  of  animals  for  which  each  of  these 
kinds  of  data  has  been  obtained  is  as  yet  small,  the  consistency  and 
the  clear  statistical  significance  of  these  intensive  studies  seem  to 
warrant  some  consideration  at  this  time. 

Findings  have  been  reviewed  which  show  a  correlation  between 
certain  electrophysiological  phenomena  and  differential  condi- 
tioned behavior.  Results  have  been  presented  from  a  number  of 
studies  primarily  designed  to  explore  two  hypotheses  based  on 
this  earlier  work:    1)   The  configuration  of  labeled  potentials  in 


Neural  Mechanisms  of  Decision  Making  275 

these  situations  reflects  the  coding  and  processing  of  information 
as  the  brain  performs  differentiated  conditioned  responses  to  two 
intermittent  photic  stimuh  differing  in  frequency;  and  2)  the 
estabhshment  and  performance  of  such  differentiated  behaviors 
involve  the  measurement  of  similarity  between  past  experience, 
as  reflected  primarily  in  the  neural  activity  of  nonspecific  regions 
of  the  brain,  and  present  stimulus  configuration,  as  refiected 
primarily  by  the  specific  sensory  systems  of  the  brain. 

The  observations  of  "assimilation  of  the  rhythm"  which  have 
been  reported  by  many  workers  suggest  that  the  brain  has  the 
capacity  to  reproduce  previously  experienced  patterns  of  neural 
activity.  Manifestation  of  such  endogenously  generated  patterns 
is  marked  in  nonspecific  systems.  During  generalization,  behavioral 
performance  seems  to  be  accompanied  by  departures  from  stimulus- 
bound  response,  notably  in  the  cortex  of  the  relevant  sensory 
modality  and  in  the  reticular  formation.  Computer  analysis  of 
waveforms  from  various  structures  during  such  behavior  shows 
that  these  two  regions  cHsplay  clear  evidence  of  endogenously 
generated  coinponents  appropriate  to  the  behavior,  while  other 
regions  of  the  brain  respond  to  the  stimulus  more  accurately. 
Analogous  observations  have  been  made  when  cfifferentially 
trained  animals  commit  errors. 

These  data,  which  are  compatible  with  the  hypotheses,  are 
contradicted  by  the  failure  to  elicit  erroneous  performance  dif- 
ferentially as  a  consequence  of  central  stimulation  at  a  frequency 
discordant  with  the  frequency  of  a  concurrent  peripheral  con- 
ditioned stimulus.  It  has,  however,  been  demonstrated  that 
temporal  patterns  of  excitation  at  a  site  can  serve  as  coded  infor- 
mation for  the  brain.  Evidence  has  also  been  presented  indicating 
that  a  crucial  step  in  cortical  data  processing  may  take  place  at 
the  time  when  influences  arrive  from  the  nonspecific  system. 

To  date,  therefore,  we  have  not  succeeded  in  establishing  an 
unequivocal  functional  role  for  labeled  potentials  as  direct  reflec- 
tions of  data  processing  in  the  brain.  However,  an  increasing  and 
consistent  body  of  evidence  does  seem  to  support  the  view  that  a 
cortical-reticular  interaction  is  an  important  component  in  the 
evaluation  of  incoming  information  in  the  context  of  past  ex- 
perience. 


276  Information  Storage  and  Neural  Control 

The  rapid  rate  of  technical  progress  in  this  problem  area  gives 
us  good  reason  to  hope  that  further  clarification  will  shortly  be 
forthcoming. 

REFERENCES 

1.  Corning,  W.  C,  and  John,  E.  R.:  Effect  of  ribonuclease  on  retention 

of  conditioned  response  in  regenerated  planarians.  Science,  134: 
1363,  1961. 

2.  Dingman,  W.,  and  Sporn,  M.  B.:  The  incorporation  of  8-azaguanine 

into  rat  brain  RNA  and  its  effect  on  maze  learning  by  the  rat; 
an  inquiry  into  the  biochemical  basis  of  memory.  J.  Psychiat.  Res., 
7.-1-11,  1961. 

3.  Eiduson,  S.,  Geller,  E.,  and  Beckwith,  W.:  Some  biochemical  cor- 

relates of  imprinting.  Fed.  Proc,  20:345,  1961. 

4.  Glickman,  S.  E.:    Perseverative  neural  processes  and  consolidation 

of  the  memory  trace.  Psychol.  Bull.,  58:218-233,  1961. 

5.  Hyden,  H.:  Biochemical  changes  in  glial  cells  and   nerve  cells  in 

varying  activity,  in  Proc.  Fourth  Internal.  Congr.  Biochem.,  Vienna, 
1958,  J;64-89,   1960. 

6.  John,  E.  R.:  High  nervous  functions:  brain  functions  and  learning. 

Ann.  Rev.  Physiol.,  2J.-451-484,  1961. 

7.  John,  E.  R.:  In,  Theories  of  the  of  Mind,  ed.   by  J.   Scher.   Glencoe, 

lUinois,  Free  Press,  1962.  (in  press) 

8.  John,  E.  R.,  and  Killam,  K.  F.:  Electrophysiological  correlates  of 

avoidance  conditioning  of  the  cat.  J.  Pharmacol.  &  Exp.  Ther., 
125:252-274,  1959. 

9.  John,  E.  R.,  and  Killam,  K.  F.:  Electrophysiological  correlates  of 

differential  approach-avoidance  conditioning  in  cats.  J.  JVerv. 
Ment.  Dis.,  737.-183-201,  1960. 

10.  John,  E.  R.,  Leiman,  A.  L.,  and  Sachs,  E.:  An  exploration  of  the 

functional  relationship  between  electroencephalographic  potentials 
and  differential  inhibition.  Ann.  New  York  Acad.  Sci.,  P2.-1160- 
1182,  1961. 

11.  Kreps,   E.,  cited  by:   Palladin,   A.   V.,   and  Vladimirov,   G.   E.,   in 

Proc.  Internal.  Conf.  Peaceful  Uses  Atomic  Energy,  vol.  12,  1955. 

12.  Liberson,   VV.   T.,   and    Ellen,    P.:    In,   Recent  Advances  in   Biological 

Psychiatry,  ed.  by  J.  Wortis,  New  York,  Grune  and  Stratton,  1960. 

13.  Livanov,   M.   N.,   and   Korolkova,   T.   A.:   ^hur.    Vyss.   Nerv.   Delal. 

Pavlova,  vol.  1,  J.-332-346,  1951;  as  reported   in  Rusinov,  V.  S., 


Neural  Mecha7iisms  of  Decision  Making  277 

and  Rabinovich,  M.  Y.,  Electroencephalographic  researches  in 
the  laboratories  and  clinics  of  the  Soviet  Union.  Electroenceph.  Clin. 
NeurophysioL,   supp.   8,    1958. 

14.  Livanov,   M.   N.,   and   Polyakov,   K.   L.:   Bull.   Acad.   Sc.    USSR,  3: 

286-307,  1945:  as  reported  in:  Rusinov  V.  S.,  and  Rabinovich, 
M.  Y.,  Electroencephalographic  researches  in  the  laboratories  and 
clinics  of  the  Soviet  Union.  Electroenceph.  Clin.  NeurophysioL,  supp. 
8,   1958. 

15.  Loucks,  R.  B.:   In,  Electrical  Stimulation  of  the  Brain,  ed.   bv  D.   E. 

Sheer,  Austin,  University  of  Texas  Press,  1961. 

16.  Majkowski,  J.:   EEG  and  EMG  pictures  of  differentiation  of  con- 

ditional reflexes.  Acta  Physiol.  Pol.,  9.-565-581,   1958. 

17.  Morrell,  F.:  Some  Electrical  Events  Involved  in  the  Formation  of 

Temporary  Connections.  In  Henry  Ford  Hospital  International 
Symposium,  Reticular  Formation  of  the  Brain.  Boston,  Little,  Brown 
and   Co.,    1958. 

18.  Morrell,  F.:  In,  symposium,  Council  for  International  Organization 

of  Medical  Sciences.  Brain  Mechanisms  and  Learning,  ed.  l^y  J.  S. 
Delafresnaye,  Oxford,  Blackwell  Scientific  Publications,  1961. 

19.  Morrell,  F.:  Electrophysiological  contributions  to  the  neural  basis 

of  learning.  Physiol.  Rev.,  4l:AAl)-A9A,  1961. 

20.  Morrell,  F.,  Barlow,  J.,  and  Brazier,  M.  A.  B.:  In,  Recent  Advances 

in  Biological  Psychiatry,  ed.  by  J.  Wortis,  New  York,  Grune  and 
Stratton,   1960. 

21.  Morrell,  F.,  and  Jasper,  H.  H.:  Electrographic  studies  of  the  for- 

mation of  temporary  connections  in  the  brain.  Electroenceph.  Clin. 
NeurophysioL,  5.-201,   1956. 

22.  Neff,  W.  D.,  Nieder,  P.  C,  and  Oesterreich,  R.  E.:  Learned  response 

elicited  by  electrical  stimulation  of  auditory  pathways.  Fed.  Proc, 
vol.  18,  supp.  3,  item  442,  1959. 

23.  Nieder,  P.  C,  and  Neff,  W.  D.:  Auditory  information  from  sub- 

cortical electrical  stimulation  in  cats.  Science,  7JJ.-1010-1011,  1961. 

24.  Stern,  J.  A.,  Ulett,  G.  A.,  and  Sines,  J.  O.:  In,  Recent  Advances  in 

Biological  Psychiatry,  ed.  by  J.  Wortis,  New  York,  Grune  and 
Stratton,   1960. 

25.  von  Foerster,  H.:  Das  Gedachtnis,  Vienna,  Deuticke,   1948. 

26.  Weiss,    M.:    Unpuljlished   master's   thesis.    College   of  Engineering, 

University  of  Rochester,  Spring  1962. 

27.  Zuckermann,  E.:  Effect  of  cortical  and  redcular  stimulation  on  con- 

ditioned reflex  acdvity.  J.  NeurophysioL,  22:633-643,  1959. 


278  Injormation  Storage  and  Neural  Control 

DISCUSSION  OF  CHAPTER  XI 

Frank  Morrell  (Palo  Alto,  California):  I  would  like  to  ask 
two  questions,  Dr.  John,  both  relating  to  your  central  theme. 
In  the  experiment  dealing  with  equivalence  of  central  stimulation 
and  a  peripheral  signal  you  found  that  highly  trained  animals 
punished  for  errors  did  not  transfer.  Do  you  think  that  the  "equiv- 
alence'' you  demonstrate  represents  generalization  rather  than 
transfer  and  actually  serves  to  prove  the  nonequivalence  of  central 
stimulation  and  the  peripheral  signal? 

On  a  more  theoretical  plane,  do  you  think  that  a  representational 
system  which  stores  information  in  the  form  of  an  ongoing  pulse 
code  is  an  efficient  method  for  storage,  or  even  for  comparison? 
And  since  it  is  clear  that  animals  can  distinguish  frequencies  in 
the  visual,  somatic  or  auditory  modalities  which  are  far  above 
the  range  of  EEG  rhythms,  is  it  necessary  to  postulate  an  entirely 
different  coding  mechanism  for  temporal  sequences  beyond  the 
limited  range  within  which  a  translation  in  terms  of  brain  wave 
response  is  possible? 

E.  Roy  John  (Rochester,  New  York):  I  believe  your  first 
question  is  directed  at  whether  information  might  be  encoded  as 
a  wavefoim  and  stored  by  a  mechanism  which  could  reproduce 
waveform.  Clearly,  at  some  level  there  must  be  a  common  domain 
of  discourse  between  information  about  immediate  experience  and 
representation  of  past  experience.  Recognition  of  an  input  requires 
such  an  interaction.  As  I  stated  in  the  beginning,  there  are  various 
possible  ways  that  this  might  be  accomplished.  Some  workers 
have  suggested  neural  filter  networks,  structured  by  experience, 
such  that  passage  of  an  input  constitutes  identification.  It  is  not 
clear  how  such  throughput  is  to  be  related  to  the  experience 
which  stipulated  the  filter  characteristics.  If  a  filter  stands  for 
experience  A  and  permits  passage  of  an  impulse  when  experience  A 
occurs,  no  mechanism  has  been  proposed  which  would  assign  to 
that  passed  impulse  the  content  of  experience  A. 

Other  workers  have  suggested  an  intracellular  macromolecular 
device  functioning  almost  as  a  tape  recorder  to  register  experience. 
To  my  knowledge,  no  mechanism  has  been  proposed  which 
would  achieve  "playback"'  from  these  molecular  recordings.  One 
can  conceive  of  networks  which  would   accomplish  coincidence 


Neural  Mechanisms  of  Decision  Alaking  279 

detection,  in  whicli  things  could  be  compared  in  the  same  coin. 
I  believe  that  the  coin  in  such  networks  might  be  the  spatio- 
temporal  distribution  of  electrical  activity. 

It  is  at  this  point  that  your  second  question  becomes  relevant. 
You  are  really  asking,  "Can  such  encoding  be  possible  for  any- 
thing other  than  the  very  artificial  situation  which  we  have 
devised?  Can  the  temporal  pattern  of  electrical  events  really  be 
suggested  for  the  representation  of  stimulus  configurations  which 
are  not  characterized  by  particular  frequencies  of  events?  Could 
a  temporal  pattern  of  neural  potentials  represent  stimulus  fre- 
quencies above  the  range  of  EEG  rhythms?"  I  think  this  is  clearly 
possible.  One  could  conceive  of  spatio-temporal  transforms  so 
that  a  characteristic  distribution  of  simultaneous  events  in  dif- 
ferent regions  of  the  brain  would  generate  a  characteristic  se- 
quence of  temporal  events  at  soine  loci.  Spatial  distributions 
can  be  transformed  to  temporal  patterns,  and  temporal  patterns 
can  be  transformed  to  spatial  distributions.  Projection  pathways 
of  different  lengths  and  diflferent  propagation  velocities  could 
conceivably  project  a  characteristic  representational  temporal 
pattern  which  would  correspond  to  the  distribution  of  simultaneous 
excitation  in  anatomically  dispersed  areas  of  a  neural  population. 
Representational  patterns  need  not  be  isomorphic  with  that  which 
they  represent. 

Your  question  also  relates  to  parsimony,  which  I  do  not  con- 
sider to  be  a  law  of  nature,  but  which  is  a  help  in  the  intuitive 
ordering  of  probabilities.  Admittedly,  the  conditions  which  we 
use  in  our  experiments  are  artificial,  and  deliberately  so.  Cats 
live  in  a  world  containing  more  information  than  simply  the 
frequencies  of  flickering  lights.  We  hope  that  this  artificial  situ- 
ation might  give  us  some  insight  into  the  processing  of  information. 
However,  once  one  reaches  the  conclusion  that  information  about 
this  carefully  constrained  and  defined  environment  may  be  handled 
by  mechanisms  related  to  the  temporal  patterning  of  electrical  po- 
tentials, a  problem  arises.  If  one  wants  to  postulate  a  diff'erent  mech- 
anism for  coding  other  kinds  of  sensory  information,  one  has 
introduced  chaos  into  the  nervous  system.  Rapid  and  accurate 
integration  seems  more  compatible  with  a  system  which  codes 
all  data  in  one  language  and  decodes  it  in  the  same  tongue  than 


280  Information  Storage  and  Neural  Control 

with  a  system  which  uses  a  different  language  for  each  kind  of 
message.  I  cannot  assert  that  information  about  diverse  events 
is  necessarily  coded  in  the  same  way,  but  I  would  prefer  to  test 
that  hypothesis  rather  than  to  accept  a  doctrine  of  specific  message 
languages.  I  have  presented  evidence  which  suggests  that  the 
temporal  pattern  of  macropotentials  may  be  related  to  the  coding 
of  information  about  flicker  frequency.  Were  I  convinced  that 
this  were  the  code  for  this  carefully  specified  stimulus,  I  would 
be  inclined  to  suggest  temporal  pattern  of  potential  as  the  most 
probable  code  for  other  sorts  of  stimuli. 

I  would  like  to  take  advantage  of  this  opportunity  to  ask  you 
a  question  about  your  paper  this  morning,  which  is  not  unrelated. 
Certain  aspects  of  the  model  which  is  implicit  in  what  you  said 
seem  to  pose  appreciable  difficulties.  If  I  understand  you,  you 
suggested  a  memory  based  on  the  specification  of  protein  sequences 
by  ribonucleic  acid.  You  did  not  touch  on  the  question  of  how 
to  obtain  a  readout  from  this  memory.  To  use  your  terms,  are  you 
sure  that  such  a  memory  would  be  more  efficient  than  a  memory 
which  would  operate  as  follows:  The  spatio-temporal  distribution 
of  activity  caused  by  a  stimulus  in  an  extensively  interconnected 
network  of  cells  is  such  that,  due  to  local  micro-environments, 
characteristic  interspike  intervals,  fiber  diameters  and  distances 
between  elements,  etc.,  there  is  some  population  of  cells  for  which 
re-entrant  pathways  exist  such  that  a  reverberation  can  be  sus- 
tained to  a  given  stimulus  configuration.  Maintaining  this  rever- 
beration for  a  sufficient  time,  which  might  be  the  duration  of  the 
consolidation  period,  might  accomplish  a  change  in  macro- 
molecular  synthesis.  As  the  average  intracellular  electrolyte  con- 
centration was  altered  by  sustained  reverberatory  activity,  a 
change  might  occur  in  the  carbon-to-carbon  bond  angle  of  RNA, 
thus  altering  the  distance  between  purine  and  pyrimidine  bases. 
Such  changes  in  the  spacing  of  the  template  would  alter  the  amino 
acid  species  which  would  fit  in  that  place.  The  concentration  of 
the  appropriate  amino  acid  in  the  environment  would  determine 
the  probability  that  the  appropriate  fit  would  be  made.  The  rate 
of  assemblage  of  a  protein  would  depend  on  the  achievement  of 
the  appropriate  concatenation  of  amino  acids  on  the  template. 
Therefore,  since  the  concentrations  of  various  amino  acids  in  the 


Neural  Mechanisms  of  Decision  Making  281 

cell  differ,  such  ion-induced  changes  in  template  might  accom- 
plish increases  or  decreases  from  the  usual  rate  of  protein  synthesis. 

In  this  view,  macromolecular  specificity  is  relevant  only  insofar 
as  the  control  of  synthesis  rate  is  concerned.  The  synthesized 
material,  no  matter  what  its  specific  configuration,  simply  binds 
charge.  The  presence  of  bound  charge  in  some  cellular  regions 
causes  an  inhomogeneous  distribution  of  diffusible  ions.  The  pro- 
portion of  diffusible  to  bound  charge  need  not  be  the  same  every- 
where within  a  cell  nor  from  cell  to  cell.  Consequently,  the 
equilibrium  concentration  for  diffusible  potassium  is  altered,  the 
rate  of  restoration  of  membrane  polarization  after  discharge  is 
variable,  the  RC  constants  are  different,  and  the  interspike 
intervals  are  different.  By  such  a  mechanism,  this  population  of 
cells  which  sustains  reverberation  could  be  shifted  away  from  the 
population  mean  with  respect  to  the  interspike  interval  and 
become  isolated. 

A  mechanism  of  this  sort  seems  to  provide  a  way  to  get  around 
some  of  the  severe  problems  which  face  a  memory  mechanism 
based  on  protein  specificity.  For  example,  if  memory  were  a 
particularly  specified  macromolecular  configuration,  what  would 
protect  it  once  it  was  built?  Further,  if  RNA  structure  is  at  the 
mercy  of  every  influence  which  impinges  on  the  cell,  if  you  can 
make  many  kinds  of  RNA  depending  on  the  afferent  stimulus 
configuration,  then  how  does  a  cell  sustain  the  enzymatic  activity 
necessary  for  its  survival?  Why  assume  that  you  can  make  any 
kind  of  RNA  or  any  kind  of  protein  or  enzyme?  Why  not  assume 
that  the  cell  can  make  only  stipulated  kinds  of  macromolecules 
with  a  facilitation  mechanism  of  the  sort  I  suggested  regulating 
the  amounts  of  each?  Such  macromolecules  could  play  many 
roles,  including  that  of  binding  charge.  There  is  no  specific  macro- 
molecular sequence  which  requires  protection  to  preserve  memory. 
One  cannot  prevent  the  cell  from  being  excited  to  enable  memory 
to  be  preserved.  Why  not  assume  that  the  protection  is  achieved 
by  randomness? 

A  memory  of  the  sort  I  describe  here  could  be  perturbed  only 
by  a  sustained  nonrandom  influence  which  impinged  on  a  sub- 
population  of  the  coupled  assemblage  of  cells  mediating  a  memory, 
and  only  if  that  influence  were  sustained  long  enough  to  alter 


282  Information  Storage  and  Neural  Control 

the  rates  of  macromolecular  synthesis.  Global  nonrandom,  as  well 
as  random  influences,  and  localized  random  influences  would 
have  no  eff"ect  on  the  relative  interspike  intervals  of  the  ensemble 
of  cells.  All  messages  are  encoded  in  the  same  language  in  such 
a  scheme.  I  see  no  sort  of  information  which  could  not  be  coded 
by  a  spatio-temporal  pattern  of  this  sort.  The  memory  readout 
which  would  result  from  the  simultaneous  activation  of  a  large 
enough  proportion  of  the  cells  in  the  assemblage  to  sustain  itself 
and  propagate  via  re-entrant  pathways  would  generate  a  spatio- 
temporal  distribution  of  electrical  activity  quite  comparable  to 
the  readin.  The  resting  memory  here  is  not  a  reverberation,  but  a 
set  of  structurally  mediated  temporal  relationships  which  generates 
a  reverberation  only  when  the  ensemble  is  excited.  Would  a 
scheme  of  this  sort  seem  to  meet  more  of  the  constraints  with 
fewer  ad  hoc  hypotheses  than  the  sort  of  mechanism  which  you 
had  in  mind? 

Morrell:  I  still  do  not  see  how  your  supposition  does  away  with 
the  notion  that  the  critical  factor  is  the  distribution  of  charged 
sites  available  for  bonding. 

John:  I  think  there  is  an  essential  difference.  Your  code  is  the 
specification  of  sequence  on  a  molecule. 

Morrell:  Only  in  a  limited  sense.  For  example,  a  possible 
alteration  might  involve  only  the  exchange  of  glutamic  acid  for 
glutamine  at  a  specific  site.  Perhaps  potassium  is  involved  as  well. 
We  cannot  say.  All  one  can  reasonably  suggest  is  that  a  change 
in  charge  distribution  on  an  impermeable  molecule  would  be 
necessary  to  influence  the  ionic  environment  permanently. 

John:  There  would  be  many  ways  to  accomplish  the  binding 
of  ionic  charge  which  did  not  require  the  specification  of  sequence. 

Morrell:  Well,  to  put  your  question  back  to  you,  would  not 
the  readout  from  such  a  system  be  a  temporal  pattern  of  cell  dis- 
charge in  space? 

John:  Yes,  the  readout  would  be  a  spatio-temporal  pattern 
of  cell  discharge,  but  it  could  be  a  much  smaller  space  than  the 
set  of  neurons  initially  excited  by  a  stimulus. 

Morrell:   Oh  yes,  there  would  be  a  difference. 


CHAPTER 
XII 

ANASTOMOTIC  NETS  COMBATING  NOISE* 

Warren  S.  McGulloch,  M.D. 


We 


E  INHERITED  from  Greek  medicine  a  recognition  that 
knowledge  depends  in  some  manner  upon  a  mixture  of  a  knower 
and  the  known.  The  Fathers  of  Medicine  supposed  that  this  mixing 
took  place  locally  in  the  anastomotic  veins  and  was  carried  by 
the  blood  to  the  general  mixture  in  the  heart.  Except  for  a  few 
chemical  messengers  like  hormones,  we  have  abandoned  this 
cardiocentric  theory  of  knowledge  for  a  cephalocentric  one.  We 
have  replaced  their  mixture  of  substances  with  an  interaction  of 
signals,  but  have  retained  the  essentially  anastomotic  quality  of 
the  net.  In  fact,  we  conceive  our  nervous  system  to  be  so  anasto- 
motic that  every  efferent  peripheral  neuron  can  be  affected  over 
a  multiplicity  of  paths  by  every  afferent  peripheral   neuron. 

For  the  purposes  of  this  paper,  I  shall  ignore  all  other  sources 
of  reliability  in  the  process  of  perception.  I  mean  such  things  as: 
1)  closed  loops  of  reflexive  and  regulatory  mechanisms;  2)  use  of 
topological  mapping  to  preserve  local  sign;  3)  redundancy  of  code 
that  is  inherent  in  the  repetition  rate  characteristic  of  those  nervous 
structures  that  determine  posture  and  motion;  and  4)  autocor- 
relative  functions  of  the  cerebellum  that  are  used  to  raise  signals 
out  of  a  background  of  noise. 

I  shall  say  nothing  about  evolution,  adaptation,  learning,  or 
repair.  My  reason  is  this:  The  nervous  system  is  state-determined; 
that  is,  at  any  one  time  its  change  into  another  state  is  determined 


*This  work  was  supported  in  part  by  the  U.S.  Army  Signal  Corps,  the  Air  Force 
Office  of  Scientific  Research,  and  the  Office  of  Naval  Research;  in  pai-t  by  the  National 
Institutes  of  Health  Grant  B-1865,  (C3);  and  in  part  by  the  U.S.  Air  Force,  Aero- 
nautical Systems  Division,  under  Contract  AF33(616)-7783. 

283 


284  Information  Storage  and  Neural  Control 

only  by  the  state  in  which  it  is  and  by  the  input  to  that  state. 
Consequently,  we  do  not  care  how  it  came  to  be  in  that  state. 

For  our  problem  of  the  moment,  perception,  we  shall  deal  only 
with  essentially  synchronous  signals  to  a  layer  of  neurons  whose 
axons  end  on  the  succeeding  layer,  for  as  many  layers  in  depth  as 
we  choose.  Such  a  net  can  be  designed  to  compute  in  any  layer 
at  any  one  time  as  many  Boolean  functions  of  its  simultaneous 
inputs  as  there  are  neurons  in  that  layer,  and  no  others.  We  shall 
not  consider  any  other  nets.  We  shall  suppose  that  our  nets  have 
been  designed  so  that  functions  computed  by  the  output  neurons 
lead  to  those  responses  that  are  most  useful  to  the  organism.  This 
assumption  simplifies  our  problem. 

Some  twenty  years  ago,  when  Walter  Pitts  and  I  began  our 
study  of  a  logical  calculus  for  ideas  that  are  immanent  in  nervous 
activity,  there  was  good  evidence  that  a  neuron  had  a  threshold 
in  the  sense  that  it  would  fire  if  adequately  excited;  that  impulses 
from  separate  sources,  severally  subthreshold,  could  add  to  exceed 
the  threshold;  and  that  the  neuron  could  be  inhibited.  For  sim- 
plicity, we  took  inhibition  as  being  absolute.  These  few  properties 
served  our  purpose,  which  was  to  prove  that  a  net  of  such  neurons 
could  compute  any  number  that  a  Turing  machine  could  compute 
with  a  finite  tape.  Some  five  years  later,  these  properties  sufficed 
for  a  theory  of  how  we  can  perceive  universals,  such  as  a  chord, 
regardless  of  key,  or  a  shape,  regardless  of  size.  These  two  papers 
were  crucial  in  the  development  of  Automata  Theory. 

But,  ten  years  ago  the  inadequacy  of  these  assumptions  came  to 
light,  theoretically,  in  von  Neumann's  paper  on  probabilistic  logic 
concerned  with  building  reliable  computers  from  less  reliable 
components. 

By  that  time  spontaneously  active  neurons  had  been  demon- 
strated in  most  parts  of  the  mammalian  nervous  system.  Inhibitions, 
like  excitations,  had  been  found  to  sum,  and  we  had  come  to  grips 
with  those  interactions  of  axons  that  are  afferent  to  a  cell  and  by 
which  signals  in  one  prevent  signals  in  another  from  reaching  the 
recipient  neuron. 

We  could  demonstrate  this  interaction  as  peripherally  as  the 
primary    bifurcation    of  afferent    peripheral    neurons.    In    Nature 


Anastomotic  Nets  Combat mg  Noise  285 

(January  6,  1962),  E.  G.  Gray  has  published  the  first  electron 
microscopic  anatomical  evidence  of  axonal  terminations  upon 
boutons  of  other  axons,  which  may  account,  as  proximally  as 
possible,  for  the  interaction. 

Interaction  of  afferents  is  of  great  theoretical  importance.  First, 
it  enables  a  neuron  to  compute  any  Boolean  function  of  its  inputs, 
i.e.,  to  respond  to  a  specified  set  of  afTerent  impulses,  not  merely 
those  functions  available  to  so-called  threshold  logic;  and,  .second, 
it  permits  a  neuron  to  run  tlirough  all  possible  sequences  of  func- 
tions as  its  threshold  is  shifted. 

The  first  is  of  great  importance  in  audition.  The  Boolean  func- 
tion is  an  exclusive  OR,  and  the  important  cells  are  in  the  superior 
olive.  Each  cell  will  respond  to  an  impulse  from  either  ear  unless 
there  is  one  from  the  other,  but  never  to  both  or  neither.  The 
utility  of  this  arrangement  is  obvious  to  anyone  with  wax  in  one 
ear.  Put  on  a  pair  of  earphones  with  a  beep  in  one  ear  and  drown 
it  10  decibels  under  with  noise.  Next,  put  the  same  noise  into  the 
other  ear  also,  and  the  beep  is  as  loud  and  clear  as  it  is  without 
the  noise.  Finally,  put  that  beep  into  the  other  ear  also  and  it 
disappears,  for  it  is  10  decibels  below  the  noise.  Please  note  that 
this  noise  is  external  to  the  central  nervous  system  and  is  not  the 
kind  that  we  shall  consider  later. 

The  second,  or  sequence  of  functions  determined  by  shifting 
threshold,  is  of  great  importance  in  respiration  but  is  not  so  easily 
stated.  As  nearly  as  I  can  tell  from  old  experiments  and  from  the 
literature,  the  rise  in  threshold  to  electrical  stimulation  that  is  due 
to  ether  is  approximately  the  same  in  all  neurons;  yet  the  respira- 
tory mechanism  continues  to  work  under  surgical  anesthesia  when 
the  threshold  is  raised,  at  least  in  cortex  and  cord,  by  approxi- 
mately 200  per  cent.  The  input-output  function  of  the  respiratory 
mechanism  remains  reasonably  constant,  although  the  threshold 
of  its  component  neurons  has  changed  so  much  that  each  is  com- 
puting a  different  function  (or  responding  to  a  diflTerent  set)  of 
the  signals  it  receives.  Von  Neumann  called  such  nets  "logically 
stable  under  a  common  shift  of  threshold,"  and  Manuel  Blum  has 
cleaned  up  the  problem  for  appropriate  nets  of  neurons  with  any 
number  of  inputs. 


286 


Information  Storage  and  Neural  Control 


To  explain  this,  I  would  like  to  introduce  to  you  the  only 
symbols  with  which  I  have  been  able  to  teach  the  necessary 
probabilistic  logic.  I  use  a  X  with  a  jot  for  true,  a  blank  or  0 
for  false,  a  dash  for  "I  don't  care  which,"  and  a  p  for  a  1  with 
probability  p.  For  "A  alone  is  true"  {i.e.,  a  sign  for  A  alone),  X; 
for  B  alone,  X;  for  both,  X;  and  for  neither,  X.  Then  I  can  write 
the  sixteen  logical  functions,  or  firing  diagrams,  of  a  neuron  with 
two  inputs,  as  shown  in  Figure  1,  and  we  can  draw  the  diagrams, 
as  in  Figure  2,  to  show  how  the  computed  function  depends  upon 
the  threshold  9. 

X  'X  X  X'  'X  ^  X-  'X'  X  'X  ^  X*  't,  X'  X'  '^ 

Figure  1 


A             B 

A 

B 

A             B 

A 

B 

+2\fKVl 

h- 

/ 

1 

\^2 

V 

^ 

4X 

3X 

2X 

1  X 

3X 

2X 

IX 

ox 

2X 

1  X 

OX 

-IX 

IX 

OX 

-1  X 

-2X 

OX 

-IX 

-2X 

-3X 

■X 


0 


X 


Figure  2 


You  will  note  that  the  first  four  neurons,  without  interaction  of 
afferents,  compute  all  but  two  of  the  sixteen  logical  functions, 
and  these  missing  ones  can  be  computed  by  the  two  neurons  at 
the  right  in  Figure  2.  The  upper  right  neuron  does  the  trick  in 
the  superior  olive. 


Anastomotic  Nets  Combating  Noise 

A  B 


287 


Figure  3 


To  explain   respiration,   we   now   use   a   net   of  three   neurons 
(Fig.  3),  and  suppose  that  we  want  [X], 

[1]  [(X)      X      (X)] 


X 


or. 


[2] 


[(X)      X      (X)]    =    [X] 


and  compute  it  as  in  Equation  1,  and  tlien  decrease  every  6  by 
one  so  as  to  compute  tlie  same  [X].  Now  every  component  is  com- 
puting a  new  function  of  its  input;  hence,  this  net  is  logically 
stable  under  common  shift  of  6  over  a  change  of  one  step.  If  we 
were  to  carry  it  a  second  step,  we  would  have  a  net  that  always 
fires  or  never  fires. 

The  maximum  range  for  neurons  with  two  afferents  is  clearly 
two  steps,  but  it  can  only  be  achieved  by  nets  with  interaction  of 
afferents,  and  then  it  can  be  achieved  always  and  for  any  number 
of  afferents  per  neuron.  For  example.  Figure  4  shows  these  expres- 


Information  Storage  and  Neural  Control 


(•^)^-(^)>[X. 


(X-)-X(X)>LaJ 

(X)X(;^)> 


x' 

— 

[x 

— 

X 

Figure  4 


B 


Figure  5 

sions  for  a  net  of  three  neurons  whose  output  neuron  goes  through 
X  and  requires  interaction,  as  does  the  left-hand  neuron.  Obviously 
no  more  is  possible,  for  the  output  would  always  or  never  fire. 

One  more  trick  served  by  interaction  is  the  use  of  separate 
shifts  in  d  that  are  produced  by  feedback  to  secure  flexibility  of 
function.  Consider  the  net  of  Figure  5  in  which  the  feathered 
arrows  indicate  feedback  affecting  0's.  This  net  can  be  made  to 


Anastomotic  Nets  Combating  Noise  289 

compute  fifteen  out  of  the  sixteen  possible  functions.  Had  I  drawn  it 
for  neurons  with  three  inputs  each,  it  could  have  been  switched  so  as 
to  compute  each  of  253  out  of  the  256  logical  functions  of  three 
arguments.  I  strongly  suspect  that  this  is  why  we  have  in  the  eye 
some  100  million  receptors  and  only  approximately  one  million 
ganglion  cells,  but  note  that  it  depends  upon  interaction  of  afferents. 

Finally,  Manuel  Blum  has  recently  proved  that  this  interaction 
enables  him  to  design  nets  that  will  compute  any  one  specified 
function  of  any  finite  number  of  inputs  with  a  fixed  threshold  of 
the  neuron  at  a  small,  absolute  value,  say,  1  or  0.  This  prevents 
the  neuron  from  having  to  detect  the  small  difference  of  two  large 
numbers,  thus  allowing  the  brain  a  far  greater  precision  of  response 
to  many  inputs  per  neuron,  despite  a  fluctuation  of  a  given  per 
cent  of  the  threshold  6.  This  fluctuation  of  6  is  the  first  source  of 
noise  which  I  wish  to  consider. 

The  effective  threshold  of  a  neuron  cannot  be  more  constant 
than  that  of  the  spot  at  which  its  propagated  impulse  is  initiated. 
This  trigger  point  is  a  small  area  of  membrane,  with  a  high 
resistance,  and  it  operates  at  body  temperature.  It  is,  therefore, 
a  source  of  thermal  noise.  The  best  model  for  such  a  trigger  is 
the  Node  of  Ranvier,  and  the  most  precise  measurements  of  its 
value  are  those  of  Verveen.  For  axons  '■^A^  in  diameter,  he  finds 
it  to  be  ^^  ±1  per  cent  of  0;  it  is  larger  for  small  axons.  Moreover, 
his  analysis  of  his  data  proves  that  the  fluctuations  have  the  random 
distribution  expected  of  thermal  noise.  There  are,  of  course,  no 
equally  good  chances  to  measure  it  in  the  central  nervous  system, 
for  one  cannot  tell  how  much  of  a  fluctuation  is  due  to  signals  or 
to  stray  currents  from  other  cells.  Our  own  crude  attempt  on  the 
dorsal  column  of  the  spinal  cord  indicates  far  greater  noise,  but 
not  its  source. 

What  goes  for  thresholds  goes,  of  course,  for  signal  strength; 
and  for  fine  fibers,  say,  0.1^,  the  root  mean-square  value  of  the 
fluctuation  calculated  by  the  equation  of  Fatt  and  Katz  is  —0.5  mv. 
If  we  accept  a  threshold  value  of  15  mv.,  this  is  several  per  cent. 
It  may  be  much  larger. 

Moreover,  it  is  impossible  that  the  details  of  synapsis  are  per- 
fectly specified  by  our  genes,  preserved  in  our  growth,  or  perfected 
by  adaptation.  They  are  certainly  disordered  by  disease  and  injury. 


290 


Information  Storage  and  Neural  Control 

A  B 


kX 


Figure  6 


Nevertheless,  it  is  possible  to  cope  with  these  three  kinds  of 
noise — 0,  signal,  and  synapsis — as  long  as  the  output  of  a  neuron 
depends,  in  some  fashion,  on  its  input  by  an  anastomotic  net  to 
yield  an  error-free  capacity  of  computation.  This  is  completely 
impossible  with  neurons  having  only  two  inputs  each.  The  best 
we  can  do  is  to  decrease  the  probability  of  error.  Consider,  for 
example,  a  net  like  that  of  Figure  6  to  compute  [X],  where  each 
neuron  is  supposed  to  have  0=3,  but  each  drops  independently 
to  2  with  a  frequency  p.  As  long  as  p  is  less  than  0.5,  the  net 
improves  rapidly  as  the  product  of  the  p's  of  successive  ranks 
decreases.  The  trick  here  is  to  segregate  the  errors. 

The  moment  we  look  at  neurons  with  three  inputs,  the  picture 
changes  completely;  but  to  describe  this  change  we  need  to  increase 
the  complexity  of  our  logical  symbols  by  putting  a  circle  on  the  X, 
so  that  inside  it  is  C,  outside  not  C,  as  in  Figure  7(a).  Now  con- 
sider a  net  to  compute  some  function,  say,  all  or  else  none.  We 
can  schematize  this,  as  in  Figure  7(b).  The  dash  is  a  "don't-care" 
condition;  it  may  be  a  1,  or  0,  or  any  p  that  you  choose.  This  net 
makes  no  mistakes.   Let  us  suppose  that  each  of  the  first  rank 


Anastomotic  Nets  Combating  Moise 

AB 

B 


291 


NONE 


ALL  OR   NONE 


(a) 


ABC 


(b) 


(c) 

Figure  7 


292 


Information  Storage  and  Neural  Control 


exerts  +2  excitation  on  the  third.  Then  its  threshold  can  vary 
harmlessly:  3  <  0  <  6,  or  nearly  50  per  cent.  Moreover,  if  the 
threshold  is  better  controlled,  then  the  strength  of  the  signals  can 
vary.  Finally,  if  both  are  fairly  well  controlled,  the  connections 
can  be  wrong,  as  in  Figure  7(c),  and  the  input-output  function 
[S5]  is  still  undisturbed. 

If  we  want  to  extend  our  symbols  to  four  arguments,  then  the 
pattern  becomes  that  of  Figure  8,  and  for  five  arguments  it  becomes 
more  complex.  In  general,  each  new  line  must  divide  all  existing 
areas  into  two;  thus  for  N  inputs  there  are  2  spaces.  Oliver  Selfridge 
and  Marvin  Minsky  have  worked  out  simple  ways  of  making  such 
symbols,  with  sine  waves,  for  any  finite  number  of  inputs. 

Eugene  Prange  has  invented  a  way  of  devising  the  distribution 
of  don't-care  conditions  so  that  there  are  as  many  as  possible  for 
a  net  of  N  neurons  in  the  first  rank  and  one  in  the  output  rank, 
each  rank  having  N  inputs  per  neuron.  The  number  of  don't-care 
conditions,  or  dashes,  depends  upon  the  number  of  ones  in  the 
spaces  for  the  function  to  be  computed.  The  dashes  are  fewest 
when  the  function  to  be  coinputed  has  exactly  one-half  its  spaces 
filled  with  ones.  Manuel  Blum  has  solved  the  following  questions: 
1)  Suppose  that  there  are  no  don't-care  conditions,  or  dashes,  in 
the  symbol  for  the  output  neuron;  what  fraction  of  the  spaces  for 
each  neuron  of  the  first  rank  can  have  dashes  and  the  calculation 
be  error-free  for  the  toughest  function  (half-filled  with  ones),  all 
as  a  function  of  N?;  2)  With  all  those  dashes  in  the  first  rank,  what 


1.0 


0.5 


T 

^""' 

1 

Figure  8 


50 

Figure  9 


100 


Anastomotic  Bets  Combating  Noise  293 

fraction  of  the  spaces  in  the  symbol  of  the  output  neuron  can 
harmlessly  be  dashes?  Figure  9  is  Blum's  diagram,  based  on 
equations  that  are  exact  if  N  is  a  perfect  square  and  fairly  good 
approximations  for  the  rest. 

You  will  notice  that  for  N  less  than  40,  the  output  neuron  (the 
solid  line)  has  fewer  dashes.  At  '^40  they  are  equal,  being  ^^80 
per  cent  of  all  spaces.  For  larger  N,  the  output  neuron  has  the 
larger  fraction;  and,  for  N  =  100,  90  per  cent  of  spaces  in  the 
input  rank  are  dashes,  and  98  per  cent  in  the  output  neuron 
are  dashes.  From  this  outcome,  it  is  very  clear  that  the  output 
neuron  cannot  be  a  majority  organ  like  the  one  for  N  =  3. 

We  all  know  that  real  nervous  systems  and  real  neurons  have 
many  other  useful  properties.  But  I  hope  I  have  said  enough  to 
convince  you  that  these  impoverished  formal  nets  of  formal  neurons 
can  compute  with  an  error-free  capacity  despite  limited  pertur- 
bations of  thresholds,  of  signal  strength,  and  even  of  local  synapsis, 
provided  the  net  is  sufficiently  anastomotic.  If  I  have  convinced 
you,  it  has  been  in  terms  of  a  logic  in  which  the  functions,  not 
merely  the  arguments,  are  only  probable.  But  even  this  prob- 
abilisitic  logic,  for  all  its  don't-care  conditions,  is  adequate  to 
cope  fully  with  noise  of  other  kinds.  Our  neurons  die — thousands 
per  day.  Neurons,  when  diseased,  often  emit  long  strings  of  im- 
pulses spontaneously  and  cannot  be  stopped  by  impulses  from  any 
other  neuions.  And,  finally,  axons  themselves  become  noisy,  trans- 
mitting a  spike  when  none  should  have  arisen  or  failing  to  transmit 
one  that  they  should  have  transmitted. 

To  handle  these  problems  in  which  the  output  of  a  neuron  has 
ceased  to  be  any  function  of  its  input,  von  Neumann  proposed 
what  is  called  "'bundling.'"'  In  the  simplest  case,  one  replaces  a 
simple  axon  from  A  by  two  axons  in  parallel.  This  alters  the  logic, 
for  now  if  all  fibers  in  the  bundle  fire,  A  is  regarded  as  certainly 
true;  if  none  fire,  as  certainly  false;  but  between  these  limits  there 
is  a  region  of  uncertainty — call  it  a  set  of  values  between  true  and 
false.  In  the  simplest  case,  there  are  two  such  intermediate  values. 
Von  Neumann  found  that  if  there  are  only  two  inputs  per  neuron, 
the  neurons  had  to  be  too  good  and  the  bundles  too  big.  To  com- 
pute, say,  X  or  ¥,  with  a  net  constructed  like  the  net  of  Figure  10, 
we  find  that  given  a  probability  of  an  error  on  the  axon,  say. 


294  Information  Storage  and  Neural  Control 

A  B 


Figure  10 

€  =  0.5  per  cent,  to  have  the  bundle  usably  correct  all  but  once 
in  one  million  times,  he  needed  5000  neurons  and  two  more  ranks 
of  5000  to  restore  his  signal  so  that  it  was  usable.  His  difficulty 
was  chiefly  the  poverty  of  the  anastomosis.  We  have  found  that, 
with  the  same  e  and  the  requirement  that  the  bundle  be  usably 
correct  all  but  once  in  one  million  times,  if  each  axon  is  connected 
to  every  neuron,  we  only  need  one  rank  of  10  neurons. 

Leo  Verbeek  has  looked  into  the  problem  of  the  death  and 
fits  of  neurons,  and  has  found  that  again  the  probability  of  an 
erroneous  output  decreases  as  the  number  of  inputs  per  neuron 
and  the  width  of  the  first  rank  (both  5  in  number)  increase,  at 
least  for  probabilities  of  death  and  fits  reasonably  under  50  per 
cent.  Figure  11  shows  his  graph,  where  5  is  the  number  of  inputs, 
p  the  probability  of  error  in  the  input  neurons,  and  a:s(p)  the 
probability  of  erroneous  output.  Even  for  a  small  5,  these  calcu- 
lations are  enormously  laborious. 

We  are  all  much  indebted  to  Jack  Cowan  for  our  knowledge  of 
many-valued  logic  for  handling  bundling,  and  for  conclusive 
evidence  that  this  is  not  the  cleverest  way  to  obtain  reliability. 
He  and  Sam  Winograd  have  made  a  much  greater  contribution, 
which  I  could  not  expound  to  you  if  I  wanted  to,  and  I  do  not 
because  it  will  probably  be  communicated  in  full  by  Professor 
Gabor  for  publication  in  the  Philosophical  Transactions.  Vaguely, 
its  purport  is  this: 


Anastomotic  Nets  Combating  Noise 


295 


0.4  - 


0.2  - 


In  the  theory  of  information  concerned  with  communication, 
there  is  a  theorem,  due  to  Shannon,  that,  by  proper  encoding 
and  decoding,  if  one  transmits  at  something  less  than  the  capacity 
of  a  noisy  channel,  one  can  do  so  with  as  small  a  finite  error  as 
one  desires  by  using  sufficiently  long  latencies.  Except  for  things 
like  X  and  X,  no  one  before  Cowan  and  Winograd  was  able  to 
show  a  similar  information-theoretic  capacity  m  computation. 
They  have  succeeded  for  any  computation  and  for  any  depth  of 
net,  limited  only  by  the  reliability  of  the  output  neurons.  The 
trick  lay  in  a  diversification  of  function  in  a  net  that  was  sufficiently 
richly  interconnected.  Their  fundamental  supposition  is  that  with 
real  neurons  the  probability  of  error  on  any  one  axon  does  not 
increase  with  the  complexity  of  its  neuron's  connections.  The 
recipients  of  most  connections  are  the  largest  and,  consequently, 
the  most  stable  neurons.  Again,  it  is  the  richest  anastomosis  that 
combats  noise  best. 


ACKNOWLEDGMENT 

I  wish  to  acknowledge  the  contributions  of  those  who  have 
worked  with  me  in  this  endeavor,  namely:  Anthony  Aldrich, 
Michael  Arbib,  Manuel  Blum,  Jack  Cowan,  Nello  Onesto,  Leo 
Verbeek,  Sam  Winograd,  and  Bert  Verveen. 


296  Information  Storage  and  Neural  Control 

DISCUSSION  OF  CHAPTER  XII 

Bernard  Saltzberg  (Santa  Monica,  California):  In  the  head- 
phone experiment,  I  assume  you  used  a  single  noise  source  which 
divided  its  power  between  the  earphones.  Was  an  experiment 
attempted  with  two  independent  noise  sources?  How  did  the 
results  come  out? 

Warren  S.  McCulloch  (Cambridge,  Massachusetts) :  It  does  not 
help  much.  It  has  to  be  the  same  noise.  Different  noise  is  no  good. 
What  they  were  trying  when  I  was  last  involved  was  lagging  one 
earphone  a  little  behind  the  other  to  see  what  phase  difference 
they  could  make  in  it  and  still  have  it  work.  As  far  as  I  know,  this 
has  not  been  cleaned  up  yet. 

Gregory  Bateson  (Palo  Alto,  California):  What  is  the  price 
of  this  increased  reliability  in  terms  of  loss  of  educability?  Obviously, 
to  obtain  a  new  function — a  new  relationship — out  of  this  net,  you 
have  to  alter  a  large  number  of  connections.  In  a  sense,  I  suspect 
that  the  more  reliable  your  new  constructions,  the  more  non- 
educable  the  net  becomes;  but  I  am  not  a  good  enough  logician 
to  know  that  this  is  so. 

McCulloch:  Look  at  the  flexibility  end  of  it.  We  have  here  a 
neuron  with  a  couple  of  inputs  (A  and  B)  and  one  output  neuron. 
Let  us  take  the  case  of  three  neurons.  Incidentally,  I  cannot  build 
this  without  the  interaction  of  afferents.  I  have  one  output  neuron. 
Now  I  can  send  signals  back  from  the  central  nervous  system  and 
tell  my  eye  what  it  is  to  look  for,  what  it  is  to  see.  You  get  256 
possible  logical  functions.  You  can  calculate  253  of  them  by  giving 
these  first  rank  neurons  a  nudge  on  the  threshold.  Reliability 
does  not  mean  that  the  net  is  inflexible.  This  is  a  remarkably 
flexible  device.  The  flexibility  goes  up  with  the  anastomosis;  it 
does  not  go  down.  That  is  one  of  the  beautiful  things  about  it. 
If  it  was  simple  majority  logic,  the  situation  would  be  impossible. 
The  stupidest  thing  to  do,  so  to  speak,  if  you  want  to  get  the 
maximum  life  out  of  a  rope  is  to  use  it  until  it  breaks  and  then 
replace  it  with  another  one.  No  mountain  climber  that  I  know 
takes  such  a  chance.  The  next  worse  thing  is  always  to  stretch 
two  ropes  from  man  to  man.  What  you  want  is  the  richness  of 


Anastomotic  Nets  Combating  Noise  297 

connections.  The  dynamics  of  the  picture  is  beginning  to  show  up, 
but  the  matliematics  is  too  comphcated  for  us  as  yet. 

Eugene  Pautler  (Akron,  Ohio) :  What  type  of  detector  would 
be  required  to  recognize  the  results  of  this  output — the  computa- 
tions inherent  in  this  output  neuron? 

McCulloch:  I  think  it  is  probably  all  done  in  the  eye.  Suppose 
you  tell  your  eye  to  look  for  four-leaf  clovers.  You  simply  send 
out  the  message,  "Find  a  particular  pattern  in  those  leaves";  and 
when  you  have  found  it  signal,  "Here  is  one!  Here  is  another!" 
You  knew  what  you  were  looking  for  so  you  set  your  filter  ac- 
cordingly. 

A  frog,  when  he  jumps,  sends  back  impulses  to  his  eye  to  give 
as  great  a  response  as  possible  to  an  affair  of  lesser  curvature  or 
greater  radius  of  curvature,  which  informs  his  eye.  This  works 
during  the  first  part  of  the  jump  while  his  eyes  are  open.  One 
tells  one's  eye  what  to  see,  what  too  look  for.  It  would  be  almost 
unthinkable  that  otherwise  one  could  go  into,  say,  Grand  Central 
Station,  look  off  across  the  hall,  and,  knowing  that  there  is  a 
chance  of  so-and-so  being  there,  find  him,  unless  one  has  in  some 
manner  set  a  filter.  Just  how  much  of  that  matching  is  done 
in  the  eye,  I  do  not  know.  The  mouse,  which  does  not  turn  its 
eyes  and  keeps  them  open,  is  another  nice  animal  to  work  on. 
His  retina  is  the  same  all  over,  and  whether  you  get  a  response 
from  a  particular  ganglion  cell  or  from  a  particular  axon  depends 
upon  whether  the  mouse  is  hungry  or  whether  it  has  smelled  its 
cheese.  If  it  has,  then  it  bothers  to  look,  but  it  will  not  look  the 
rest  of  the  time.  The  mouse  shows  very  little  response  to  any 
visual  stimulus.  The  situation  is  far  too  complicated  to  be  solved 
with  a  set  of  electrodes. 

Homer  F.  Weir  (Houston,  Texas):  In  the  use  of  the  injured 
neuron,  you  are  apparently  producing  noise  from  non-input 
sources.  Is  it  correct  to  say  that  your  injured  neuron  is  putting 
out  output  without  input? 

McCulloch:  Yes.  Either  it  is  doing  that  or  it  is  dead. 

Weir:  At  what  level  would  this  have  to  occur,  relatively  speak- 
ing, before  it  would  override  this  protective  error  mechanism  that 
you  were  speaking  of? 


298  Information  Storage  and  Neural  Control 

McCuUoch:  I  have  not  seen  my  own  cerebellum,  but  I  have 
seen  that  of  many  a  man  my  age.  I  am  in  my  second  century, 
and  I  expect  that  at  least  10  per  cent  of  the  Purkinje  cells  in  my 
cerebellum  are  replaced  by  nice  holes  at  my  age,  but  I  can  still 
touch  my  nose.  It  is  incredible  how  little  brain  has  to  be  left  in 
order  for  it  to  function. 


PART  IV  — THE  HUMAN  NERVOUS  SYSTEM 

Moderator:  Wavne  H.  Holtzman,  Ph.D. 


CHAPTER 
XIII 

THE  INDIVIDUAL  AS  AN  INFORMATION 
PROCESSING  SYSTEM 


James  G.  Miller,  M.D.,  Ph.D. 


c 


CONSIDERING  human  beings  as  information  processing"  sys- 
tems has  in  tiie  last  decade  proved  useful  in  both  experiment  and 
theory.  Some  of  the  hoary  old  problems  of  behavior  and  learning 
theory  have  received  a  new  form  or  have  been  bypassed,  and  some 
fruitful  approaches  to  human  individual,  group,  and  social  be- 
havior have  arisen. 

It  has  been  estimated  (1)  that  in  fifty  years  of  waking  life  an 
individual  may  process  10"^  (ten  thousand  trillion)  bits  of  infor- 
mation. A  person  may  be  looked  upon  as  a  component  in  an 
interpersonal  system  in  which  messages  are  sent  from  one  node 
to  another  along  channels  and  through  nets.  As  an  individual, 
he  may  be  studied  as  a  "black  box''  whose  input-output  relation- 
ships can  be  detei mined,  or  as  a  system  of  interrelated  components 
whose  performance  and  capacities  are  increasingly  available  to 
experimental  investigation. 

At  the  Mental  Health  Research  Institute  of  The  University  of 
Michigan  some  of  us  work  within  the  general  systems  orientation 
which  regards  all  life  as  a  part  of  the  physical  space-time  continuum. 
We  consider  this  continuum  to  be  organized  into  a  hierarchy  of 
levels  of  systems,  all  of  which  have  subsystems  and  are  themselves 
subsystems  of  larger  organizations  or  supersystems.  The  smallest 
living  system,  the  cell,  is  composed  of  nonliving  molecules.  These 
may  be  free-living  or  may  be  components  of  organs,  which  in 
turn  are  organized  into  more  complex  individual  systems.  These 

301 


302  Injormaiion  Storage  and  Neural  Control 

may  band  into  face-to-face  groups  or  larger  social  organizations 
and  societies. 

There  is  continuity  and  there  are  cross-level  similarities  in 
structure  and  process  at  all  levels  of  this  hierarchy,  even  though 
there  are,  of  course,  at  the  same  time,  many  specific  differences 
among  individual  systems,  species,  and  levels.  We  have  sought 
for  and  found  cross-level  "formal  identities"  which  can  be  studied 
experimentally. 

All  living  systems  are  open  systems.  That  is,  they  maintain 
steady  states  of  several  variables  and  counteract  entropic  dis- 
integration by  means  of  inputs  and  outputs.  Living  systems  at 
all  levels  process  both  energy  and  information.  These  always  flow 
together.  For  example,  energic  inputs  such  as  food  convey  infor- 
mation in  the  patterning  of  their  molecular  structures,  and  coded 
verbal  communications  are  carried  on  the  energy  of  sound  waves. 
Energic  and  informational  inputs  are  distinguished  by  whether 
the  receiver  responds  to  their  energic  or  their  informational 
aspects.  Sometimes  the  response  is  to  both. 

SUBSYSTEMS 

Living  systems  at  any  level  require  certain  crucial  subsystein 
functions  in  order  to  survive,  unless  they  exist  in  a  relationship  of 
parasitism  or  symbiosis  with  another  system  which  supplies  them. 
Free-living  cells,  for  example,  may  be  shown  to  have  subsystems 
that  accomplish  all  the  essential  functions,  while  cells  which  are 
part  of  organs  may  lack  some  of  them.  Groups  which  survive  over 
time  isolated  from  other  people  have  all  these  subsystems  while 
groups  which  are  parts  of  organized  societies  almost  never  do. 
Subsystems  may  be  either  local,  like  the  eye,  or  dispersed,  like 
the  reticuloendothelial  system. 

There  are  essential  subsystems  which  deal  with  the  processing 
of  energy  and  others  which  process  information.  The  essential 
energy-processing  subsystems  in  the  general  order  of  their  operation 
are:  boundary,  ingestor,  distributor,  decomposer,  producer,  energy 
storage,  excretor,  and  mover  or  output  transducer. 

The  essential  subsystems  in  information  processing,  listed  in  the 
general  order  of  flow  in  information  processing  are: 


The  Individual  as  an  Information  Processing  System  303 

1)  Boundary.  This  may  be  the  limits  of  the  sense  organ  of  a 
cell  or  animal  or  the  mechanisms  of  a  group  or  society  which 
receive  information  from  outside  the  system. 

2)  Input  Transducer.  A  transducer  changes  energy  from  one 
form  to  another.  The  sense  organ  of  an  animal  transduces  patterned 
energic  inputs  to  nerve  impulses.  There  are  analogs  at  the  society 
level  in  the  translaters  that  receive  and  recode  information  from 
outside  the  society. 

3)  Internal  Transducer.  This  subsystem  receives  and  passes 
on  information  from  within  the  system,  as  the  input  transducer 
does  from  without.  In  an  animal  there  is  the  system  of  internal 
sense  organs  and  chemical  sensitivities  which  activate  control 
mechanisms.  There  are  analogs  at  the  group  and  society  levels. 

4)  Channel  and  Net.  The  channel  is  the  route — neuron,  wire, 
air  or  ether — over  which  a  message  is  sent  from  a  transmitter  to 
one  or  more  receivers.  In  the  individual  the  sensory  nerves  are 
channels  over  which  the  input  is  transmitted  to  the  central  nervous 
system.  Channels  may  intersect  at  points  called  nodes  and  may 
be  interconnected  to  form  a  net.  The  nervous  system  of  individuals 
is  an  information  processing  net.  The  blood  and  lymph  of  the 
individual  also  act  as  information  carrying  channels  as  well  as 
energy  distributors.  There  are  two  distinct  common  uses  of  the 
word  "channel."'  The  more  restricted  meaning  includes  only  the 
flow  route  for  the  information,  without  intervening  subsystems 
of  any  other  sort  (such  as  transducers,  decoders,  or  encoders). 
The  other,  broader  meaning  includes  such  components  together 
with  the  intervening  flow  routes.  "Channel"  is  employed  in  both 
these  senses  in  electronics  and  little  confusion  appears  to  result. 
We  follow  the  second  usage. 

5)  Decoder.  The  decoder  alters  input  information  into  a  code 
or  language  which  can  be  transmitted  and  "interpreted"  inside 
the  systein. 

6)  Learner.  This  subsystem  establishes  a  reliable  and  enduring 
association  between  certain  information  inputs  and  other  infor- 
mation from  outside  or  inside  the  system.  Thereafter,  the  system 
will  make  an  altered  output  to  an  input  which  previously  elicited 


304  Information  Storage  and  Neural  Control 

another  response,  or  no  response,  or  make  the  same  output  to  a 
different  input. 

7)  Memory.  This  subsystem  stores  information  over  time. 

8)  Decider.  A  given  set  of  inputs  may  ehcit  two  or  more 
alternate  outputs.  The  decider  selects  the  one  that  is  put  into 
action.  Each  of  the  subsystems  of  a  system  is  also  a  system  at  its 
own  level  and  must  make  its  own  decisions,  as  well  as  carry  out 
other  critical  functions.  The  neuron  has  the  binary  decision  to 
fire  or  not  to  fire,  which  is  based  upon  the  strength  and  charac- 
teristics of  its  inputs  and  the  present  state  of  the  neuron.  The 
individual  has  a  central  decision-making  subsystem  which  deter- 
mines output  for  the  whole  system. 

9)  Encoder.  This  prepares  information  for  output  by  putting 
it  into  a  code  which  can  be  transmitted  to  and  interpreted  by 
other  systems  in  the  environment. 

10)  Motor  or  Output  Transducer.  The  motor  in  an  animal 
is  the  same  for  both  energy  and  information  outputs.  Nervous 
impulses  trigger  activities  like  gross  physical  movements,  speech, 
ingestion,  or  excretion. 

11)  Reproducer.  This  is  capable  of  giving  rise  to  other  systems 
similar  to  the  one  in  which  it  is  found.  We  consider  it  an  infor- 
mation processing  subsystem  because  its  primary  activity  is 
transmission  of  information  or  patterning.  The  reproducer,  while 
not  essential  for  the  survival  of  the  individual,  is  necessary  for 
the  continuation  of  the  species  and  all  social  organizations  which 
endure  for  more  than  one  generation. 

Each  of  these  subsystem  functions  is  carried  out  within  the 
individual,  but  as  we  have  seen  in  this  symposium,  it  is  not  possible 
at  present  to  show  the  precise  localization  of  all  of  them.  The 
specific  neural  arrangements  for  decoding,  learning,  memory, 
perception,  deciding,  and  encoding,  for  example,  are  all  being 
studied  but  are  not  yet  understood. 

We  have  emphasized  the  use  of  standard  centimeter-gram- 
second  or  information  theory  units,  or  units  which  are  derivatives 
of  these,  rather  than  the  welter  of  unrelated  measures  which 
have  been  used  in  the  different  fields  of  behavioral  science.  Since 
we    are   looking   for   cross-level    measurable    uniformities   or   dif- 


The  Individual  as  an  Information  Processing  System  305 

ferences,  the  quantitative  study  of  tiiese  requires  the  use  of  com- 
parable measures  at  different  levels,  and  the  units  of  the  natural 
sciences  seem  best  suited,  though,  of  course,  all  sorts  of  phenoinena 
cannot  yet  be  expressed  in  them. 

THE  ROUTE  OF  INFORMATION  FLOW 

Each  one  of  a  person's  subsystems  may  participate  in  the 
preparation  of  the  output.  Input  of  appropriate  kind  and  strength 
crosses  the  individual  boundary  and  is  transduced  into  the  proper 
form  for  nervous  transmission.  If  a  language  or  code  is  involved, 
it  is  translated  by  the  decoder  and  classified  by  the  perceiver  in 
terms  of  a  perceptual  schema  which  represents  the  world  as  the 
individual  has  experienced  it.  Reference  may  be  made  to  stored 
memories.  There  may  be  some  recoding  or  other  preparation  of 
all  or  part  of  the  input  for  storage  in  the  memory.  On  the  output 
side,  a  decision  is  made  from  among  the  alternate  possible  outputs; 
encoding  for  external  transmission  is  carried  on,  and  the  nervous 
message  is  transduced  into  physical  response,  through  either  the 
speech  mechanism  or  other  musculature.  There  is  a  large  literature 
on  each  of  these  functions  and  it  is  impossible  to  do  more  than 
give  a  brief  review  of  some  of  the  material  on  some  of  the  sub- 
systems. Not  all  input,  of  course,  is  channeled  through  all  the 
subsystems.  A  reflex  response  to  an  input  may  involve  only  a 
small  number  of  subsystems.  Complex  decisions  may  make  use 
of  the  whole  range  of  individual  subsystems. 

Throughout  the  system  there  is  a  continual  and  cumulative 
loss  of  information.  One  important  aspect  of  the  response  of 
biological  systems,  as  both  Gerard  (2)  and  Piatt  (3)  have  recog- 
nized, is  amplification.  That  is,  the  energy  in  the  signal  is  very 
small  compared  to  the  energy  in  the  response.  At  the  same  time 
there  is  a  loss  of  dimensionality  from  input  to  output  in  all  am- 
plifiers, which  must  select  in  order  to  amplify,  since  they  have 
limited  power  available.  There  is  distortion  of  information  at 
each  boundary  that  is  crossed,  and  furthermore,  noise  alters  the 
signal.  The  sense  organ  reacts  only  to  part  of  the  information 
present  in  the  environment.  The  perceiver  screens  and  organizes 
the  input  further,  and  in  the  process  ignores  that  part  of  it  which 


306  Information  Storage  and  Neural  Control 

seems  irrelevant.  Channel  capacity  may  be  lower  than  the  capacity 
of  the  components.  When  the  behavior  is  organized  the  central 
decision  represents  only  a  small  part  of  the  original  input  in- 
formation. 

OVERVIEW  OF  RESEARCH  ON  SUBSYSTEM  FUNCTIONS 

For  an  input  to  cross  the  boundary  into  the  system  its  energy 
must  be  great  enough  to  cause  the  external  transducer  to  fire. 
The  signal-to-noise  ratio  also  must  be  sufficiently  high.  Other 
environmental  conditions  which  may  influence  the  permeability 
of  a  boundary  to  an  input  are  competing  signals,  and  the  simi- 
larity of  the  background  to  the  signal — for  example,  a  white 
stimulus  on  a  white  ground  may  not  be  detected.  McCulloch  gave 
an  example  of  this  in  the  experiment  he  mentioned  in  his  paper 
on  detecting  a  signal  against  monaural  and  binaural  background 
noise. 

Classical  psychophysics  in  its  study  of  the  threshold  has  tended 
to  ignore  some  of  the  important  aspects  of  signal  detectability 
or  to  assume,  sometimes  incorrectly,  that  these  other  things  are 
held  constant.  Swets,  Tanner  and  Birdsall  (4)  have  pointed  out 
that  this  classical  concept  of  the  threshold  is  unreasonable  because 
it  ignores  the  control  which  is  exerted  by  sensory  and  psychological 
variables.  That  is,  it  neglects  the  participation  of  subsystems  other 
than  the  boundary. 

The  characteristics  of  human  sense  organs  as  input  transducers 
or  internal  transducers  may  be  specified  just  as  the  characteristics 
of  electronic  transducers:  by  transfer  function,  band  width,  phase 
shift,  and  signal-to-noise  ratio.  In  these  respects  various  sensory 
subsystems  or  modalities  perform  quite  differently. 

The  transfer  function  of  a  transducer  refers  to  its  ratio  of  output 
to  input.  In  the  visual  system  this  is  the  relationship  between 
intensity  of  light  and  reported  brightness.  In  the  auditory  system 
it  is  the  relationship  between  intensity  of  sound  and  reported 
loudness.  Some  engineers  in  designing  apparatus  for  man-machine 
systems  have  mistakenly  assumed  that  the  cu'-ve  of  perceived  in- 
tensity rises  linearly  with  the  increase  in  strength  of  the  stimulus. 
As  Stevens  has  pointed  out,  the  subjective  intensity  increases  as  a 


The  Individual  as  an  Information  Processing  System  307 

power  function  of  the  stimulus  magnitude.  The  exponent  of  this 
function  for  loudness  is  about  0.3,  while  it  is  about  3.5  for  the 
apparent  intensity  of  electric  current  applied  to  the  fingers. 
Stevens  (5)  notes  that:  "In  three  modalities  investigated  .  .  . 
transducers  .  .  .  have  three  radically  different  operating  charac- 
teristics. The  slow  growth  of  loudness  (exponent  less  than  one) 
suggests  that  the  ear  behaves  as  a  'compressor'  .  .  .  This  com- 
pressor action  probably  helps  to  make  it  possible  for  the  ear  to 
respond  to  an  enormous  range  of  sound  pressures — range  of 
millions  to  one.  The  apparent  intensity  of  vibration  on  the  finger 
tip  grows  almost  linearly  with  vibration  amplitude — as  though 
the  transducer  were  approximately  linear.  The  effective  range  of 
vibration  amplitudes  to  which  the  finger  is  sensitive  is  of  the  order 
of  hundreds  to  one.  (Incidentally,  vibration  on  the  arm  does  not 
follow  a  simple  power  law.)  The  steep  operating  characteristic 
for  electric  shock  suggests  the  action  of  an  'expander'  of  some 
sort;  doubling  the  current  increases  the  sensation  about  tenfold. 
And  correlated  with  this  rapid  expansion  is  a  narrow  operating 
range  of  stimuli  of  the  order  of  only  tens  to  one." 

In  input  transducers  the  output  signal  usually  differs  from  the 
input  signal  in  bandwidth  characteristics.  For  instance,  light  of 
different  wave  lengths  and  sound  of  different  frequencies  are 
subjectively  reported  as  various  colors  and  pitches.  Sensitivity 
over  the  range  of  light  waves  and  sound  waves  is  not  uniform. 

Also  input  transducers  are  active  over  only  a  limited  range. 
There  are  light  waves  above  and  below  the  visible  spectrum  and 
sounds  which  the  human  ear  cannot  hear. 

Phase  shift  refers  to  the  lag  in  phase  of  the  output  signal  over 
the  input  signal.  Input  transducers  differ  in  speed  of  transmission. 
For  example,  sound  waves  travel  through  the  atmosphere  quite 
slowly  but  are  transmitted  rapidly  through  the  auditory  organ, 
while  light  waves,  which  reach  the  eye  very  speedily,  are  processed 
through  a  slow  input  transducer.  Input  transducers  also  differ  in 
the  amounts  and  kinds  of  noise  they  insert  into  the  signal. 

Channel  and  Net 

Broadbent  (6)  suggests  that  the  whole  individual  may  be  re- 
garded as  a  single  channel  which  performs  a  selective  operation 


308  Information  Storage  and  Neural  Control 

upon  the  input,  stores  part  of  it,  filters  it,  and  transmits  it  over  a 
limited-capacity  channel  to  long-term  storage,  to  the  output  trans- 
ducer, or  to  both.  Here  Broadbent  includes  all  the  components  of 
the  individual  system  in  one  channel,  which  is  one  possible  way 
to  view  the  system.  This  results,  however,  in  ascribing  to  channel 
activity  some  things  which  we  have  analyzed  as  subsystem  functions. 
Within  the  channel  he  analyzes  components  which  filter,  store, 
decide,  and  so  forth. 

Quastler  (7)  analyzes  the  activity  of  specific  channels  in  terms 
of  speed,  diversity,  order  of  complexity,  range,  and  other  factors. 
Electronics  engineers  measure  in  channels  the  variables  of  process- 
ing time,  channel  capacity,  bandwidth,  signal-to-noise  ratio,  and 
phase  shift  or  lag.  These  can  all  be  usefully  applied  to  the  animal 
or  human  being. 

The  processing  time  through  neurons  is  brief  compared  to  the 
total  response  time.  The  duration  of  neural  propagation  of  an 
impulse  differs  with  the  length  of  the  channel  and  the  type  and 
size  of  the  neuron.  Longer  transmission  delays  occur  at  the  per- 
ceiver  and  the  decider. 

Channel  capacity  is  a  valuable  concept  in  behavior  theory. 
Broadbent  (8)  says:  "...  perhaps  the  point  of  permanent  value 
which  will  remain  in  psychology  if  the  fashion  for  communication 
theory  wanes,  will  be  the  emphasis  on  problems  of  capacity.  The 
latter,  in  communication  theory,  is  a  term  representing  the  limiting 
quantity  of  information  which  can  be  transmitted  through  a  given 
channel  in  a  given  time  .  .  .  the  fact  that  any  given  channel  has 
a  limit  is  a  matter  of  central  importance  to  communication  engi- 
neers, and  it  is  correspondingly  forced  on  the  attention  of  psy- 
chologists who  use  their  terms." 

Quastler  (9)  was  interested  in  finding  how  much  information 
man  can  process  at  best.  His  research,  therefore,  was  designed  so 
that  neither  the  visual  input  nor  the  muscular  output  were  in 
any  way  hampered.  In  these  tasks  all  inputs  came  from  a  single 
source,  all  output  choices  were  mechanical,  and  all  displays  and 
operations  were  thoroughly  familiar.  He  studied  rates  at  which 
information  is  transmitted  by  reading,  typing,  playing  the  piano, 
doing  mental  arithmetic,  or  assimilating  by  glancing  at  displays 


The  Individual  as  an  Information  Processing  System  309 

of  letters,  playing  cards,  scales,  or  dials.  His  research  was  designed 
to  establish  the  principal  factors  limiting  performance. 

With  these  experimental  conditions,  the  performances  which 
were  obtained  were  at  peak  rates  which  could  have  been  achieved 
only  under  favorable  conditions.  Quastler  (10)  says:  "We  find 
that  people  can  make  up  to  five  to  six  successful  associations  per 
second,  can  transmit  about  twenty-five  bits  per  second,  can  operate 
efficiently  over  a  range  of  about  thirty  possible  values  and  can 
assimilate  some  fifteen  bits  at  a  glance.  We  do  not  expect  that 
they  will  reach  such  perfonnance  levels  with  every  kind  of  activity; 
in  fact,  we  know  that  they  usually  do  not.''  In  bits  per  second,  he 
and  his  colleagues  found  peak  performances  for  piano  playing  of 
twenty-two  bits;  for  reading  aloud,  twenty-four  bits;  and  for 
mental  arithmetic,  twenty-four  bits.  They  concluded  that  the 
peripheral  input  mechanisms  were  not  responsible  for  limitations 
upon  information  processing.  Quastler  (11)  notes:  "The  capacity 
of  the  optic  nerve  is  many  orders  of  magnitude  higher  than  twenty 
or  forty  bits  per  second ;  a  much  wider  range  of  symbols  could  be 
accommodated  with  the  resolving  power  of  the  retina.  As  to  speed 
limitations,  it  is  known  that  about  three  symbols  are  grouped  in 
the  act  of  reading,  and  that  about  four  such  groups  can  be  assimi- 
lated in  a  second;  this  gives  twelve  syinbols  per  second,  con- 
siderably more  than  the  highest  useful  speed  in  typing  or  piano 
playing.  On  the  output  side,  it  is  easy  to  see  that  the  limitations 
of  the  actual  speed,  both  alone  and  in  combination  with  precision, 
cannot  be  attributed  to  mechanical  difficulties.  In  all  tests,  ob- 
served speeds  would  have  been  much  improved  by  rehearsing. 
Thus  the  mechanisms  which  limit  the  observed  performance  must 
be  connected  with  the  speed  of  processing  information." 

Signal-to-noise  ratio  can  be  important  in  the  specification  of 
channels  where  minimal  energies  are  involved.  Barlow  (12)  has 
shown  that  the  limiting  factor  in  the  absolute  threshold  for  vision 
is  fluctuation  in  the  noise  in  the  visual  pathways. 

The  Decoder 

If  information  is  to  be  used  by  the  individual,  it  must  be  suitably 
coded.  That  is,  it  must  be  in  a  language  or  signal  system  which 
he  can  understand.  Deininger  and  Fitts  (13)  have  experimented 


310  Information  Storage  and  Neural  Control 

upon  the  relationship  between  the  input  code  and  performance. 
They  found  that  an  inefficient  or  inadequate  code  can  retard  the 
transmission  of  information  in  perceptual-motor  performance. 
Decoding  time,  therefore,  can  make  a  measurable  difference  in 
information-processing  rate.  Coding,  of  course,  is  important  in 
the  formation  of  concepts  since  this  involves  the  classification  of 
various  things  under  categories  which  ignore  differences  among 
them  and  emphasize  similarities.  Brown  and  Lenneberg  (14)  found 
that  when  subjects  were  asked  to  name  colors  as  quickly  as  possible, 
the  average  reaction  time  was  shorter  and  the  degree  of  agree- 
ment among  subjects  was  higher  when  there  was  a  word  which 
described  the  color.  When  the  color  had  no  special  name  but 
had  to  be  called  "greenish-yellow,"  or  something  like  that,  there 
was  hesitation  and  inconsistency.  Their  matrix  of  intercorrelations 
yielded  a  general  factor  which  they  called  codability.  There  is  a 
large  literature  on  semantic  problems  of  coding. 

The  contributions  of  the  learner  to  information  processing  are 
both  more  familiar  and  less  easy  to  distinguish  from  other  functions 
than  the  more  peripheral  processes.  Some  have  tried  to  make 
learning  theory  cover  nearly  all  of  psychology.  There  has  been 
much  research  on  learning,  but  little  strictly  in  terms  of  infor- 
mation theory,  in  which  it  should  be  viewed  as  the  associating 
of  two  or  more  signals. 

Competing  theories  about  the  memory  have  been  treated  in 
detail  by  other  speakers  in  this  symposium.  Just  how  information 
is  stored  over  time,  and  how  it  is  searched  for,  still  is  not  known. 

Deciding 

Deciding,  as  we  have  said,  goes  on  in  each  subsystem,  as  well 
as  at  the  system  level.  Much  of  psychology  concerns  choices  and 
judgments  of  various  sorts — psychophysical  judgments,  sociometric 
choices,  economic  and  social  decisions,  and  so  forth.  Recent  work 
in  game  theory,  utility  theory,  statistical  decision  theory,  and  group 
effects  on  judgments  of  their  members  is  clarifying  the  processes 
of  complex  decisions.  In  complex  reaction-time  experiments  it  is 
possible  to  calculate  accurately  the  amount  of  time  which  is  added 
to  the  response  time  when  a  choice  of  behaviors  is  involved.  This 
time  falls  to  zero  as  the  task  is  better  practiced  and  the  choice 
becomes  automatic  (15). 


The  Individual  as  an  Information  Processing  System  311 

We  have  been  interested  in  one  aspect  of  channel  capacity 
which  can  be  studied  at  five  levels  of  living  systems.  What  happens 
at  each  level  when  a  channel  is  overloaded? 


INFORMATION  INPUT  OVERLOAD 

From  a  review  of  the  literature  we  were  able  to  draw  a  curve 
which  appeared  to  apply  at  each  level.  The  general  shape  of 
this  performance  curve  shows  the  output  (in  bits  per  second)  rising 
as  a  more  or  less  linear  function  of  input  until  channel  capacity 
is  reached,  then  leveling  off  and  finally  decreasing  in  the  con- 
fusional  state.  This  cross-level  generality  appeared  fairly  con- 
vincingly in  the  empirical  work  of  others,  even  though  it  was  not 
recognized  as  such  by  them.  At  the  same  time,  we  also  found 
suggestions  as  to  hierarchical  differences  among  the  levels.  The 
overall  impression  of  the  findings  is  that  channel  capacity  decreases 
from  cells  to  organs,  to  individuals,  to  groups,  to  social  organi- 
zations. Processes  of  adjustment  appear  to  be  comparable  at 
different  levels. 

We  have  hypothesized  that  there  are  limited  numbers  of  such 
adjustment  processes  which  behaving  systems  can  enlist  as  stresses 
on  them  increase.  The  following  adjustment  processes,  or  mech- 
anisms of  defense,  seein  to  be  used  by  living  systems  against  the 
stresses  of  information  input  overload.  Not  all  living  systems  have 
all  these  inechanisms.  The  smaller  systems,  like  neurons,  appear 
to  have  fewer  than  the  larger  systems,  like  societies,  which  not 
only  have  all  of  them  but  also  have  complicated  variations  of 
them.  These  appear  to  be  the  fundamental  mechanisms,  but  this 
may  not  be  an  exhaustive  list: 

1 )  Omission,  which  is  simply  not  processing  information  whenever 
there  is  an  extreme  of  overload; 

2)  Error,  which  is  processing  incorrectly,  then  not  making  the 
necessary  adjustment; 

3)  Queuifig,  which  is  delaying  responses  during  peak  load  periods 
and  then  catching  up  during  lulls; 

4)  Filtering,  which  is  systematic  omission  of  certain  categories 
of  information,  according  to  soine  priority  scheme; 


312  Information  Storage  and  Neural  Control 

5)  Approximation,  which  is  an  output  mechanism  whereby  a  less 
precise  or  less  accurate  response  is  given  because  there  is  no  time  to 
be  precise; 

6)  Multiple  channels,  which  parallel  transmission  subsystems  that 
can  do  comparable  tasks  at  tiie  same  time  and  consequently  to- 
gether can  handle  more  information  than  a  single  channel  can 
transmit  alone; 

6a)  Decentralization,  which  is  a  special  case  of  this;  and,  finally 
there  is 

7)  Escape,  which  is  leaving  a  situation  entirely  or  taking  any 
other  steps  that  effectively  cut  off  the  flow  of  information. 

Thus  we  have  searched  for  quantitative  similarities  and  differ- 
ences among  living  systems  at  all  levels  in  the  way  they  react  to  in- 
formation input  overload,  and  have  given  special  attention  to  a) 
performance  characteristics  of  a  system  as  an  information  processing 
channel;  and  b)  associated  adjustment  processes  used  to  relieve 
stress  on  the  information  processing  subsystem  and  maintain  per- 
formance. 

Our  original  intention  in  approaching  the  problem  of  over- 
loading living  systems  with  information  was  to  study  a  single 
variable — the  input-output  rate  relationship — postulating  a  formal 
identity  of  this  function  in  channels  at  all  levels  of  living  systems. 
But  this  proposition  turned  out  to  involve  numerous  others  about 
many  variables  representing  other  aspects  of  systems.  The  whole 
problem  ramified  in  a  fascinating  way. 

We  built  apparatuses  and  designed  procedures  which  we  hoped 
would  provide  stable  conditions  for  collecting  performance  data 
from  the  systems  we  selected  for  study,  attempting  to  hold  con- 
stant as  many  of  the  variables  not  under  investigation  as  possible. 
We  were  not  concerned  primarily  with  obtaining  the  maximum 
possible  transmission  rates  from  our  systems,  but  rather  attempted 
to  create  a  stable  situation  in  which  we  could  test  our  overload 
proposition  and  be  sure  we  knew  when  overload  occurred.  Later 
we  could  study  as  independent  variables  those  functions  which 
change  a  given  system's  maximum  channel  capacity. 

Since  information  bits  per  second  had  been  used  by  others  in 
researches  at  all  five  levels,  we  believed  this  to  be  a  suitable  measure 


The  Individual  as  an  Information  Processing  System  313 

of  performance.  We  realized  that  at  each  level  we  would  encounter 
a  complex  statistical  problem  if  we  used  limited  sequences  of 
inputs.  We  also  met  other  problems  in  calculating"  bits,  particularly, 
in  knowing  what  code  was  employed  at  the  cell  and  organ  levels, 
and  in  knowing  the  exact  size  of  the  implicit  ensemble  at  all  levels. 
We  hope  our  methods  at  least  begin  to  cope  with  these  issues. 

Cellular  Research 

A  stimulator  was  constructed  which  could  administer  pulses  to 
a  neuron  at  various  average  rates,  and  at  various  intensities  at 
each  of  these.  A  single  fiber  in  the  sciatic  nerve  of  the  frog  was 
isolated  by  microdissection,  and  was  stimulated  at  the  rates  of 
100,  200,  400,  600,  800,  and  1,000  pulses  per  second,  using  four 
different  values  of  stimulus  voltage  (1,  5,  2,  0,  2.5,  and  3.0  times 
the  threshold  value).  We  recorded  the  output  of  the  fiber  thus 
stimulated  from  microelectrodes  in  the  same  cell  and  across  a 
synapse  in  the  next  cell. 

As  the  input  rate  was  increased,  the  fiber  eventually  ceased  to 
follow  every  input  and  started  missing  some.  Among  the  fibers 
which  we  have  studied,  three  different  types  of  responses  have 
been  observed.  Some  fibers,  when  they  reach  the  point  at  which 
they  can  no  longer  follow  every  stimulus,  start  skipping  every 
other  stimulus.  As  the  rate  is  further  increased  they  respond  only 
to  every  third  or  fourth  stimulus  in  a  regular  fashion.  Other  fibers 
skip  in  a  perfectly  random  manner,  so  that  at  a  given  rate  the 
number  of  pulses  skipped  will  have  a  Poisson  distribution.  Still 
other  fibers  transmit  several  adjacent  stimuli  and  then  fail  to 
transmit  any  stimuli  at  all  for  a  long  period,  after  which  they 
again  fire  repeatedly.  Sometimes  all  three  types  of  functions  are 
found  in  the  same  fiber  at  different  times  and  at  different  rates  of 
stimulation. 

Two  other  phenomena  were  also  noted.  As  the  rate  of  stimu- 
lation was  increased,  there  was  a  fall  in  the  amplitude  of  the 
response  and  a  decrease  in  the  lag  between  the  occurrence  of  the 
input  and  the  start  of  the  response  pulse.  The  amplitude  decrease 
is  probably  related  to  the  energetics  of  membrane  recovery;  the 
lower  recovery  time  leads  to  a  lower  potential.  The  decrease  in 
latency  must  have  a  similar  explanation;  it  makes  the  fiber  able 


314  Information  Storage  and  Neural  Control 

to  cope  with  a  greater  overload,  enabling  it  to  follow  at  much 
higher  rates  than  would  otherwise  be  possible.  Our  findings  are 
in  harmony  with  others  who  have  worked  in  this  field. 

In  order  to  measure  the  maximum  information  transmission 
capacity  of  a  nerve  fiber  which  employs  pulse-interval  coding, 
we  must  be  able  to  stimulate  the  neuron  with  an  input  source 
which  can  deliver  trains  of  two  or  more  pulses  at  diff^erent  intervals. 
This  follows  from  information  theory,  since  in  an  evenly  spaced 
pulse  train  there  is  no  uncertainty  about  the  time  of  arrival  of  the 
next  pulse,  and  hence  no  information.  Maximum  uncertainty  is 
available  only  in  a  random  source  in  which  the  pulse  is  equally 
likely  to  occur  at  any  time.  It  is  also  necessary  to  determine  the 
minimum  interpulse  interval  which  can  be  discriminated  by  the 
neural  system.  We  can  determine  this  by  measuring  the  standard 
deviation  of  the  latent  period  in  a  fiber,  or  its  "jitter." 

We  proceeded,  therefore,  to  use  an  electronic  timer,  accurate 
to  1  microsecond,  to  measure  the  jitter  of  single  sciatic  nerve 
fibers  of  the  frog,  studying  variation  as  tiie  time  between  two  ad- 
jacent pulses  was  reduced.  This  turned  out  to  be  of  the  order  of 
2-5  microseconds,  and  adding  a  third  pulse  before  the  other  two 
did  not  aff^ect  this  value.  Using  a  mathematical  model  developed 
by  Rapoport  and  Horvath  (16),  we  were  able  to  calculate  the 
curve  of  maximum  channel  capacity  of  such  a  neuron  at  various 
input  rates.  We  found  that  the  output  increased  as  a  function  of 
the  input  up  to  4,000  bits  per  second  (an  astonishingly  high 
capacity  for  such  a  small  system — assuming  optimal  pulse-interval 
coding);  then  leveled  off"  and  decreased,  thereafter,  as  the  input 
rate  increased.  This  performance  curve  is  shown  in  Figure  1. 

As  for  adjustment  processes,  the  skipping  of  pulses  which  we 
found  at  the  higher  input  rates  was,  of  course,  omission.  The 
lower  output  intensities  could  be  called  erroneous  processing  if 
they  were  not  intense  enough  to  cross  the  threshold  of  the  neuron 
on  the  other  side  of  the  synapse.  That  threshold,  incidentally, 
can  be  considered  a  sort  of  filtering.  For  other  neuronal  adjustment 
processes  to  information  overload,  we  have  no  evidence. 

Organ  Research 

We  used  the  same  electronic  timer  to  stimulate  the  optic  nerve 
of  the  white  rat,  recording  the  output  from  a  macroelectrode  on 


The  Individual  as  an  Information  Processing  System 


315 


O      T3 


< 

X 

o 


Fig.  1. 
mation 


0  5         10        15        20       25       30       35       40       45       50 

AVERAGE  INPUT    RATE   PULSES   PER    SECOND   X   10^ 

The  channel  capacity  of  a  model  neuron  calculated  by  continuous  infor- 
theory  for  a  Gaussian  noise  distribution  with  a  standard  deviation,  <x, 
or  jitter  =  5  ^  sec.  Refractory  period  taken  to  be  1  msec. 


^  00        200         300        400        500        600        700        800         900       1000 

2  AVERAGE  STIMULATION  RATE  IN  PULSES    PER    SECOND  (  Poisson  Input) 

Fig.  2.  The  channel  capacity  of  a  model  organ  system  calculated  by  continuous 
information  theory  for  a  Gaussian  noise  distribution  with  a  standard  deviation,  cr, 
or  jitter  =  1  msec.  Refractory  period  taken  to  be  50  msec. 


the  optic  cortex.  Similar  calculations  from  this  experiment  gave 
us  a  curve  of  comparable  shape.  The  channel  capacity,  however, 
was  of  the  order  of  fifty  bits  per  second  (Fig.  2). 


316 


Information  Storage  and  Neural  Control 


lllllllll 
■  llllltii 

■  I  III  111  I* 
■■lillllll 

f  ii!!:iiif 


Fig.  3.  Subject  at  IOTA  Apparatus. 


The  same  adjustment  processes  appear  at  the  organ  level, 
except  that  multiple  channels,  of  course,  are  also  used. 

Individual  Research 

For  this  study  we  designed  and  built  an  IOTA  (Information 
Overload  Testing  Aid)  apparatus  (Fig.  3). 

This  is  a  piece  of  equipment  by  which  stimuli  are  presented 
to  a  subject  on  a  transparent  ground-glass  screen,  about  3x4  feet 
in  size.  The  apparatus  is  placed  on  a  table  in  front  of  the  subject, 
who  responds  by  pushing  appropriate  buttons  arrayed  before  him. 
Stimuli  are  thrown  on  the  back  of  the  screen  by  a  Perceptoscope, 
which  is  a  projector  capable  of  showing  movie  film  at  rates  of 
from  one  to  twenty-four  frames  per  second.  The  film  contains  a 
program  which  presents  black  arrows  on  a  white  background, 
which  can  appear  in  from  one  to  eiglit  of  the  eight  two-inch  wide 
vertical  slots  which  run  down  the  screen.  Arrows  can  assume  any 
one  of  eight  angular  positions,  like  clock  hands.  Before  the  subject 


The  Individual  as  an  Information  Processing  System  31  7 

is  a  set  of  eight  buttons  for  each  of  the  slots  being  used.  Since  he 
can  see  stimuH  in  a  maximum  of  four  slots  at  once,  altogether  he 
has  thirty-two  buttons,  four  sets  of  eight  buttons  each.  If  an  arrow 
in  Position  B  appears  in  Slot  3,  the  correct  response  is  to  push 
Button  B  of  the  set  for  Slot  3.  Any  other  response  is  an  error. 
If  the  subject  pushes  none,  that  is  an  omission. 

Queuing  is  also  possible.  The  subject  has  a  foot  pedal  with 
which  he  can  lower  or  raise  opaque  strips  behind  each  of  the  slots. 
At  the  beginning  of  each  test,  only  the  top  square  in  each  of  the 
slots  being  used  is  open  so  that  light  can  come  through.  If  the 
subject  pushes  the  pedal,  he  can  move  the  opaque  strips  to  open 
as  many  as  eleven  more  squares,  a  maximum  of  twelve;  or  by  push- 
ing the  pedal  in  the  other  direction  he  can  close  these  up  again,  as 
he  wishes.  The  moving  picture  film  is  made  so  that  if  an  arrow 
appears  in  Position  B  in  Slot  3  in  Frame  1  of  the  film,  it  goes  to 
the  next  lower  position  in  that  slot  in  Frame  2,  and  to  the  next 
lower  position  in  Frame  3,  until  having  gone  through  all  twelve 
positions,  it  finally  disappears  from  the  screen.  In  the  meantime 
other  stimuli  may  be  appearing  higher  in  the  same  slot,  or  in 
other  slots.  Therefore,  when  the  subject  pushes  his  queuing  pedal 
he  gives  himself  more  time  to  respond  to  the  stimulus  before  it 
disappears.  He  can  filter  by  paying  attention  only  to  the  arrows 
pointing  up,  or  to  those  pointing  to  the  left,  rather  than  to  those 
pointing  to  all  eight  positions.  He  can  approximate  by  pushing 
all  four  left  buttons  in  Slot  3,  if  he  is  not  certain  in  which  of  the 
four  left  directions  the  arrow  pointed,  but  knows  it  pointed  toward 
the  left;  or  by  pushing  all  eight  buttons  for  Slot  3  if  he  simply 
saw  an  arrow  but  has  no  idea  of  its  direction.  On  occasion,  he 
can  use  multiple  channels  by  working  with  both  hands  at  the 
same  time.  Finally,  escape  is  possible,  if  he  gives  up  and  refuses  to 
continue  the  task.  So  all  the  mechanisms  of  adjustment  that  we 
have  mentioned  are  possible  on  the  IOTA. 

This  apparatus  can  increase  the  amount  of  information  per 
second  in  several  ways:  1)  by  increasing  the  ensemble,  or  the 
number  of  alternate  positions  for  the  arrows  from  two  (1  bit)  to 
eight  (3  bits);  2)  by  speeding  the  movie;  3)  by  increasing  the 
range,  or  raising  the  number  of  slots  used  simultaneously;  or  4)  by 
altering  the  degree  of  regularity  or  randomness  of  the  presentations. 


Information  Storage  and  Neural  Control 


INPUT     RATE      (BITS/SEC) 
Fig.  4.  Performance  curves  for  Subjects  A  and  B 


Two  male  college  students  were  used  as  subjects  in  this  study. 
Before  being  tested,  they  were  thoroughly  trained  in  the  procedure, 
including  the  use  of  all  the  adjustment  processes. 

The  button-pushing  performance  of  each  subject  was  recorded 
on  a  kymograph  and  was  compared  with  the  program  of  stimuli 
presented.  These  raw  data  were  fed  into  a  computer  programmed 
to  calculate  the  input  (stimulus  presentation)  and  output  (subject's 
response)  rates  in  bits  of  information  per  second,  using  the  Shannon 
information  statistic. 

Data  obtained  with  this  equipment  (Fig.  4)  produce  a  curv^e  of 
the  same  general  shape  as  do  data  at  the  level  of  cell  and  organ  when 
input  in  bits  per  second  is  plotted  against  output  in  bits  per  second. 
Within  the  range  tested,  there  is  some  question  as  to  whether  the 
output  falls  below  channel  capacity  at  high  input  rates,  or  whether 
some  cut-off  mechanism  prevents  this  from  happening.  It  is  prob- 


77?^'  Individual  as  an  Information  Processing  System 


319 


INPUT     RATE      (BITS/SEC) 
Fig.  5.  Average  utilization  of  adjustment  processes  by  both  subjects  at  various 

input  rates. 


able,  however,  that  uhimately,  at  very  high  input  rates,  output 
does  falL  The  maximum  channel  capacity  for  an  individual 
operating  the  IOTA  was  determined  to  be  about  six  bits  per 
second.  Other  experimenters  have  found  maximum  channel 
capacities  for  random  material  up  to  about  thirty  bits  per  second. 
With  the  IOTA,  however,  output  rate  is  limited  by  partial  in- 
compatibility between  the  nature  of  the  stimulus  and  the  organiza- 
tion of  the  response  mechanism,  by  the  difficulty  of  making  the 
response,  and  by  many  other  factors. 

Our  subjects  were  trained  regarding  the  possible  mechanisms 
of  adjustment  available  to  them,  and  were  free  to  select  them  as 
they  saw  fit.  They  used  few  or  none  of  the  mechanisms  at  slow 
rates  of  transmission.  They  tended  to  attempt  them  all  at  medium 
rates.  At  higher  rates,  under  our  experimental  conditions,  the 
subjects  showed  preference  for  filtering  and,  particularly,  for 
omission  (Fig.  5).  Whether  this  preference  is  genetically  deter- 
mined or  learned,  we  do  not  know. 


320 


Information  Storage  and  Neural  Control 


3.0 


-     2.5 
<_) 

CO 


m 


< 
ir 


2.0 


1.5 


z)     0.5 
O 


O— O    GROUP    A 
•— •    GROUP   B 


I  2  3  4  5  6  7 

INPUT   RATE  (BITS/SEC.) 

Fig.  6.  Performance  curves  for  Groups  A  and  B. 


8 


Group  Research 

We  also  used  the  IOTA  apparatus  with  two  four-man  groups. 
The  procedure  was  as  follows:  Three  members  of  the  group,  A, 
B,  and  D,  face  the  screen.  A  calls  out  the  number  of  the  slot  in 
which  an  arrow  appears,  and  B  calls  out  a  letter  representing  the 
position.  C,  who  is  facing  the  buttons,  but  whose  back  is  turned 
to  the  screen,  then  pushes  the  button  indicated  by  the  information 
he  got  from  A  and  B.  When  C  pushes  a  button  a  small  red  light 
appears  over  one  of  the  slots,  indicating  which  button  he  pushed. 
If  his  push  is  correct,  D  says  nothing.  If  the  push  is  incorrect,  D 
corrects  C  and  C  tries  to  push  the  right  button.  The  performance 
curves  from  our  pretest  runs  with  two  groups  have  the  same 
general  appearance  as  the  performance  curves  of  the  individual 
subjects,  though  at  lower  channel  capacities — about  3  bits  per 
second   (Fig.   6). 

This  is,  of  course,  a  very  specialized  sort  of  small  group  in  which 
roles  are  strictly  differentiated.  Only  some  members  can  receive 
sensory  inputs,  while  others  make  responses  or  perform  other 
tasks.  There  are  structural  similarities  to  certain  lole-diflfeientiated 
groups  in  military  life,  like  tank  crews,  bomber  crews,  or  sub- 


The  Individual  as  an  Information  Processing  System 


321 


8.0 
7.5 

7.0 


ADJUSTMENT  PROCESSES 

• •  OMISSIONS 

o o  ERRORS 

* «  FILTERING 

* *  APPROXIMATION 


I  2  34  56  789         10         II 

INPUT     RATE       (BITS/SEC) 

Fig.  7.  Average  utilization  of  adjustment  processes  by  both  groups  at  various 

input  rates. 


marine  crews.  We  recognize  this  as  only  one  type  of  group,  just 
as  we  used  only  one  type  of  individual  in  our  individual-level  study. 
The  use  of  adjustment  processes  by  groups  was  comparable  to 
their  use  by  individuals,  although  queuing  was  not  employed.  The 
reason  for  this  is  not  clear.  Figure  7  presents  these  findings. 

Social  Institution  Research 

My  colleagues   and    I    have   conducted    two   investigations   on 
informational   overload    of  social   organizations.    In   one   project, 


\77 


Information  Storage  and  Neural  Control 


much  of  which  was  carried  out  by  Jay  and  McCornick,  we  studied 
a  simulator  of  the  air  raid  warning  system  of  the  United  States 
and  Canada.  The  simulator,  at  System  Development  Corporation 
in  California,  which  cooperated  in  the  study,  consisted  of  three 
groups  of  three  men,  each  in  three  separate  rooms.  Because  of 
these  three  echelons,  and  because  all  members  were  not  face  to 
face,  we  called  this  an  institution  rather  than  a  group,  though 
it  was  a  very  small  unit.  The  first  room  simulated  a  radar  station 
in  the  air  raid  warning  network;  the  second  room  simulated  the 
room  at  headquarters  in  which  the  message  from  the  local  station 
was  received;  and  the  third  room  simulated  a  plotting  board  on 
which  the  location  of  planes  was  indicated  at  headquarters.  Dots, 
presumably  representing  airplanes  in  geographical  sectors,  ap- 
peared randomly  on  a  21x21  matrix  (Fig.  8).  These  dots,  each 
with  an  associated  message  number,  were  thrown  on  a  board  by 


I' 


•iff!'!1WRfT.^'!n«!Pf?!fS"!l'flR  '32SS5-S 


Fig.  8.  Display  board  for  Air  Raid  Warning  Simulator. 


The  Individual  as  an  hiformation  Processing  System 


323 


5.0 

4.0h 

OUTPUT 
RATE         30 
IN   BITS 

PER  2.0 

SECOND 

I.Oh 


2  3         4  5  6  7  8 

INPUT   RATE    IN   BITS    PER    SECOND 

Fig.  9.  Average  performance  curves  for  teams  in  Social  Institution  Experiment. 


a  movie  projector.  Each  of  three  readers  in  the  first  room  was 
responsible  for  one-third  of  the  total  board.  When  a  dot  and  its 
number  appeared  in  his  sector,  the  appropriate  reader  wrote  down 
on  a  card  the  coordinates  of  the  cell  in  which  the  dot  appeared 
and  also  the  number  of  the  dot.  He  then  presented  the  caid  to 
his  corresponding  teller  in  the  next  room  by  passing  it  through  a 
slot.  The  teller  in  turn  read  the  card  by  telephone  to  the  cor- 
responding plotter  in  the  third  room,  who  wrote  the  message 
number  in  the  proper  cell  on  the  plotting  board.  This  board  was 
photographed  automatically  at  6-second  intervals,  so  that  a  con- 
tinuous record  of  the  appearance  of  numbers  on  the  plotting  board 
could  be  obtained.  Thus,  there  were  three  entirely  separate  chan- 
nels in  this  system,  since  Reader  A  always  gave  his  information 
only  to  Teller  A,  who  passed  the  message  only  to  Plotter  A,  and 
so  on  for  Team  2  and  Team  3.  The  performance  curves  for  these 
teams  had  shapes  similar  to  those  curves  obtained  at  the  individual 
and  group  levels  when  input  in  bits  per  second  was  plotted  against 
output  in  bits  per  second.  Maximum  channel  capacity  was  about 
four  bits  per  second,  approximately  in  the  range  of  the  group, 
probably  because  information  passed  through  about  as  many  com- 
ponents as  in  the  group,  rather  than  through  more,  as  would 
have  been  the  case  in  a  larger  social  institution  (Fig.  9).  These 
subjects  also  had  much  more  practice  than  those  in  our  group 


324 


Information  Storage  and  Neural  Control 


% 

100 

90- 
80- 
70 


UTILIZATION 

OF 

ADJUSTMENT     ^° 

PROCESSES 


40+18 


• •  OMISSIONS 

O— O  ERRORS 

A A  QUEUING 

>> K  FILTERING 


AVERAGE 
SECONDS   OF 
QUEUING 

50  +  21 


123456789 
INPUT    RATE    IN   BITS    PER    SECOND 
Fig.  10.  Average  utilization  of  adjustment  processes  by  teams  in  Social  Institution 

Experiment. 


research  and  had  greatly  improved  their  transinission  rates  since 
their  earher  trials. 

Four  adjustment  processes  were  used  by  the  teams  in  these 
studies — omission,  error,  queuing,  and  filtering.  The  experimental 
instructions  prevented  use  of  approximation  and  multiple  channels. 
Utilization  of  all  adjustment  processes  was  measured  in  percent- 
ages, except  for  queuing,  which  was  measured  in  average  number 
of  seconds  of  delay  (Fig.  10). 

An  associated  study  directed  by  Meier  (17)  dealt  with  the 
effects  of  overloads  of  demands  upon  the  Undergraduate  Library 
of  The  University  of  Michigan  at  periods  of  peak  use.  The  inflow 
of  students  and  faculty  into  this  library,  each  person  with  special 
needs,  is  not  an  overload  of  energy  or  matter,  for  the  library  is 
never  actually  physically  unable  to  hold  them.  The  demands  upon 
members  of  the  library  staff  for  service,  however,  can  constitute 
what  is  essentially  an  information  overload. 

Participant  observation  and  other  operations  research  procedures 
were  employed  to  find  how  much  the  library  was  used  at  top  load 
periods  and  what  changes  occurred  in  library  functions  at  such 
times.  Since  no  significant  difference  in  average  time  of  getting 


Thf  Indii'idiial  as  an  Information  Pmcessing  System  325 

a  book  was  found  between  periods  of  light  and  of  heavy  use,  the 
library  may  not  have  been  under  real  performance  overload  at 
any  time.  Rough  efforts  were  made  to  calculate  the  number  of 
bits  of  information  flowing  through  the  library.  It  was  determined 
that  the  average  book  title  in  the  card  catalog  contains  about 
135  bits  of  information,  and  that  the  average  reader  processes 
between  50,000  and  90,000  bits  per  hour  of  reading. 

Perhaps  the  most  significant  finding  by  Meier  and  his  colleagues 
was  that  a  series  of  adjustment  processes  occurred,  or  could  occur, 
in  the  library  to  cope  with  the  overload.  He  recognizes  the  simi- 
larity of  his  list  to  the  one  presented  earlier  in  this  chapter.  How- 
ever, he  found  more  complex  forms  of  these  adjustment  processes, 
or  "policies"  as  he  calls  them,  in  this  complicated  social  institution 
with  its  many  subsystems  carrying  out  numerous  activities.  His 
list  follows:  Queuing;  priorities  in  cjueues  and  backlogs;  destruction 
of  low  priority  inputs  (filtering);  omission;  reduction  of  processing 
standards  (approximation) ;  decentralization  (a  special  case  of  use 
of  multiple  channels) ;  formation  of  independent  organizations  near 
the  periphery  (multiple  channels);  mobile  reserve  (multiple  chan- 
nels); rethinking  procedures;  redefinition  of  boundaries  of  the 
system;  escape;  retreat  to  formal,  ritualistic  behavior;  and  dis- 
solution of  the  system  with  salvage  of  its  assets.  Whether  there 
are  new  adjustment  processes  here,  or  simply  special  cases  of  those 
we  have  listed  is  a  question  for  debate;  but  that  such  adjustment 
policies  are  used,  there  can  be  no  question. 

Summary  of  Our  Research 

For  five  levels  of  organization,  or  systems,  viewed  as  information 
processing  channels,  the  following  propositions  appear  to  have 
support: 

a)  When  information  input  in  bits  per  second  is  increased,  the 
output  at  first  follows  the  input  more  or  less  as  a  linear  function, 
then  levels  off  at  a  channel  capacity,  and  finally  falls  off"  toward 
zero.  We  have  yet  to  deteimine  whether  the  larger  systems  have 
a  cut-off  mechanism  which  prevents  the  final  fall  in  output. 
Though  such  a  mechanism  may  delay  this  fall,  the  weight  of 
evidence  suggests  that  it  must  finally  occur. 


326  Information  Storage  and  Neural  Control 

This  decrease  of  information  output  rate  in  living  systems  is 
not  the  result  of  destruction  of  the  system  by  an  overload  of  the 
energy  which  conveys  the  information  because  1)  the  process  is 
reversible — decrease  of  input  rate  immediately  raising  output  rate 
back  to  channel  capacity,  and  2)  final  irreversible  change  of  such 
systems  by  energy  input  undoubtedly  occurs  when  the  energy  is 
orders  of  magnitude  greater  than  that  involved  in  informational 
overload. 

b)  There  is  a  hierarchical,  cross-level  difference  in  maximum 
channel  capacity.  Assuming  pulse-interval  coding,  we  found  this 
to  be  of  the  order  of  4,000  bits  per  second  for  neurons  in  the 
frog  sciatic  nerve,  and  about  fifty  bits  per  second  for  a  single 
channel  in  the  visual  nervous  system  of  the  rat.  It  was  six  bits 
per  second  for  the  individual,  three  bits  per  second  for  a  single- 
channel  group,  and  three  to  four  bits  per  second  per  channel  in 
a  small  social  institution  with  about  the  same  number  of  com- 
ponents in  each  channel  as  there  were  in  the  group. 

Apparently  the  more  components  there  are  in  an  information 
processing  system,  the  lower  is  its  channel  capacity.  There  are 
several  reasons  for  this.  Two  of  the  most  obvious  are  that  recoding 
of  information  is  necessary  at  the  border  between  each  component 
and  the  next,  and  that  such  recoding  always  results  in  loss  of  a 
certain  amount  of  information.  Moreover,  if  there  are  n  com- 
ponents in  any  system,  one  must  have  a  lower  channel  capacity 
than  the  others,  and  the  statistical  probability  of  there  being  such 
a  slow  component  is  always  greater  as  n  increases.  This  sluggish 
component  constitutes  a  bottleneck,  since  no  channel  is  faster  than 
its  slowest  component. 

c)  Several  of  the  adjustment  processes  are  used  by  all  of  these 
systems,  the  use  increasing  as  input  rate  rises. 

d)  Fewer  adjustment  processes  seem  to  be  available  to  the 
systems  at  the  lower  levels.  Those  employed  at  the  higher  levels 
appear  to  be  more  complex  as  well  as  more  numerous,  although 
their  fundamental  similarity  to  the  lower  level  processes  is  clear. 

Of  course  the  findings  for  other  types  of  systems  at  each  of  the 
levels  might  be  difTerent  in  significant  ways  from  our  findings  in 
the  particular  systems  we  chose  to  study.  The  goal  of  these  projects 


The  Individual  as  an  Information  Processing  Sy stein  327 

was  to  determine  whether  a  cross-level  formal  identity  could  be 
confirmed  for  any  examples  of  systems  at  different  levels. 

It  is  apparent  that  interesting  insights  arise  when  not  only 
individuals,  but  all  living  organisms  and  organizations  are  viewed 
as  information  processing  systems. 

REFERENCES 

1.  Barlow,  H.  B.:  Sensory  mechanisms,  the  reduction  of  redundancy 

and  intelligence.  In:  Mechanisation  of  Thought  Processes,  Proceedings 
of  Symposium  at  National  Physical  Laboratory,  Teddington,  Eng- 
land. London:  Her  Majesty's  Stationery  Office,  1959,  p.  542. 

2.  Gerard,   R.   VV.:   Organism,   society  and  science.   Sci.   Monthly,  50: 

340-350,  403-412,  530-535,  1940. 

3.  Piatt,  J.  R.:  Amplification  aspects  of  biological  response  and  mental 

activity.  Arner.  Sci.,  44.- 180-1 97,  1956. 

4.  Swets,  J.  A.,  Tanner,  W.  P.,  Jr.,  and  Birdsafi,  T.  G.:  The  evidence 

for  a  decision-making  theory  of  visual  detection.  Technical  Report 
JVo.  40,  Electronic  Defense  Group,  LIniversity  of  Michigan,  Ann 
Arbor,  April,  1955. 

5.  Stevens,    S.    S.:    Cross-modality  validation   of  suljjective   scales  for 

loudness,  vibration,  and  electric  shock.  J.  Exp.  Psychol.,  57:201- 
209,    1959. 

6.  Broadbent,  D.  E.:  Perception  and  Communication.  New  York,  Pergamon 

Press,  1958,  p.  297. 

7.  Quastler,  H.:  In  Human  Performance  in  Information  Transmission.  Con- 

trol Systems  Laboratory,  Report  No.  R-62.  Urbana,  University 
of  Illinois,   1955. 

8.  Broadbent:  op.  cit.,  p.  5. 

9.  Quastler:  op.  cit. 

10.  Quastler:  ibid.,  p.  62. 

11.  Quastler:  ibid.,  pp.  62-63. 

12.  Barlow,  H.  B.:  Increment  thresholds  at  low  intensities  considered  as 

signal  noise  discriminations.  J.  Physiol.  (London),  141  :?>'il-?)SO, 
1958. 

13.  Deininger,  R.  L.  and  Fitts,  P.  M.:  Stimulus-response  compatibility, 

information  theory,  and  perceptual-motor  performance.  In,  H. 
Quastler  (Ed.),  Information  Theory  in  Psjchology.  Glencoe,  The  Free 
Press,  1955. 

14.  Brown,  R.  and  Lenneberg,  E.:  A  study  in  language  and  cognition 

J.  Abnorm.  Soc.  Psychol.,  ^P.-454-462,  1954. 


328  Information  Storage  and  Neural  Control 

15.  Mowbray,  G.  H.  and  Rhoades,  M.  V.:  On  the  reduction  of  choice 

reaction  times  with  practice.  Qiiart.  J.  Exp.  Psychol.,  7 7;  16-22,  1959. 

16.  Rapoport,  A.  and  Horvath,  W.  J.:  The  theoretical  channel  capacity 

of  a    single    neuron   as   determined    by   various   coding   systems. 
Inform,  and  Control,  3:335-350,  1960. 

17.  Meier,  R.  L.:  Social  change  in  communications  oriented  institutions. 

Mental  Health  Research  Institute,  Preprint  No.  10,  March.  1961. 


CHAPTER 
XIV 

INFORMATION  PROCESSING  IN  THE 
TIME  DOMAIN 

Neil  R.  Burch,  M.D.  and  Harold  E.  Childers 


T« 


HIS  paper  briefly  outlines  the  work  we  are  conducting  in 
the  Department  of  Psychiatry,  Baylor  University  College  of  Medi- 
cine, and  in  the  laboratories  of  the  Houston  State  Psychiatric 
Institute.  The  basis  for  this  research  is  the  theory  that  a  special 
case  of  analysis  in  the  time  domain  has  something  to  offer  both  in 
time  resolution  and  in  economy  of  information  processing  that 
cannot  be  readily  obtained  from  frequency  analysis  or  from  more 
conventional  time  sampling  procedures.  The  analytical  process  to 
be  described  we  have  called  period  analysis  (1). 

Given  an  amplitude  function  distributed  in  tiine,  there  are  a 
limited  number  of  questions  that  may  be  asked  of  the  function 
to  yield  an  analysis  or  to  undertake  data  reduction.  Consider  the 
following  four  cases:  1)  One  inay  focus  on  the  amplitude  and  ask 
the  question  "how  much"  over  a  time,  T;  one  may  focus  on  both 
time  and  amplitude  and  ask  the  question  "how  much"  at  par- 
ticular points  in  time,  either  2)  points  at  fixed  intervals  or  3)  points 
related  to  an  event;  finally,  4)  one  may  focus  on  selected  events 
and  ask  the  question  "when." 

A  theorem  in  information  theory  tells  us  that  if  we  take  this 
amplitude  distribution  in  time  and  sample  it  every  so  often,  we 
will  retain  complete  information  about  the  signal.  Presented  more 
formally,  the  theorem  reads,  "If  a  function  G{t)  contains  no  fre- 
quencies higher  than  W  cycles  per  second,  it  is  completely  deter- 
mined by  giving  its  ordinatcs  at  a  series  of  points  spaced  jrr 
seconds  apart,  the  series  extending  throughout  the  time  domain" 
(2)   (Case  2).  A  corresponding  theorem  for  sampling  in  the  frc- 

329 


330  Information  Storage  and  Neural  Control 

quency  domain*  requires  exactly  the  same  number  of  sampling 
points  plus  one,  or: 

2TW  +  1 

where  T  is  the  duration  of  a  signal,  W  is  the  spectral  band  width, 
and  the  sampling  points  are  spaced  at  fixed  intervals. 

If  we  can  say  in  the  electroencephalographic  (EEG)  signal,  as 
an  example,  that  the  highest  frequency  which  carries  neuro- 
physiological  information  is  100  cycles  per  second,  we  know  that 
we  must  sample  at  least  200  times  a  second — perhaps  much  more 
often  if  there  is  considerable  noise  in  the  system — in  order  to 
retain  all  of  the  information.  While  we  are  now  satisfied  that  in 
both  the  time  and  frequency  domain  we  may  ask  of  the  EEG 
signal  "how  much"  at  200  points  per  second,  we  have  also  posed 
ourselves  a  massive  problem  in  data  handling.  Further,  we  have 
not  learned  anything  of  the  optimum  analytic  procedure,  since 
to  retain  all  information  in  the  original  signal  "defeats  the  very 
purpose  of  analysis,  which  is  to  abstract  and  emphasize  only 
significant  changes."   (4). 

There  remain  the  case  of  coding  "how  much"  related  to  a 
selected  event  and  the  even  simpler  case  of  asking  only  "when" 
the  event  occurs.  In  both  of  these  remaining  cases,  the  coding 
event  must  be  defined  and  the  assumption  made  that  all  200  points 
per  second  in  the  EEG  do  not  contain  the  same  amount  of  infor- 
mation. Let  us  for  a  moment  suppose  that  some  of  these  points 
contain  ten  times  the  information  of  other  points.  Then  we  may 
drop  the  low  information  points,  retain  the  high  information 
points  and  sacrifice  a  unique  characterization  of  the  wave  for  a 
good  approximation.  Such  a  process  would  be  highly  economical 
in  terms  of  handling  the  data. 

The  critical  problem  is,  of  course,  the  generation  of  the  coding 
event  which  acts  as  a  "metasignal"  in  the  sense  in  which  Gregory 
Bateson  used  the  term  for  us  earlier.  In  the  first  paper  of  this 
symposium,  Bernard  Saltzberg  introduced  you  to  Maxwell's 
demon.  I  would  like  to  propose  another  hypothetical  information 
demon,  one  that  might  look  at  each  of  our  points  and  say,  "We'll 


*If  fi(w)  represents  the  spectrum  of  a  function  G(t),  which  is  zero  everywhere 
except  in  the  range  Tj  <  t  <  T2,  then  i2(w)  is  exactly  determined  for  all  values  of 
w  by  giving  its  values  at  a  series  of  points  ^  / {Ti  —  T2)  cycles  per  second  apart  in  fre- 
quency, the  series  extending  throughout  the  frequency  domain.  (3)  (Case  1). 


Information  Processing  in  the  Time  Domain  331 

take  those,"  or  "No,  that's  low  information,  drop  that."  If  we 
speculate  that  our  demon  is  extremely  conservative  and  expects 
the  signal  to  be  linear  as  a  function  of  time,  a  straight  line  deter- 
mined by  two  or  more  points,  then  any  point  that  agrees  with  this 
assumption  is  a  low  information  point.  The  demon  has  predicted 
that  the  signal  will  not  change  from  positive  to  negative  values, 
will  not  change  its  sense  of  positive-negative  direction,  will  not 
even  change  its  sense  of  curvature.  The  high  information  points 
now  become  zero  points,  minimax  points  and  points  of  inflection 
in  the  primary  signal.  The  coding  points  generated  are  at  the 
baseline  cross  of  the  primary  signal,  of  its  first  derivative  and  of 
its  second  derivative.  We  might  have  a  second  type  of  demon, 
a  neurophysiological  demon,  that  can  tell  us  when  a  significant 
neurophysiological  event  is  reflected  in  the  signal.  This  demon 
identifies  our  semantic  information  as  contrasted  to  statistical 
information.  It  would  be  nice,  of  course,  if  both  these  demons 
were  the  same.  In  order  to  "twin"  our  two  friends,  we  would  be 
forced  to  assume  that  the  brain  sees  change  and  rate  of  change  of 
the  electrical  potentials  in  its  subpopulations  as  highly  significant  in- 
formation. We  would  also  conclude  that  the  wave  shape  of  our  EEG 
signal  is  rich  in  semantic  information  as  compared  to  characteri- 
zations in  the  frequency  domain  such  as  frequency  or  power  spectra. 

Defining  the  coding  points  for  amplitude  sampling  as  the  baseline 
cross  of  the  primary  and  its  first  and  second  deri\^atives  allows  us 
to  take  discrete  data  in  a  definite  but  not  uniformly  spaced  pattern 
(Case  3).  This,  on  the  average,  should  result  in  fewer  sampling 
points  than  the  folding,  or  Nyquist,  frequency  requirement  (5) 
discussed  previously.  Period  analysis  is  a  further  simplification  of 
this  general  process  in  that  the  amplitude  of  the  function  is  not 
sampled  at  all.  The  theoretical  justification  for  this  approach  has 
been  developed  in  terms  of  the  Gram-Charlier  series  (6).  The 
remainder  of  this  paper  will  explore  period  analysis  as  a  special 
case  (Case  4)  of  information  processing  in  the  time  domain.  The 
questions  to  be  asked  concern  retention  of  both  statistical  and 
semantic  information  during  period  analysis  of  several  bio- 
electronic  signals. 

Figure  1  illustrates  the  characteristics  of  the  first  and  second 
derivatives.  The  function  /(.v)  in  the  upper  right  hand  corner  of 
the   figure   represents   an   evoked    potential   which   feeds   into   a 


332  Information  Storage  and  Neural  Control 

DERIVATIVE         CHARACTERISTICS 


"to  56  66 

TfMSSSKt  (CTCIXB  m  SBXMD) 


Fig.  1.  Electronic  Parameters  of  the  Mathematical  Derivative.  The  90°  phase  shift  and 
linear  doubhng  of  ampHtude  per  octave  is  illustrated  as  the  electronic  definition 
of  a  first  derivative.  The  sharply  increasing  amplitude  of  the  second  derivative 
with  increase  in  frequency  emphasizes  the  accentuation  of  high  frequency  com- 
ponents. The  three  functions  on  the  right  of  the  figure  graphically  illustrate  the 
eff'ect  of  derivative  processing. 

differentiating  network  to  yield  the  first  derivative, /  (v) .  The  first 
derivative,  through  an  identical  differentiating  network,  gives  the 
first  derivative  of  the  first  derivative,  or  second  derivative,  /"(v), 
of  the  primary  evoked  potential.  These  functions,  after  Lorente 
de  No  (7),  illustrate  the  external  action  potential  of  bullfrog 
alpha  fibers  and  its  first  two  derivatives.  It  is  clear  that  the  high 
frequency  components  of  the  primary  evoked  potential  are  greatly 
accentuated  by  double  differentiation.  The  electronic  definition 
of  a  derivative  is  the  same  as  the  mathematical  definition  except 
that  it  is  couched  in  different  parameters.  The  phase  shift  required 
in  a  sine  wave  is  90°  for  the  first  derivative  and  180°  for  the  second 
derivative.  The  important  parameter  for  our  purpose  is  the 
amplitude  characteristic  as  illustrated  in  Figure  1.  Given  a  mixed 
sine  wave  made  up  of  equal  amplitude  twenty  cycle  per  second 
and  forty  cycle  per  second  components,  the  first  derivative  will 
yield  twice  as  much  amplitude  for  the  forty  cycle  per  second 
component  because  it  is  twice  the  frequency  of  the  twenty  cycle 


Information  Processing  in  the  Time  Domain 


333 


per  second  component.  This  linear  relationship  holds  throughout 
the  band  pass  range.  The  second  derivative  multiplies  the  forty 
cycle  component  by  a  factor  of  4,  the  eighty  cycle  component  by 
a  factor  of  16,  etc.,  in  this  example.  Figure  2  illustrates  this  deriva- 
tive processing  as  it  is  applied  to  the  electroencephalogram.  The 
faster  frequency  components  present  in  a  complex  primary  wave 
become  full-fledged  baseline  crosses  because  of  the  relative  accentu- 
ation of  the  faster  frequencies.  Period  analysis  proceeds  by  generat- 
ing square  waves  at  the  baseline  cross  of  the  primary,  the  first 
derivative  and  the  second  derivative.  As  can  be  seen  in  Figure  2, 
the  square  wave  train  designated  as  major  period  reflects  the  domi- 


Fig.  2.  Pulse  JVidth  Conversion:  EEC.  The  process  of  period  analysis  applied  to 
the  left  parieto-occipital  electroencephalogram.  The  60  cycle  per  second  artifact 
superimposed  on  the  original  primary  trace  is  markedly  reduced  by  the  rejection 
notch  of  the  selective  frequency  amplifier,  as  seen  in  the  filtered  primary.  The 
"fragmented"  appearance  of  the  second  derivative  minor  period  results  from  the 
high  inertia  pen  system  which  cannot  foUow^  a  true  square  wave  at  these  fre- 
quencies. Paper  speed  60  millimeters  per  second. 


334 


Information  Storage  and  Neural  Control 


nant  rhythm  of  the  analog  primary.  The  second  derivative  square 
wave  train,  referred  to  as  the  minor  period,  carries  information 
reflecting  superimposed  fast  activity,  desynchrony,  and  waveshape. 
It  is  of  particular  importance  to  know  how  much  wave  shape 
information  is  retained  or  lost  in  this  processing,  because  it  is 
probably  the  wave  shape  which  triggers  recognition  in  the  human 
computer  in  clinical  electroencephalography.  We  propose  that 
much  of  the  wave  shape  information  is  retained  in  the  three  square 
wave  trains  as  they  relate  to  one  another  in  time,  as  we  have 
attempted  to  illustrate  in  Figure  3.  The  top  trace  is  a  synthetic 
function  made  up  of  a  "dominant"  nine  and  one-half  cycle  per 
second  sine  wave  mixed  with  a  lower  amplitude  "superimposed" 


PERIOD    RECONSTITUTION 
SYNTHETIC  FUNCTION 


\nj^j\r'iS\j\s^v^'\rvriP^^ 


/Wu^^ 


mmmmmm 


\\m\mw 


nfunnn^n, 


m 


wmW' 


/ 


Fig.  3.  Mixed  Sine  Function.  K^Vi  cycles  pei-  second  sine  wave  niLxed  with  a  lower 
amplitude  sine  wave  of  approximately  36  cycles  per  second  simulates  a  "dominant 
alpha"||with  "superimposed  fast  frequency  components."  Smoothing,  mixing, 
and  smoothing  of  the  tliree  square  wave  trains  result  in  tlie  reconstituted  signal 
of  tlie  bottom  trace.  Similarity  between  reconstituted  and  original  signal  suggests 
that  wave  shape  information  is  retained  by  the  processing. 


Information  Processing  in  the  Time  Domain 


335 


fast  frequency  of  approximately  thirty-six  cycles  per  second.  Again 
we  see  the  primary  square  wave,  or  major  period,  reflecting  the 
"dominant  rhythin"  and  the  second  derivative  scjuare  wave  or 
minor  period  rather  clearly  reflecting  the  fast  component.  If  these 
three  trains  of  square  waves  are  smoothed  individually  by  an 
integration  filtering  operation,  inixecl,  and  smoothed  once  more, 
the  reconstituted  analog  signal  may  be  written  out  as  shown  on 
the  bottom  trace.  In  a  way,  this  reconstitution  is  an  inversion  of 
the  operations  which  generated  the  square  waves  in  the  first  place. 
There  is  a  rather  striking  resemblance  between  the  reconstituted 
primary  and  the  original  signal,  although  careful  inspection  will 
reveal  some  discrepancies  in  both  phase  and  amplitude.  However, 
the  wave  shape,  by  and  large,  has  been  retained.  If  the  process 

PERIOD   RECONSTITUTION 
ELECTROENCEPHALOGRAM 


PRI«UC  WAV? 


'JFJ^lTi 


v^/^ 


•jn;  LTUir^ JV"  Jin^a  nf L 


Wl*UW  SQUARE  WA7E 


SECt-i 


\^V 


tEcyi 


'A:^s\jur[mirTKf\^ 


^IR3T 

i 


'^rwrus^AT 


nfi/n4Mi4ifiAaa 


isfUJW/lM^LJUU 


\iWv/\ 


Fig.  4.  Left  Parieto-Occipital  EEC.  Period  analysis  of  the  electroencephalogram 
yields  the  three  square  wave  trains  shown.  The  three  square  wave  trains  are 
smoothed,  mixed  and  smoothed  to  reconstitute  the  wave  forms  in  the  bottom 
trace.  The  complex  EEG  signal  retains  enough  information  in  the  processing 
to  allow  clinical  interpretation. 


336  Information  Storage  and  Neural  Control 

does  this  well  on  a  simple  wave,  what  may  be  expected  from  the 
rather  more  complex  signal  of  the  electroencephalogram? 

Figure  4  shows  that  the  electroencephalogram  is  not  recon- 
stituted as  successfully  as  the  simple  mixed  sine  waves.  Some  of 
the  amplitude  modulation  features  are  lost,  the  envelope  is  not 
as  clearly  evident  on  the  reconstituted  signal,  and  some  phase 
shift  is  apparent  as  distortion  in  a  number  of  the  waves.  Here  again, 
however,  the  resemblance  of  the  reconstituted  wave  to  the  original 
one  is  rather  striking.  The  clinical  electroencephalographer  would 
probably  interpret  the  reconstituted  EEG  in  much  the  same  way 
as  he  would  the  original,  and  would  render  much  the  same  clinical 
impression  after  reading  the  reconstituted  forms. 

Interpretation  of  the  major  and  minor  periods  may  be  accom- 
plished in  the  same  way  as  interpretation  of  an  EEG  but  with 
less  equivocation.  Anyone  who  has  attempted  to  reduce,  quantita- 
tively, long  stretches  of  EEG  record  by  any  form  of  hand  analysis 
will  appreciate  the  significance  of  this.  The  square  waves  may  be 
further  processed  and  displayed  in  several  difTerent  ways,  depending 
on  the  physiological  event  under  investigation.  One  system  we 
have  used  quite  extensively  distributes  the  major  and  minor 
periods  in  a  ten  second  epoch  over  ten  bands  in  the  major  period 
and  ten  bands  in  the  minor  period.  Table  I  defines  the  bands  we 
are  currently  using  in  terms  of  equivalent  frequency.  A  square 
wave  of  the  same  duration  as  the  square  wave  generated  by  an 
eight  cycle  per  second  sine  wave  falls  into  band  4  of  the  major 
period  and  band  1  of  the  minor  period.  The  major  period  bands 

TABLE  I 

Band  Distribution  (As  Equivalent  FREquENCv)  Currently  Being  Used 

IN  "Spectral  Display"  of  the  Square  Wave  Trains  Generated  by  the 

Process  of  Period  Analysis 


Major  Period 

Minor  Period 

{Eq 

uivalent  Frequency) 

{Equivalent  Frequency) 

Band 

in  cps 

in  cps 

1 

1.5-3.5 

1.5-10 

2 

3.5-5 

10-20 

3 

5-7.5 

20-30 

4 

7.5-10.5 

30-40 

5 

10.5-13.5 

40-50 

6 

13.5-18.5 

50-60 

7 

18.5-30 

60-70 

8 

30-50 

70-80 

9 

50-80 

80-90 

10 

80-100 

90-100 

Information  Processing  in  the  Time  Domain 


337 


SPECTRAL   ANALYSIS 


BPINEPHRINE   EFFECT 


Fig.  5.  Histogi am-like  Display  Resulting  from  ''Band  Breakdown''  of  Square  Wave 
Trains.  Small  downward  spikes  indicate  ten  second  epochs.  The  upward  spikes 
are  proportional  in  their  amplitude  to  the  percentage  time  occupied  in  tlie 
previous  ten  seconds  by  square  waves  which  fell  into  a  particular  band   (see 

Table  I). 

approximate  the  frequency  breakdown  employed  in  clinical  electro- 
encephalography. Major  period  band  1  covers  delta  activity,  band 
2  is  a  "slow"  theta,  band  3  is  a  "fast"  theta,  etc.  Minor  period  is 
distributed  as  ten  cycle  increments  in  each  of  the  ten  bands. 

The  histogram-like  display  resulting  from  this  band  breakdown 
process,  which  we  call  spectral  analysis,  is  illustrated  in  Figure  5. 
In  both  traces  the  small  downward  spikes  are  ten  second  epoch 
markers.  The  upward  spikes  are  proportional  in  their  amplitude 


338  Information  Storage  and  Neural  Control 

to  the  percentage  time  occupied  during  the  previous  ten  seconds 
by  square  waves  which  fell  into  a  particular  band.  Reading"  from 
left  to  right  between  epoch  markers,  the  first  band  of  the  minor 
period  is  the  percentage  time  of  all  minor  period  square  waves 
from  one  and  one-half  to  ten  cycles  per  second  (equivalent  fre- 
quency). The  second  band  of  ten  to  twenty  cycles  per  second  has 
been  reduced  from  as  high  as  30  per  cent  time  (full  scale  equals 
50  per  cent  of  full  time)  in  portions  of  the  pre-drug  record  to  an 
insignificant  percentage  during  the  epinephrine  effect.  The  higher 
frequency  bands  on  the  right  of  the  spectrum  have  increased  in 
amplitude  some  30  to  50  per  cent.  This  sort  of  change  we  refer 
to  as  a  "shift  to  the  right."  It  is  characteristic  of  mild  arousal 
such  as  may  be  simulated  by  five  micrograms  intravenous  epi- 
nephrine per  minute.  The  major  period  clearly  shows  a  decrease 
in  band  4  as  the  alpha  activity  is  suppressed  and  replaced  by 
higher  frequencies  and  some  delta  activity. 

We  suspect  that  level  of  sleep  can  be  followed  quantitatively 
by  a  simple  measure  of  the  percentage  delta  time  as  well  as  more 
precisely  by  the  spectral  epochs.  We  say  "suspect,"  since  to  prove 
that  sleep  can  be  fractionated  into,  say,  fifty  distinct  levels  would 
require  an  independent  measure  of  the  state  of  consciousness  having 
the  same  order  of  resolution  as  the  variable  we  are  trying  to 
demonstrate.  Unfortunately,  we  are  unaware  of  a  performance 
measure  or  any  other  measure  that  allows  quantitation  of  the 
state  of  consciousness  or  state  of  arousal  with  as  high  resolution 
as  we  think  is  possible  with  period  analysis  of  the  EEG. 

Several  bioelectronic  measures  other  than  EEG  may  be  amen- 
able to  period  analysis  or  to  some  modification  of  the  process. 
Figure  6  shows  the  first  derivative  of  the  galvanic  skin  response 
(GSR)  signal  as  it  is  employed  to  generate  square  waves  coinci- 
dent with  the  onset-to-peak-amplitude  time  in  the  primary  wave. 
We  regard  the  onset-to-peak-amplitude  time  as  "active  GSR 
time"  since  it  is  the  time  of  depolarization  of  the  membrane  which 
is  the  effector  site  of  this  phenomenon.  Automatic  analysis  of  the 
GSR  produces  two  parameters  of  real  importance  in  psycho- 
physiological interpretation.  The  number  of  square  waves  gen- 
erated per  epoch,  perhaps  ten  seconds  or  perhaps  five  minutes, 
and  the  duration  of  the  active  GSR  time  for  the  given  epoch  are 
partially  independent  parameters  which  seem  worth  considering 


Information  Processing  in  the  Time  Domain 


339 


AtTIVi 


■,     GS8   SIGl 

Shh 

^''-^^ 

~  ' 

DERIVATIVE 


l!  High  Frequency  Noise 

ii    Rejected  by  Pulse 

11     Width  Dis-'-criraination 


IjOW  Frequency,  Low  Level 
Noise  Not  Converted  to 
Rectangular  Pulses 


Effective  Response  Level 


Fig.  6.  Period  Analysis  of  the  Galvanic  Skin  Response.  The  first  derivative  of  the 
primary  GSR  clearly  shows  the  accentuation  of  "noisy"  fast  components.  The 
effective  response  level  or  threshold  for  square  wave  generation  "filters"  out 
slow  components  of  insufficient  amplitude.  Square  waves  of  less  than  one  second 
are  filtered  out  by  the  "period  filter"  (see  text). 


340 


Information  Storage  and  Neural  Control 


in  the  interpretation  of  states  of  arousal.  Figure  6  also  illustrates 
the  use  of  a  "period  filter"  (digital  filter),  which  is  analogous  to  a 
resonant  frequency  filter  in  the  frequency  domain,  but  which  does 
not  have  the  inherent  disadvantages  of  time  lag  for  energy  buildup 
and  decay.  If  the  square  wave  is  of  less  than  five  milliseconds  dura- 
tion in  the  case  of  EEG  and  of  less  than  one  second  duration 
for  the  GSR,  the  period  filter  will  not  "pass"  it  for  further  proces- 
sing. The  GSR  recording  is  often  plagued  by  relatively  high 
frequency  noise  from  movement  artifact.  In  a  noisy  record  this 
high  frec^uency  artifact  may  produce  as  much  as  85  per  cent  of 
the  square  waves.  It  is  a  great  convenience  to  be  able  to  "filter" 
them  out. 

While  the  system  just  described  is  of  real  practical  value  in  the 
analysis  of  the  GSR  activity  of  a  single  subject,  it  is  indispensable 
to  the  "coincidence"  analysis  of  GSR's  from  four  subjects  in  group 
interaction.  Figure  7  is  a  record  of  this  type  of  analysis  in  which 


The  ESToajNE-ANGus  Co 


Fig.  7.  Coincidence  Analysis  of  Group  GSR.  Lines  two  through  five  show  the  square 
wave  trains  generated  by  the  baseline  cross  of  the  first  derivative  of  the  GSRs 
recorded  from  Subjects  A  through  D.  The  coincidence,  or  overlap,  of  "active" 
GSR  time  between  all  pairs  of  subjects  is  shown  in  lines  six  through  twelve. 
Three  of  a  "kind"  and  four  of  a  "kind"  can  be  seen  in  lines  thirteen  through 

seventeen. 


Information  Processing  in  the  Time  Domain  341 

a  special  purpose  digital  computer  is  employed  to  measure  the 
amount  of  coincidence  between  the  active  GSR  time  of  the  in- 
dividual subjects  in  various  combinations.  The  square  wave  of 
Subject  A  is  compared  for  coincidence  or  overlap  in  time  with  the 
square  waves  from  the  GSR  of  Subject  B,  C,  D,  etc.  In  a  four- 
man  group,  the  coincidence  of  all  four  GSR's  at  one  time  is  a 
relatively  rare  event.  Such  a  ''four  of  a  kind"  coincidence  is 
usually  the  result  of  a  rather  strong  stimulus  which  has  been 
experienced  by  all  four  members. 

Consideration  of  coincidence  analysis  has  led  us  tentatively  tc 
formulate  a  model  of  group  interaction  predicated  on  "overlap 
of  value  systems'"  among  the  individuals  of  the  group.  The  specific 
GSR  represents  perception  and  reaction  to  a  specific  stiinulus. 
Generally,  it  may  be  assumed  that  a  stimulus  has  meaning  or  an 
"aflfect  investment"  for  the  individual  if  it  produces  a  GSR.  The 
GSR,  as  an  indicator  of  "investment,"  is  taken  as  a  "yes-no" 
index  without  regard  for  the  afi'ect  polarity.  That  is  to  say,  the 
occurrence  of  a  GSR  reveals  that  the  stimulus  is  "invested"  but 
does  not  reveal  whether  the  stimulus  produces  a  positive  aflfective 
response  or  a  negative  affective  response.  We  are  aware  of  some 
of  the  diflficulties  implicit  in  this  rather  simplified  interpretation 
of  the  GSR  in  relation  to  the  psychological  variables  of  affect 
and  investment.  We  suspect  that  under  certain  circumstances 
extreme  high  negative  affect  may  "freeze"  the  GSR  and  wipe 
out  all  response.  It  may  be  that  this  sort  of  inhibition  effect  is  an 
idiosyncratic  response  of  the  individual  or  that  such  a  phenomenon 
may  be  seen  more  often  in  the  schizophrenic  than  in  the  normal 
patient.  The  group  interaction  is  seen  as  a  continuously  moving 
field  which  presents  a  sequence  of  stimuli  to  all  individuals  in  the 
group.  Some  individuals  may  not  perceive  a  given  stimulus  or 
may  derive  no  meaning  from  it.  When  two  or  more  individuals 
perceive  and  are  invested  in  a  given  stimulus,  it  is  postulated 
that  each  will  produce  a  GSR  and  that  these  GSR's  will  be 
approximately  coincident.  Insofar  as  two  individuals  have  coinci- 
dent GSR's  to  a  finite  but  large  stimulus  array,  our  hypothesis 
would  suggest  that  they  have  "overlap  of  value  systems." 

In  clinical  group  therapy  we  might  ask  the  following  question: 
"In  the  course  of  group  therapy,  will  two  patients  in  the  same  diag- 


342  Information  Storage  and  Neural  Control 

nostic  category  show  more  GSR  coincidence  as  a  pair  than  would 
one  of  tiiese  patients  and  a  third  patient  in  a  different  diagnostic 
category?"  Also  of  interest  is  the  total  number  of  overlaps  for  a 
particular  group.  At  present  we  are  able  to  analyze  only  four 
people  at  a  time,  even  if  the  group  is  composed  of  8  to  10  individ- 
uals; but  unfortunately  in  a  four-man  group  or  in  a  four-man 
subgroup  there  are  various  degrees  of  "coupling"  and  communi- 
cation between  members  that  may  change  the  number  or  degree 
of  coincident  GSR's.  For  such  interpretation  the  total  GSR 
population  should  be  taken  into  account  because  the  number  of 
overlaps  must  be  soine  function  of  the  total  number  of  GSR's 
generated.  In  a  very  loosely  coupled  group,  such  as  four  people 
in  four  different  rooms  without  communication,  there  is  a  certain 
probability  of  overlap  that  can  be  computed  theoretically.  A 
somewhat  more  difficult  theoretical  problem  is  that  of  the  expected 


1500 

. 

VBU.  <»OUP  OW  TWW8  O*  OVMUr  n  MTO 

• 

1300 

- 

• 

• 

1100 

- 

• 

• 

900 

- 

1 

1'°° 

- 

.      1 

'  •  .     • 

500 

~ 

200 

- 

... 

100 

.L 

• 

I 

1 

1       1       1 

i       1 

1500 
Sl«tt  1 


Fig.  8.  Scatter  Diagram  Representing  Approximately  50,000  GSR's  Recorded  in  Group 
Therapy.  The  number  of  coincident  GSR's  between  pairs  of  subjects  per  group 
(Sigma  2)  plotted  against  total  number  of  GSR's  per  group  (Sigma  1)  shows  a 
linear  relationship  which  may  be  used  as  a  baseline  for  interpretation  and  cor- 
rection for  overlap  expected  on  a  probability  basis. 


Information  Processing  in  the  Time  Domain 


343 


overlap  value  in  the  moderately  coupled  group  of  four  people  in 
the  same  room  in  therapeutic  group  interaction.  In  an  empirical 
approach  to  the  problem  of  moderate  coupling,  we  have  plotted 
the  total  number  of  GSR's  generated  by  a  group  against  the 
total  number  of  overlaps  for  that  group.  Figure  8  is  the  scatter 
diagram  of  two  different  groups  in  therapy.  These  data  represent 
approximately  50,000  GSR's.  The  rather  good  linear  relationship 
in  a  fair-sized  population,  with  respect  to  number  of  subjects  and 
hours  of  interaction,  suggests  an  expected  value  of  GSR  coinci- 
dence which  may  be  used  as  a  baseline  for  the  interpretation  of 
overlap  between  subjects  for  small  increments  of  time.  This  tech- 
nique may  allow  us  to  reconsider  group  process  studies  in  terms 
of  this  new  approach. 

The  final  application  of  period  analysis  which  we  would  like 
to  describe  is  its  use  in  connection  with  the  electrocardiogram 
(EKG).  Figure  9  summarizes  some  of  the  parameters,  relationships 
and  cjuestions  which  are  of  interest  to  us  in  reduction  of  the  EKG. 

D/iTA    OBTAINABLE    FROM   PERIOD   ANALYSIS 


m 


PRIMARY   EKG 


^^ 


Q     S 

f,    FIRST   DERIVATIVE 


ANALYSIS 
OF 


PRESENT 
MEASURES 


OTHER  MEASURABLE 
PARAMETERS 


QRS 

T 

PR  INTERVAL 

PR  SEGMENT 

QT  INTERVAL 

ST  SEGMENT 

ST  INTERVAL 


FIRST 
DERIVATIVE 


PR  SEGMENT 

PR  INTERVAL 

QRS 

ST  SEGMENT 


SECOND 

DERIVATIVE      PP"  SEGMENT 

PR  INTERVAL 

QRS 

ST  SEGMENT 


PRIMARY  a 

FIRST 
DERIVATIVE 


RO-T  INTERVAL 
RO-U  INTERVAL 
P-RO  SEGMENT 


a  TIME   DURATION  OF 

ANY    WAVE 
bTIME  DURATION  OF 

ANY   INTERVAL 
(  ZERO  CROSSING  ) 
c  SIGNATURE 

RECOGNITION 


a  TIME  DURATION  OF 

ANY   WAVE 
b  TIME  DURATION  OF 

ANY   INTERVAL 
c  SIGNATURE 
RECOGNITION 


ANALYSIS    QUESTIONS 


I.  IS  THE  P  WAVE  INVERTED? 
2  IS  THE  R  WAVE  INVERTED? 
3,IS  THE  T  WAVE  INVERTED? 
4  IS  THE  MAGNITUDE  OF  P,Q,' 

R.S.AND  T  GREATER   THAN 

SOME   CONSTANT? 
5.Q  =  S?  (TIME) 


1  IS   THE   RATE  OF  CHANGE 
IN   p  Q.R.S.T,  AND  U  WAVES 
GREATER    THAN   SOME 
CONSTANT? 

2  IS  THE  P  WAVE  SYMMETRICAL? 
3.  ARE  CERTAIN   WAVES 

INVERTED  ? 


1  IS  THE   ACCELERATION  OF  P, 
Q,R,S,T,  AND  U  WAVES  GREAT- 
ER THAN  SOME  CONSTANT? 

2  HIGH  FREQUENCY  ACTIVITY? 


1  SYMMETRY  OF  T  ? 

2  ARE  THERE  NOTCHES  IN  P,R, 
AND  T  WAVES  ? 


Fig.  9.  Classical  EKG  Wave  Shape  and  Derivatives.  Parameters  employed  in  clinical 

interpretation  are  related   to  other  parameters  not  usually  considered   and   to 

questions  which  might  be  posed  in  the  analysis. 


344 


Information  Storage  and  Neural  Control 


PERIOD  ANALYSIS       OF       EKG 


PRIMARY  WAVE 


T^ 

i-ST 

»tc— 

$T     INT 

BASE     LINE 


1     m [ 


BASE        LINE      CROSSINGS 
OF       PRIMARY        WAVE 


m   rri 


POSITIVE 
WAVE 


NEGATIVE 
WAVE 


FIRST      DERIVATIVE 


BASE     LINE      CROSSINGS 
OF     FIRST      DERIVATIVE 


_ra R  R  r 


m  Ri  R EZi 


BASE     LINE 


POSITIVE 
WAVE 


NEGATIVE 
WAVE 


SECOND     DERIVATIVE 


BASE     CROSSING 

OF     SECOND      DERIVATIVE 


>"^..,/\j —      BASE      LINE 


nn       13  f?     n   n      positive 

'   '  '  '  wave 


nn  nw\n nn    nega^t^.ve 


Fig.   10.  EKG  Positive  and  Negative  Square   Wave   Trains.  The  square  wave  trains 
generated  by  baseline  crosses  of  the  primary,  first  derivative  and  second  deriv- 
ative of  the  EKG  are  detailed  in  this  ilkistration.  All  durations  and  intervals 
are  available  for  computation. 


Information  Processing  in  the  Time  Domain 


345 


Figure  10  presents  the  square  waves  generated  by  both  the  positive 
and  negative  portions  of  tlie  EKG  and  its  derivatives.  The  physio- 
logical information  contained  in  these  square  waves  and  in  their 
relations  to  one  another  is  still  largely  unknown.  Reports  of  re- 
cent studies  employing  general  purpose  computers  and  utilizing 
coding  points  similar  to  period  analysis  indicate  success  in  charac- 
terizing and  classifying  normal  and  pathological  subjects  (8).  We 
would  like  to  expand  one  particular  problem  of  EKG  analysis 
as  we  have  approached   it  in  our  laboratories.   Both  low  wave 


Fig.  11.  Recognition  of  a  'fat-thin-faV'  Square  Wave  Set.  The  synthetic  function  of 

mixed  sine  waves  slowly  changes  wave  shape  over  time.  If,  and  only  if,  the  wave 

shape  generates  a  square  wave  sequence  within  acceptable  limits,  the  complex 

is  "recognized,"  as  indicated  by  the  spike  in  the  recognition  pulse  trace. 


346 


Information  Storage  and  Neural  Control 


y^^ 


Jl/^aJL/^^J 


F:ltai*d  irljirx  IID 


MiJor 


Riieasiiltlia  FlIm 


Fig.  12.  Artifact-Contaminated  EKG.  Slow  wave  artifact  distorting  wave  shape 
and  high  frequency  "pop"  artifact  are  "filtered"  out  by  absence  of  "recognition." 
Only  those  complexes  within  acceptable  limits  are  held  in  intermediate  storage 
for  further  computation.  Sixty-cycle  artifact  is  filtered  by  conventional  resonant 

rejection  circuits. 


sway  artifact  and  high  frequency  movement  artifact  demand, 
for  practical  analysis,  automatic  rejection  of  the  contaminated 
complex.  The  "filter"  we  have  employed  is  a  system  of  signature 
recognition  or  pattern  recognition.  A  somewhat  different  type  of 
pattern  recognition,  as  defined  in  the  recent  work  of  Steinberg, 
et  al.  (9),  is  a  hybrid  combination  of  Cases  3  and  4,  and  again 
utilizes  several  coding  points  of  period  analysis. 

Figure  11  again  presents  a  synthetic  function  of  mixed  sine 
waves.  The  two  oscillators  drift  in  relation  to  one  another  over 
time,  and  the  wave  shape  pattern  changes  with  this  phase  shift. 
The  square  wave  train  generated  by  the  baseline  cross  of  the 
primary,  the  major  period,  is  presented  to  digital  circuitry  which 
"recognizes"  a  complex  if,  and  only  if,  it  is  made  up  of  a  "fat- 
thin-fat"   square  wave  set.   The  definition   of  "thin"    and   "fat" 


Information  Processing  in  the  Time  Domain 


347 


square  waves  and  the  combination  of  these  square  waves  may  be 
set  up  with  any  desired  hmits  or  sequence;  we  adjust  them  for 
a  given  EKG  signal.  The  hmits  for  this  particular  example  are 
320  to  66  milliseconds  for  a  "fat"  square  wave  and  88  to  2.7 
milliseconds  for  a  "thin"  square  wave.  The  trace  designated 
"recognition  pulse''  in  Figure  11  illustrates  by  the  absence  of  a 
pulse  the  rejection  of  an  improper  sequence  and  individual  square 
waves  not  falling  within  the  defined  acceptable  limits. 

Figure  12  presents  signature  recognition  as  applied  to  the  EKG. 
Both  high  frequency  artifact  and  baseline  sway  distort  this  signal 
and  are  rejected,  so  that  in  this  figure  only  three  complexes  have 
been  "recognized."  Figure  13  is  the  identical  EKG  signal  taken 
at  a  slower  paper  speed  to  display  more  clearly  the  rejection  of 
sway  artifact. 


ii)iwffiiHfH'H'H'y|i'iiff!|i"''!"|'<!'fH'i'!>Hiif^iH''tti^ir^-vt^  ')'^^ll^|  n  ''''t^jWf^ni'^ 


Major   Period 


!rj^v4ru-i_irjr|i*i r|rv4j-t/  jrjrTff  !rjri'-lrv-u-|rjj-lrun/-inr|ri^ — r^-^j^ r|n r-Y^ryy^^ 


Recogr.ltlon  pjlse 


Fig.  13.  Slow  Writeout  of  Artifact  Contaminated  EKG.  Rejection  of  those  portions  of 
the  record  distorted  by  sway  artifact  is  demonstrated  by  the  absence  of  recog- 
nition pulses  in  die  lower  trace. 


348  Information  Storage  and  Neural  Control 

SUMMARY 

Period  analysis  has  been  described  as  a  special  case  of  informa- 
tion processing  in  the  time  domain.  Illustrations  have  been  offered 
of  the  application  of  period  analysis  to  the  electroencephalogram, 
the  galvanic  skin  response  and  the  electrocardiogram.  The  period 
filter,  coincidence  analysis  of  GSR,  and  signature  recognition  of 
EKG  have  been  detailed  as  special  techniques  appropriate  to 
information  processing  in  the  time  domain. 


ACKNOWLEDGMENTS 

The  authors  would  like  to  thank  Messrs.  W.  A.  Spoor,  A.  J. 
Welch,  and  R.  J.  Edwards  for  their  creative  contribution  in 
relation  to  the  work  reported. 


REFERENCES 

1.  Burch,   N.   R.,  and  Childers,  H.  E.:   Physiological  data  acquisition. 

In.  Psychophysiological  Aspects  of  Space  Flight,   ed.    by   Col.   Bernard 
E.  Flaherty.  New  York.  Columbia  University  Press,  1961. 

2.  Goldman,  S.:  Information  Theory.  New  Yoik,  Prentice-Hall,  1955,  p.  67. 

3.  Goldman,  S.:  Information  Theory.  New  York,  Prentice-Hall,  1955,  p.  73. 

4.  Burch,   N.   R.:   Automatic  analysis  of  the  electroencephalogram:   A 

review  and  classification  of  systems.  EEC  &  Clin.  Neurophysiol.,  71: 
827-834.  1959. 

5.  Blackman,  R.  B.,  and  Tukey,  J.  \V.:  The  Measurement  of  Power  Spectra, 

New  York,  Dover  Publications,  1958. 

6.  Saltzberg,  B.,  and  Burch,  N.   R.:  A  rapidly  convergent  orthogonal 

representation  for  EEG  time  series  and  related  methods  of  auto- 
matic analysis.  IRE  WESCON  Convention  Record,   Part  8,   1959. 

7.  Lorente  de  No,   R.:   A  study  of  nerve   physiology.   Studies  From   the 

Rockefeller  Institute  for  Medical  Research,  752.- 384-482,  1947. 

8.  Rikli,  A.  E.  et  al.:  Computer  analysis  of  electrocardiographic  measure- 

ments.  Circulation,  24:643-649,   1961. 

9.  Steinberg,  C.  A.,  Abraham.  S.,  and  Caceros,  C.  A.:  Pattern  recog- 

nition in  the  clinical  electrocardiogram.   IRE   Trans,   on  Bio-Med. 
Elect.,  9:23-30,  1962. 


Information  Processing  in  the  Time  Domain  349 

DISCUSSION  OF  CHAPTER  XIV 

H.  W.  Shipton  (Iowa  City,  Iowa):  May  I  ask  two  questions 
please.  First,  have  you  used  the  advantages  of  your  period  analysis 
system  to  study  the  so-called  "squeak"  effects  that  were  reported 
by  Storm  van  Leeuwuen  about  two  years  ago?  Second,  what  is 
your  approach  to  the  inherent  difficulty  with  all  these  systems  of 
analysis  of  presenting  multichannel  displays?  Have  you,  for 
example,  written  out  the  records  for  two  channels  recorded 
simultaneously? 

Neil  R.  Burch  (Houston,  Texas):  The  answer  to  your  first 
cjuestion  is  no.  We  have  not  investigated  the  '"'squeak"  effect 
reported  by  W.  S.  van  Leeuwuen.  The  answer  to  your  second 
question  is  that  the  single-channel  processing"  we  have  been  doing 
for  a  number  of  years  has  been  directed  toward  trying  to  quantify 
changes  in  the  state  of  consciousness.  We  are  particularly  interested 
in  minimal  shifts  in  the  state  of  consciousness  rather  than  in  con- 
ditions when  a  man  is  in  coma  or  in  a  state  of  panic.  The  work 
we  have  done  in  the  last  year  and  a  half  has  been  directed  toward 
the  problem  you  ask  about.  For  the  display  of  multiple  channel 
information  and  for  better  display  of  the  single  channel,  we  are 
using  a  type  of  analysis  that  is  the  inverse  to  the  overlap  analysis 
of  the  group  GSR.  We  generate  a  train  of  square  waves  with 
signal  A.  These  square  waves  are  minor  period  square  waves 
gaited  by  the  major  period.  This  yields  a  burst  of  minor  period 
square  waves,  a  blank  space,  a  burst  of  minor  period  square  waves, 
a  blank  space,  etc.  The  duration  and  positioning  of  these  waves 
are  characteristic  of  the  wave  shape  in  this  signal.  We  then  take 
signal  B  and  do  exactly  the  same  thing.  Now  we  have  two  trains 
of  square  waves.  We  put  them  into  norlogic  circuits  and  ask  the 
question:  "How  much  anticoincidence  is  present?"  If  these  are 
identical  waves,  we  get  no  readout  at  all.  If  there  is  dissimilarity 
between  signal  A  and  Signal  B,  even  in  very  minor  phase  shifts, 
then  this  system  reads  out  either  the  exact  amount  of  instantaneous 
anticoincidence  or  the  sum  over  one  second  or  more.  We  also 
plan  to  display  this  information  toposcopically,  and  hope  to  be 
able  to  handle  up  to  10  channels  in  this  way. 


PART  V  — SUMMARY  AND  GENERAL  DISCUSSION 

Moderator:   Ralph  W.  Gerard,  M.D.,  Ph.D. 


CHAPTER 
XV 

SUMMARY  AND  GENERAL  DISCUSSION 

Ralph  W.  Gerard,  M.D.,  Ph.D. 


I 


AM  not  confronted  here  with  the  problem  that  so  often  emerges 
in  trying  to  summarize  a  symposium  of  this  kind,  because  Drs. 
Fields  and  Abbott  have  so  clearly  exhibited  the  logical  bones  of 
the  organization.  I  think  it  has  been  beautifully  planned  and,  on 
the  whole,  beautifully  executed.  There  have  been  many  good 
talks  and  many  interesting  lines  of  thought  developed,  not  all  of 
which,  obviously,  can  I  allude  to;  nor  shall  I  attempt  to  mention 
all  the  participants  in  the  course  of  my  discussion,  although  I 
shall  refer  to  things  said  by  practically  all.  A  few  items  to  start 
us  off. 

Dr.  Lindsay,  in  the  opening  theory  session,  made  rather  a  point 
of  distinguishing  product  theories  from  process  theories.  I  had  not 
previously  heard  the  dichotomy  in  that  particular  form,  but  I 
liked  it.  It  is  equivalent,  I  should  think,  to  molar  and  molecular 
theories  and  to  the  term  introduced  by  Mainz,  order-analytical 
interpretations  and  cau.sal-analytical  interpretations;  and  it  does, 
as  Lindsay  suggested,  imply  a  progressive  reduction  from  one 
level  to  another.  He  seemed  to  think  this  is  primarily  because 
psychologists  are  reaching  out  hands  toward  neurophysiologists. 
I  think  the  hands  are  coming  from  both  sides  of  the  gap;  and, 
indeed,  still  partly  an  act  of  faith,  I  am  quite  convinced  that  the 
hands  have  about  touched. 

At  the  level  of  genes,  Kit  and  Echols,  gave  the  beautiful  evidence 
showing  that  the  genetic  code  is  about  to  be  broken;  and,  as  I 
listened,  it  seemed  that  here,  also,  interest  was  moving  from  one 
level  of  thought  to  another.  There  was  again  reductionism;  prob- 

353 


354  Information  Storage  and  Neural  Control 

lems  that  started  pretty  clearly  as  biological  ones  have  now 
become  of  interest  almost  entirely  at  the  level  of  pure  chemistry. 
The  problems  here  are  very  sharp  and,  therefore,  will  very  soon 
become  dull;  because,  when  it  is  possible  to  formulate  the  issues 
as  clearly  as  it  now  is,  getting  the  answers  is  a  matter  of  hard  work 
but  often  lacks  major  intellectual  excitement.  I  think  the  great 
epoch  of  the  nucleotides  is  rapidly  drawing  to  a  close,  although 
several  Nobel  prizes  are  still  lurking  there;  I  am  not  denigrating 
it,  I  assure  you.  I  think  the  most  exciting  area  for  the  future  is 
rather  in  reducing  behavior  to  neurophysiology.  The  questions 
here  are  still  fuzzy  enough  so  that  almost  any  kind  of  answer  is 
likely  to  be  exciting. 

Going  on  with  the  group,  Bateson  gave  us  his  charming  presen- 
tation as  raconteur  and  experimenter.  He  exemplified  beautifully 
the  story  that  psychologists  love  to  pass  around:  One  rat  says  to 
another,  "By  golly,  I've  got  my  experimenter  trained  now!  Every 
time  I  push  the  lever,  he  feeds  me."  He  discussed  the  fact  that  one 
deals  with  metasignals  for  information  as  to  the  kind  of  world 
one  is  facing,  and,  in  this  connection,  there  are  several  points 
that  I  cannot  resist  making. 

There  is  an  obvious  experimental  prediction,  which  perhaps 
has  been  checked.  (I  understand  such  experiments  do  give  the 
predicted  results.)  Bateson  compared  the  classical  conditioning 
experience  of  one  rat  with  the  instrumental  conditioning  experi- 
ence of  another,  and  said  that  each  rat  then  allowed  free  experience 
in  the  world  would  find  his  experiment-induced  expectations  more 
or  less  reinforced.  This  is  part  of  establishing  a  particular  learning 
set.  An  animal  given  a  learning  set  in  terms  of  experience  with 
classical  conditioning  should  learn  an  instrumental  conditioning 
situation  less  easily  than  would  a  naive  animal,  and  vice  versa. 

At  the  human  level,  we  at  Michigan  have  an  interdisciplinary 
study  on  schizophrenics,  attempting  to  break  them  up  into  sub- 
categories. Our  social  scientist  came  to  the  interesting  conclusion 
that  the  social  space  in  which  a  schizophrenic  subject  lives  (in 
contradistinction  to  the  non-schizophrenics  in  the  same  hospital 
and  under  the  same  conditions) — his  social  world — is  different 
from  that  of  non-schizophrenics  and  that  the  behavior  of  the 
schizophrenic,  so  abnormal  relative  to  our  world,  may  not  be  too 


Summary  and  General  Diseussion  355 

inappropriate  to  his.  This  is  closely  related  to  what  Mr.  Bateson 
was  saying". 

In  the  section  on  the  nervous  system,  Dr.  Brazier  gave  us  an 
excellent  picture  of  the  whole  field,  with  some  emphasis  on  how 
spontaneous  wave  generation  might  give  an  internal  comparison 
standard.  Dr.  John  picked  this  up  in  his  research  report,  then  he 
and  Dr.  Morrell  had  a  good  discussion  on  the  mechanism  of 
fixation,  to  which  I  shall  return.  At  the  human  level,  Dr.  Miller 
contrasted  the  problem  of  energy  and  information  flow  and  intro- 
duced the  concept  of  levels,  and  Dr.  Burch  discussed  similar 
problems  in  connection  with  his  technicjue  of  extracting  informa- 
tion from  a  complex  temporal  signal. 

Now,  what  can  one  do  to  integrate  all  these  fine  materials? 
I  should  like  to  conduct  this  discussion  in  terms  of  four  major 
headings:  l)the  question  of  order  and  information  in  general,  and 
as  applied  to  organisms;  2)  the  role  of  the  environment;  3)  the 
problem  of  malleability;  and,  4)  the  problem  of  fixation.  At  the 
end  I  shall  say  a  word  about  our  own  work  on  fixation. 

I  am  not  an  information  theoretician,  but  it  seemed  to  me  when 
I  began  to  put  this  summary  together  that  organizing  the  material 
as  follows  gave  me  further  clarification  of  the  session  on  information: 
Think  of  a  deck  of  cards  in  any  particular  order;  obviously  the 
energy  in  it  is  exactly  the  same  for  any  order.  If  you  burn  the 
deck,  the  calories  obtained  are  the  same  whatever  the  order. 
Furthermore,  any  particular  order  in  a  well-shuffled  deck  is  just 
as  probable  as  any  other  particular  order.  Certain  orders  are  of 
more  interest  than  others,  but  any  order  would  be  of  great  interest 
to  a  player  for  it  determines  the  hands  that  are  dealt. 

I  think  it  is  useful  to  distinguish  a  structural  order  such  as  the 
kind  of  order  in  which  the  cards  come  from  the  manufacturer 
(ace  through  king  and  one  suit  after  another).  Such  structural 
order  we  easily  recognize  in  architecture.  Usually  it  implies  some 
regularity  and  symmetry  and  repetitiveness,  and  ordinarily  we 
are  likely  to  call  this  "order."  But  I  can  easily  demonstrate  to 
you  another  very  diff'erent  order  which  I  might  call  functional 
order — an  apparent  "disorder"'  in  arrangement  that  emits  ordered 
behavior.  You  may  have  played  this  little  trick  as  a  child:  Organize 
the  cards  so  that  by  moving  the  top  card  to  the  bottom  at  each 


356  Information  Storage  and  Neural  Control 

letter  and  turning  up  one  at  the  word  you  spell  out  o-n-e — one; 
the  ace  appears:  t-w-o — two,  the  two  is  turned;  and  so  on,  right 
through  the  deck,  ending  with  the  last  two  cards  of  the  last  suit. 
Examination  of  the  cards  as  they  have  been  ordered  in  the  deck 
so  as  to  give  this  functional  output,  which  recreates  the  structural 
order  of  the  original  package,  reveals  nothing"  at  all;  the  deck 
seems  to  be  completely  messed  up. 

Either  kind  of  order  is  produced  by  some  operation  of  the 
environment  on  tlie  system,  on  the  deck  of  cards;  and  the  amount 
of  information  contained  in  it,  in  the  technical  sense,  is  a  matter 
of  how  well  we  know  the  rules  that  produced  that  particular 
order.  If,  for  example,  one  gives  the  value  of  tt  to  many  hundreds 
of  digits,  the  number  of  bits  needed  to  transmit  it  would  increase 
without  limit  at  the  rate  of  over  three  bits  per  digit.  But  if  the 
formula  for  calculating  tt  is  given,  very  few  bits  are  needed  for 
a  limitless  number  of  digits. 

I  suggest  that  one  sees  structural  order  quite  easily  and  recognizes 
the  rule  almost  intuitively;  whereas,  one  does  not  see  functional 
order  nearly  so  easily  nor  tumble  at  once  to  the  rule.  But  when 
we  do  find  the  rule,  the  information  collapses  and  we  no  longer 
have  the  element  of  surprise.  Certainly  the  whole  history  of  scien- 
tific development  has  followed  such  lines.  In  every  area  we  have 
recognized  structural  elements,  structural  entities,  and  regularities 
long  before  we  have  paid  attention  to  functional  ones. 

Turning  now  to  organisms  in  this  connection,  stored  information 
need  not  require  any  expenditure  of  energy.  It  may,  of  course, 
if  storage  is  dynamic,  but  it  need  not,  as  in  the  structural  storage 
of  books  or  pictures.  Information  flow  does  take  energy,  but 
negligible  amounts  will  ordinarily  suffice.  One  can  think,  in 
organisms,  of  an  overall  structural  information,  seen  in  the  total 
morphology  that  has  been  built  up.  This  is  what  Patten  was 
concerned  with  in  his  study  of  the  morphology  of  an  ecosystem, 
a  kind  of  epiorganism.  This  is  of  interest  per  se  to  the  anatomist, 
the  structuralist;  but  to  the  behaviorist,  the  physiologist,  it  is  of 
interest  more  in  terms  of  what  it  can  yield  as  patterned  behavior. 
If  the  system  is  suddenly  made  unable  to  behave,  if  it  is  killed, 
most  of  this  information  remains  present,  at  least  for  a  time,  but 
it  is  no  longer  of  any  functional  use  or  interest.  In  a  way,  what 


Summary  and  General  Discussion  357 

I  have  just  called  structural  information  is  the  same  as  stored 
information;  but  we  tend  to  think  of  these  gross  structures  a  little 
differently  from  the  micro  ones  of  ordinary  memory,  to  which 
I  shall  return.  In  all,  of  course,  storage  of  information  is  a  matter 
of  past  experience,  either  of  the  race,  with  phylogeny  and  ontogeny 
laying  down  structures  that  ate  essentially  uniform  from  individual 
to  individual  in  the  species,  or  of  individual  experience  and 
learning,  with  the  attendant  high  variance. 

The  flow  of  information  was  discussed  fully  by  Dr.  Miller,  but 
I  shall  add  a  few  general  comments.  First,  all  the  informational 
aspects  of  organisms  are  induced  originally  by  the  environment 
acting  upon  the  system,  and  changes  in  these  aspects  are  over- 
whelmingly the  result  of  continued  environmental  influence.  There 
are,  therefore,  two  extremely  interesting  questions  to  raise  about 
such  influence.  The  first  concerns  the  sensitivity  of  the  system 
to  environmental  influence;  the  second,  the  establishment  of  an 
enduring  change.  Sensitivity  can  be  of  two  kinds:  1)  quantitative — 
what  threshold  of  an  environmental  disturbance  or  alteration  is 
necessary  for  the  system  to  recognize  it,  so  to  speak;  and  2)  quali- 
tative— what  specificity  exists,  what  discrimination  is  made  be- 
tween different  kinds  of  environmental  influences — which  is  per- 
haps even  more  interesting.  So  we  have  the  subquestions  of 
threshold  and  of  specification. 

The  other  large  question  has  to  do  with  the  conditions  under 
which  a  transient  action  of  the  environment  leads  to  a  response 
of  the  system.  The  environmental  action,  although  originally 
ephemeral,  may  become  irreversible  and  lead  to  a  permanently 
altered  system.  When  and  how  does  a  reversible  response  of  the 
system  become  an  irreversible  change?  This  is  the  essential  prob- 
lem of  evolution,  of  individual  development,  of  group  history,  and, 
of  course,  of  individual  learning;  and  I  have  liked  the  term 
"becoming"  for  this  collectivity  of  irreversible  change  of  the 
system  over  time — the  "becoming"  of  the  system.  The  architecture, 
essentially  constant  in  time,  is  its  "being,"  the  reversible  changes  in 
time,  its  "behaving,"  and  the  irreversible  changes  its  "becoming." 
Let  us  look  at  the  environment  system  in  a  little  more  detail. 

The  environment  alone  is  able  to  induce  inhomogeneities  in  a 
homogeneous  system;  and  if  the  latter  is  appropriately  responsive 


358  Information  Storage  and  Neural  Control 

to  particular  inhomogeneities,  there  will  be  a  morphogenetic 
action  and  internal  structure  will  result.  Some  of  you  may  not 
remember  the  vast  argument  that  occurred  near  the  turn  of  the 
century  when  the  German  zoologist  Driesch  shook  apart  the  two 
half  cells  of  a  fertilized  egg.  Normally,  of  course,  one  would  become 
the  right  side,  say,  of  a  frog  and  the  other  the  left  side;  but  after 
separation,  each  became  an  intact  frog  with  perfectly  good  right 
and  left  sides.  The  outer  cell  surfaces  exposed  to  pond  water 
developed  skin  in  the  proper  fashion,  but  the  medial  surfaces, 
which  became  backbone  and  nervous  system  when  left  stuck 
together,  now  also  developed  skin.  This  phenomenon  caused 
Driesch  to  turn  vitalistic  and  invoke  guiding  entelechies,  but  it 
was  explained  decades  later  by  the  American  zoologist  Child  in 
terins  of  concentration  gradients  from  outside  to  center.  In  the 
intact  embryos  the  medial  cell  surfaces  are  at  the  low  or  high 
end  of  a  gradient  of  oxygen,  carbon  dioxide,  or  any  other  sub- 
stance that  must  diffuse  from  or  into  the  environment;  but  in  the 
separated  cells  the  end  of  the  gradient  has  moved  to  the  center 
of  each  cell  instead  of  the  center  of  the  double  cell  mass.  So, 
provided  the  cell  is  more  than  a  sac  of  water  and  is  able  to  respond 
to  different  oxygen  concentrations  by  different  morphological 
responses,  the  organized  morphology  results  from  these  quantita- 
tive changes  imposed  by  the  environment. 

The  same  sort  of  thing  operates  throughout  embryonic  develop- 
ment. With  further  cell  divisions  the  germ  layers  become  differ- 
entiated and  then  organs  are  specified.  Often  it  is  only  a  matter 
of  minutes  between  the  appearance  of  the  endoderm  and  the 
irrevocable  commitment  of  a  given  endoderm  cell  to  become  a 
bit  of  liver  or  of  gut.  In  tliis  particular  case  we  know  what  the 
environmental  determiner  is:  if  the  cell  is  near  heart,  it  becomes 
liver;  if  not,  it  becomes  gut.  So  environmental  influences  operate 
all  the  way  through  ontogenesis,  in  gated  time  periods,  to  produce 
firm  outcomes. 

We  are  thoroughly  familiar  with  this  in  many  other  areas  as 
well.  We  can  tell  what  kind  of  environment  a  person  has  lived  in 
if  he  has  thick  soles  or  horny  hands  or  a  weathered  face.  Frown 
or  smile  wrinkles  are  inorphological  consequences  of  oft-repeated 
behaviors.  In  this  case,  the  environment  of  the  skin  is  internal  to 


Summary  arid  General  Discussion  359 

the  system  (the  facial  muscles),  but  this  does  not  alter  the  principle. 
The  ontogenesis  of  an  ecological  community,  i.e.,  the  evolution 
of  the  group  roles  and  structures  that  form  during  community, 
is  similar.  Such  roles  and  structures  can  form  only  in  certain 
sequences  and  at  certain  stages  in  the  interactions  of  the  indi- 
viduals that  constitute  the  "cells"  of  society,  and  in  time  can 
become  irreversible.  These  include  customs  and  rules,  libraries, 
and  all  sorts  of  appurtenances  that  form  a  morphological  substrate 
and  channel  social  behavior.  And,  of  course,  the  engram  in  the 
brain  is  entirely  comparable  to  horny  skin  or  to  bowed  legs  or 
to  wrinkles.  It  is  interesting  that  a  time-gated  period  of  specifica- 
tion has  more  recently  been  found  not  only  in  differentiation  of 
cells  but  in  "imprinting"  the  nervous  system  and  in  fixation  of 
experience  in  still  other  areas.  One  is  inclined  to  raise  the  question 
of  whether  the  units  involved  are  in  a  sort  of  soft-shelled  state, 
like  a  molting  crab,  all  at  the  same  time,  or  whether  different 
units,  particular  neuron  groups,  become  impressionable  in  separate, 
temporally  ordered  periods.  This  also  relates  to  the  earlier  argu- 
ment on  memory,  and  I  shall  come  back  to  it. 

Now  a  word  about  malleability.  This,  you  will  remember,  refers 
to  the  sensitivity  and  the  specificity  of  an  organism  relative  to  its 
environment,  particularly  to  the  rain  of  information  from  the 
environment.  Over  evolutionary  sequences  there  develops  greater 
ability  to  respond,  with  greater  chscrimination,  to  more  kinds  and 
lesser  amounts  of  such  information.  In  fact,  I  would  urge  that 
the  major  theme  of  organic  evolution  is  what  I  have  called  the 
epigenetic  inode  and  is  not  just  the  ability  to  respond  to  the 
environment,  to  learn,  or  to  be  molded  by  it;  beyond  that,  it  is 
also  the  ability  to  be  molded  more  and  more  easily — to  learn  to 
learn.  This  learning  to  learn  occurs,  I  think,  at  all  levels  and  in 
all  systems  in  the  course  of  "becoming,"  not  only  in  evolution 
and  history  but  also  in  the  individual,  as  psychologists  well  know. 

Several  major  inventions  of  life  have  favored  this  successful 
increase  in  the  ability  to  learn.  Perhaps  the  first,  certainly  one  of 
the  very  early  and  important  ones,  was  the  invention  of  an  array 
of  molecules  able  to  replicate  themselves  and  to  produce  other 
particular  molecules  (in  other  words,  the  invention  of  an  array 
of  genes  with  sufficient  stability  and  sufficient  mutability).  This 


360  Injormation  Storage  and  Neural  Control 

permitted  very  slow  evolution.  A  great  speeding  up  of  modi- 
fication of  the  system  by  environmental  impact,  i.e.,  an  enhance- 
ment of  response  to  the  information  available,  allowed  a  second 
forward  step — the  invention  of  sex.  This  latter  maneuver  made 
it  possible  to  mix  the  genes  in  two  individuals,  to  shuffle  the  cards, 
and  so  get  an  almost  infinite  number  of  hands  with  the  same  small 
array  of  individual  items. 

The  third  major  landmark  was  the  invention  of  multicellularity. 
This  made  possible  the  setting  off  of  groups  of  cells,  tissues,  and 
organs  for  particular  functions,  including  susceptibility  to  environ- 
mental influences.  Multicellularity  made  possible  a  meaningful 
nervous  system,  the  appearance  and  steady  improvement  of  which 
is  the  most  important  invention  for  us.  This  evolution  over  suc- 
cessive epochs  probably  involved  an  initial  improvement  of  the 
individual  unit  neurons  from  decrementing  to  all-or-none  con- 
duction, from  reciprocal  to  irreciprocal  synapses,  from  lower  to 
higher  speeds,  from  higher  to  lower  thresholds,  and  all  the  rest. 
Then  there  developed  better  circuitry  between  the  neurons, 
including  such  effective  physiological  devices  as  the  simple  reflex, 
the  reverberating  loop,  the  negative  feedback  loop,  etc.  Two  of 
the  circuits  already  mentioned  are  worth  a  moment. 

Dr.  Brazier,  particularly,  referred  to  one  as  the  "inhibitory 
surround."  This  term  emphasizes  recent  work  by  investigators  such 
as  Hartline,  Hubel,  and  many  others,  dealing  with  the  sensory 
input,  but  the  mechanism  really  goes  back  to  Sherrington's 
reciprocal  inhibition.  This  mechanism  not  only  cuts  in  a  clean 
group  of  motor  neurons  to  give  a  shaiply  integrated  act,  very 
possibly  via  the  feedback  inhibition  by  Renshaw  cells,  but  it  also 
operates  all  through  the  nervous  system.  I  have  suggested  in 
The  Handbook  of  Neurophysiology  that  it  functions  in  giving  attention 
to  one  or  another  sensory  input  or  thought  train  and  in  shifting 
mood  sharply,  as  well  as  in  selecting  a  behavior.  This  device 
(active  units  blocking  out  nearby  ones  that  could  have  become 
engaged  in  the  activity  but  are  in  this  way  kept  inactive)  is  the 
basic  mechanism  for  dissecting  a  graded  continuum  into  sharp 
classes.  "Nature  doesn't  come  as  clean  as  we  can  think  it,"  as 
Whitehead  said,  but  our  whole  nervous  system  and  our  sense 
organs   are  designed   to  clean   it  up   for   our   thought  processes. 


Summary  and  Ge?ieral  Discussion  361 

Perception  of  an  object  comes  through  clean  and  sharp,  and  an 
act  comes  through  clean  and  sharp  without  conflict  or  blurring 
by  opposing"  elements.  Sometimes  we  err  grievously  by  over- 
commitment to  a  typology,  as  did  the  scholastic  philosophers;  but 
without  such  a  commitment  we  could  not  think  at  all,  and  with 
sophistication  we  can  return  to  graded  or  probabilistic  thinking. 
The  mechanisms  are  standard  orthodox  neurophysiology;  their 
behavioral  consequences  are  still  being  explored. 

The  second  neural  circuitry  worth  mentioning — it  has  received 
much  attention  here — is  the  double  system,  discrete  and  diffuse. 
The  diffuse  system  gives  the  metasignals  which  are  the  set.  It  acts 
like  the  basic  adjustments  of  the  television  set  that  make  a  picture 
possible:  adjusting  brightness  and  discrimination,  locking  in  the 
vertical  and  horizontal,  etc.,  but  not  giving  the  actual  picture. 
The  discrete  system  presents  the  picture,  the  particular  pattern 
that  receives  our  attention.  I  have  probably  oversimplified  this 
(an  example  of  oversharpening  nature)  but  there  is  much  evidence 
for  it.  The  diffuse  system  can  modulate  thresholds  and  responses 
of  the  cortical  neurons  that  are  thrown  into  action  initially  by 
the  discrete  system;  and  the  diffuse  system  does  affect  mood,  set, 
emotional  background,  even  level  of  conscious  awareness  and 
attention.  The  whole  question  of  novelty,  stress,  anxiety,  and 
performance  has  been  discussed  (Gerard,  R.  W.:  Neurophysiology; 
an  integration,  in,  Handbook  of  Physiology — Neurophysiology  III, 
Victor  E.  Hall  et  al.,  eds.,  Amer.  Physiol.  Society,  I960,  p.  1919) 
in  relation  to  the  interaction  of  the  two  systems  in  modifying  the 
size  of  a  "physiological  neuron  reserve." 

Returning  to  the  overall  evolution  of  the  nervous  system,  the 
third  stage,  after  improved  units  and  organized  circuits,  is  increase 
in  number.  The  great  rise  in  capacity  of  the  vertebrates,  and 
particularly  of  the  mammals,  is  attended,  so  far  as  I  know,  neither 
by  improvements  in  the  neurons  and  their  connections  nor  by 
any  better  circuitry.  It  is  a  remarkable  consequence  of  simply 
adding  more  of  the  same.  While  this  is  surprising  at  first,  a  little 
thought  recognizes  that  more  of  the  same  can  add  entirely  new 
dimensions  of  richness  in  performance.  In  fact,  I  was  struck  by 
the,  I  am  sure  accidental,  parallel  in  the  number  of  base  pairs 
in  genes  and  of  neurons  in   brains.   The  small  virus  has  about 


362  Information  Storage  and  Neural  Control 

6,000  base  pairs,  the  mammal  close  to  10'",  according  to  Dr.  Kit. 
The  simplest  animals  possess  a  few  hundred  or  thousand  neurons, 
man  about  10'".  Adding  more  of  the  same  does,  indeed,  multiply 
richness  and  capacity. 

The  next  major  breakthrough  in  increasing  overall  malleability 
of  living  things  became  possible  only  when  the  nervous  system 
had  become  large  enough  and  sufficiently  complex  to  generate 
those  new  capacities  of  interaction  which  led  to  culture.  Culture, 
while  not  completely  limited  to  man,  is  tremendously  more 
enveloping  for  this  social  animal,  and  I  suggest  four  sub-epochs 
in  its  development.  The  first  stage  of  culture  probably  can  be 
dated  from  the  invention  of  the  symbol,  the  use  of  an  arbitrary 
sign  for  a  thing,  a  communicable  representation  of  the  outside 
world.  Next  came  organized  symbols,  which  are  language,  as  a 
tremendous  advance,  and  tested  organized  symbols,  which  are 
science,  as  a  further  great  step.  I  strongly  suspect  that  we  are 
just  entering  a  fourth  epoch  in  increased  malleability  of  collective 
man  with  the  invention  and  rapid  growth  of  the  computer,  a 
prosthetic  instrument  for  thinking,  much  as  bulldozers  are  for 
muscles  and  telescopes  and  microphones  are  for  receptors. 

In  fact,  perhaps  the  most  interesting  thing  about  present-day 
man  is  that  the  world  in  which  he  lives,  the  one  that  matters, 
that  gives  problems  and  satisfactions,  is  no  longer  very  much  a 
material  world  of  "things."  These  have  been  taken  care  of.  We 
have  established  homeostatic  control  of  our  physical  and  biological 
environment  so  that  these  no  longer  present  our  primary  problems. 
We  live  as  social  beings  in  an  ocean  of  information,  information 
that  did  not  exist  before  we  created  it.  Languages  of  all  sorts, 
pictures  of  all  sorts,  a  great  variety  of  communication  means  and 
contents — these  are  the  things  that  matter  to  us.  Our  interactions 
with  other  human  beings,  mainly  at  the  symbolic  level,  are  what 
we  care  about.  Indeed,  the  storing,  processing,  and  retrieving  of 
information  at  the  machine  level  are  undergoing  such  tremendous 
advances  that  the  entire  transmittal  and  use  of  the  information 
which  is  the  corpus  of  our  culture  will  soon  be  revolutionized. 

There  is  still  another  exciting  aspect  of  the  evolution  of  mallea- 
bility that  requires  mention.  In  the  earlier  phases,  this  evolution 
took  place  primarily  by  a  biological,  Darwinian  kind  of  process; 


Sumtnarv  and  General  Discussion  363 

later  it  continues  primarily  by  an  environmental  social,  Lamarckian 
kind.  I  shall  return  to  this  shortly,  but  must  first  examine  the 
last  major  topic,   the  fixation  of  information. 

For  experience  to  be  fixed  or  information  to  be  stored,  there 
must  be  a  material  change  of  some  kind.  If  a  system  is  to  retain 
an  enduring  difference  induced  by  the  environment,  not  just  a 
relatively  ephemeral  change  in  dynamic  state,  as  a  spinning  top, 
the  different  responsiveness  must  rest  on  a  morphological  differ- 
ence. Such  a  material  change  can  be  only  in  the  number  or  kind 
or  position  of  units,  suc'i  as  ions,  molecules,  organelles,  cells,  or 
perhaps  all  of  these.  One  is  tempted  to  look  at  the  macromolecules 
because,  at  least  at  that  level,  they  are  the  only  units  that  have 
considerable  endurance  in  cells.  It  is  by  no  means  excluded  that 
the  lipids,  which  endure  very  well  (some  of  them,  once  formed, 
apparently  have  no  turnov^er  during  the  life  of  the  brain),  or  the 
proteins  might  be  involved;  but  most  investigators  interested 
in  this  field  have  a  strong  predilection  for  the  polynucleotides. 
Moreover,  as  pointed  out  earlier  in  this  symposium,  there  is 
growing  evidence  that  implicates  them,  and  there  is  an  especially 
intriguing  reason  for  interest  in  RNA  and  memory. 

DNA  molecules  produce  another  generation  of  DNA,  these 
produce  another  generation,  and  so  on.  For  a  series  of  generations, 
the  important  thing,  of  course,  is  that  means  of  replication  do 
exist  and  that  they  are  precise  enough  to  give  both  great  stability 
and  appropriate  freedom  for  change.  Change  is  produced  very 
gradually  over  generations  with  the  environment  acting  primarily 
by  means  of  selection.  The  environment  normally  does  not  alter 
the  DNA  molecules,  although  it  is  ultimately  responsible  for  the 
rare  and  random  genetic  mutations.  Rather,  it  selects  one  or 
another  set  of  these  molecules  in  terms  of  the  phenotypes  produced 
and  of  the  relative  degree  of  their  adaptation  to  the  environment. 
This  is  Darwinian  evolution  — natural  selection  of  certain  molecules 
from  an  array  of  possible  DNA  molecules  or  groups  of  molecules. 
But  when  a  given  DNA  molecule  starts  to  operate  in  a  given 
organism,  it  produces  messenger  RNA  and  ribosome  RNA  and 
proteins  and  enzymes  and  all  tlie  rest;  and  somehow  or  other  this 
sequence  is  under  pretty  direct  control  of  the  environment.  Indeed, 
it  looks  as  if  there  is  here  a  Lamarckian  kind  of  influence  by  the 


364 


Injormatwn  Storage  and  Neural  Control. 


DNA 


Darwinian  selection  by  Environment 


DNA  — ^^  Messenger  RNA  — ^-    RNA    — ^-   protein 


y 


Lamarckian  modification  by  environnnent 


DNA 


Ontogeny 

Figure  1 


environment  (Fig.  1).  Just  where  in  the  sequence  it  acts,  we  do  not 
know;  but  a  reasonable  guess  would  be  that  it  operates  on  the 
messenger  RNA,  which  is  small  in  amount  and  relatively  unstable, 
to  modify  it  in  kind  or  amount  or  distribution. 

This,  I  think,  reveals  the  nub  of  the  earlier  discussion  between 
Dr.  John  and  Dr.  Morrell.  The  extremely  basic  question  arises: 
Must  we  assume,  or  is  it  better  to  assume,  that  the  environment 
operates  here  by  modifying  the  RNA  (or  other)  molecules,  which 
is  Lamarckianism;  or  is  it  possible  that,  as  in  genetic  selection, 
there  is  a  large  array  of  molecules,  say  a  gene-like  array  of  RNA's, 
on  which  environment  operates  by  some  kind  of  selection?  I  am 
sure  nobody  knows  the  answer  at  the  moment;  the  situation  does 
not  have  quite  the  feel  of  selection  to  a  biologist,  but  feelings  can 
be  very  wrong.  Moreover,  I  would  point  out  that,  if  molecular 
modification  is  involved,  we  have  not  solved  the  critical  problems 
when  we  recognize  that  this  occurs.  It  is  important  to  get  this 
far;  but  some  workers  have  talked  as  if  identifying  a  memory 
trace  with  a  change  in  RNA  is  essentially  the  solution  of  the 
engram.  Rather,  we  are  then  at  the  very  beginning  of  our  troubles. 
Exactly  the  same  problems  face  us  here  that  faced  Lamarck  in 
getting  the  giraffe's  neck  longer.  Let  mc  point  out  what  these 
problems  are.  The  environment  leads  the  giraffe  to  stretch  his 
neck;  somehow  stretching  the  neck  generates  a  substance,  or 
influence,  which  goes  from  the  neck  to  the  gonads  and  produces 


Summary  and  General  Discussion  365 

a  change  in  the  sex  cells;  this  change  specifically  favors  the  develop- 
ment of  a  longer  neck  in  the  offspring  giraffe — a  truly  formidable  re- 
quirement, which  alone  made  Lamarckian  inheritance  improbable. 

Our  demands  are  no  less.  We  require,  also,  transduction  from 
a  process  to  a  structure  and  back  to  a  process,  from  information 
fiow  to  information  storage  to  information  retrieval.  Nerve  messages 
and  events  must  be  fixed  in  some  kind  of  stable  architectural 
alteration  which  favors  regeneration  of  comparable  events  from 
the  system.  The  flow  of  information  is  a  matter,  essentially,  of 
action  at  synapses  where  nerve  cells  junction.  Synapses  can  vary 
only  in  number,  or  intensity,  which  is  really  equivalent;  position; 
kind,  to  some  extent,  as  excitatory  or  inhibitory;  and,  of  course 
the  temporal  phase  of  their  activity.  There  are  no  other  parameters, 
for  these  synaptic  attributes  also  express  the  patterns  of  neuron 
connectivities. 

The  storage  occurs  during  a  period  of  fixation,  as  I  have  called 
it,  or  consolidation,  as  Dr.  Morrell  called  it,  during  which  a 
reversible  change  becomes  irreversible  and  an  enduring  memory 
is  established.  This  engram  probably  includes  a  molecular  change 
and,  as  just  discussed,  may  involve  production  of  an  altered  mole- 
cule or  selection  of  a  particular  molecule  from  a  pre-existent  array. 
Selection  might  be  in  position  or  in  number  as  well  as  in  archi- 
tecture of  molecules.  Given  the  molecular  change,  still  further 
consolidation  processes  over  time  might  well  involve  more  gross 
morphological  changes,  such  as  enlargement  of  end-feet  or  actual 
sprouting  of  axon  branches  (there  are  many  more  in  old  neurons 
than  in  young  ones);  but  this  is  all  guess  work.  Perhaps  there  are 
only  a  given  number  of  slots,  so  to  speak,  in  which  memories  can 
last,  although  any  notion  of  one  memory  in  one  slot  is  untenable. 
There  is  conclusive  physiological  and  psychological  evidence  that, 
at  most,  there  are  different  arrays  or  patterns  of  neuron  groups 
which  subserve  different  memories,  with  some  spatial  separation 
as  well  as  overlap. 

Then,  finally,  we  must  account  for  the  ability  of  the  particular 
morphological  residue  left  by  a  given  pattern  of  impinging  im- 
pulses in  turn  to  make  the  neuron  sensitive  to  just  that  pattern  of 
impulses,  so  that  in  the  future  this  input  can  fire  the  cell  more 
easily  than  other  inputs. 


366  Information  Storage  and  Neural  Control 

Dr.  John  made  a  noble  effort  to  reduce  all  this  to  a  single 
quantitative  picture  by  pointing  out  that  an  increase,  say,  in  total 
cellular  RNA  would  bind  more  ions  and  thereby  cut  down  intra- 
cellular potassium  which  would  slow  the  discharge  of  the  neuron 
membrane  and  the  optimal  frequency  at  which  it  would  respond. 
Explanations  of  this  sort  we  eagerly  welcome.  Many  workers  are 
engaged  in  such  efforts  to  push  understanding  further.  My  own 
feeling  is  that  if  one  reduces  the  RNA  change  to  a  single  overall 
quantitative  parameter,  even  if  parceled  out  to  different  cell 
regions  or  membrane  areas,  there  does  not  remain  the  necessary 
great  specificity;  but  this  is  certainly  a  matter  of  opinion  at  the 
moment.  In  any  event,  here  are  the  active  growing  points  of 
experiment,  as  well  as  theory,  in  this  field. 

I  shall  take  a  final  moment  to  add  to  those  facts  already  before 
you  a  few  new  ones  regarding  fixation.  Dr.  Morrell  referred  to 
our  earlier  work,  paralleled  independently  by  others,  of  giving 
an  animal  a  certain  learning  experience  and  then,  after  different 
intervals,  stopping  the  activity  of  the  brain.  We  found  that  if 
brain  activity  was  stopped  early  enough,  either  by  abrupt  cooling 
or  by  massive  electric  shock,  there  had  not  been  time  for  the 
experience  to  become  fixed  in  the  nervous  system.  A  hamster  or 
rat  given  an  electric  shock  within  a  few  minutes  of  an  experience 
had  no  recollection  of  the  experience;  the  animal  learned  nothing, 
much  like  the  retrograde  amnesia  of  man  after  a  concussion.  The 
fixation  time  so  established  was  fifteen  minutes,  although  changes 
continued  for  fully  an  hour.  To  grapple  more  firmly  with  the 
engram,  we  wished  a  more  localizing  preparation,  but  without 
encroaching  on  MorrelTs  elegant  mirror  spot  technicjue  in  the 
cortex.  There  has  been  much  argument  as  to  whether  the  cord 
can  or  cannot  fix  experience,  or  learn.  Chamberlain,  Haleck,  and 
I  decided  to  follow  a  clue  provided  by  an  Italian  physiologist, 
Di  Giorgio,  relating  to  enduring"  postural  asymmetries  after  uni- 
lateral lesions  in  the  cerebellum  or  other  cephalad  structure. 
Many  mammals  show  the  phenomenon.  We  have  used  rats  mainly. 

After  an  asymmetrical  lesion,  the  right  hind  leg  is,  say,  more 
flexed,  the  left  one  more  extended.  Now,  of  course,  if  the  cord  is 
cut,  the  asymmetric  streams  of  descending  impulses  are  stopped 
and   cord    discharges   should    lapse    back   to   symmetry.    This   is, 


Summary  and  General  Discussion  367 

indeed,  what  happens  if  the  cord  is  cut  within  three  quarters  of 
an  hour  after  the  start  of  asymmetry.  But  if  the  asymmetry  has 
been  allowed  to  persist  longer  than  this,  and  the  time  discon- 
tinuity at  forty-five  minutes  is  too  sharp  for  comfort,  then  the 
asymmetry  remains  for  hours  or  days  after  the  cord  is  cut.  Clearly, 
physiological  activity  has  been  fixed  in  cord  neurons;  and  one 
has  an  obvious  place  to  look  for  shifts  in  DC  potentials  across  the 
cord,  in  unit  activity  of  motor  neurons,  in  RNA  and  enzyme 
content  in  various  cells  in  the  cord,  and  the  like.  Further,  we  are 
examining  the  influence  on  fixation  time  of  drugs  which  speed  or 
slow  the  formation  of  RNA,  and  Rothschild  is  making  comparable 
studies  on  the  learning  abilities  of  rats  and  mice  in  various  maze 
and  avoidance  situations.  It  does  look  as  if  8-azaguanine,  which 
slows  RNA  formation,  slows  learning  and  prolongs  fixation  time; 
and  that  a  malononitrile  dimer  (Upjohn  U9189),  which  is  reported 
to  speed  RNA  formation,  may  have  the  reverse  eflfect.  But  results 
are  still  coming  in  and  all  this  is  very  preliminary. 

In  any  event,  inany  workers  are  zeroing  in  on  many  prepara- 
tions, including  the  flatworm,  and  we  are  really  beginning  to  come 
to  grips  with  the  problems  of  information  processing"  and  storing 
by  the  nervous  system. 


DISCUSSION  OF  CHAPTER  XV 

Ralph  W.  Gerard  (Ann  Arbor,  Michigan):  I  would  like  to 
invite  questions  and  comments  from  those  who  participated  in 
the  symposium.  All  those  in  the  audience  will  wish  to  hear  the 
views  of  the  participants  on  what  some  other  speaker  has  said. 

Frank  Morrell  (Palo  Alto,  California):  Dr.  Gerard,  may  we 
ask  you  to  amplify  the  details  of  this  beautiful  experiment.  I 
would  particularly  like  to  know  the  details  of  how  the  operation 
to  produce  asymmetry  was  done,  whether  drugs  influence  this, 
and  whether,  for  exainple,  the  same  relations  exist  if  such  an 
operation  is  performed  using  anesthesia. 

Gerard:  The  preparation  is  made  using  anesthesia,  and  time 
is  from  the  appearance  of  asymmetry,  not  from  the  time  of  the  cord 
cut.  Anesthesia  (ether  or  nembutal)  is  light,  and  the  animal  is  ordin- 


368  Information  Storage  and  Neural  Control 

arily  pretty  well  out  when  asymmetry  appears.  Before  that,  presum- 
ably, impulses  coming  down  the  cord  have  not  been  effective. 

E.  Roy  John  (Rochester,  New  York):  I  would  like  to  mention  a 
couple  of  experiments  related  to  your  remarks  and  ask  if  you 
would  react  to  them.  I  ain  sure  the  first  one  will  be  of  interest 
to  you,  although  it  is  not  directly  related  to  the  question  of  memory 
in  the  nervous  system,  but  rather  to  your  comments  on  the  loss 
of  plasticity  and  functional  specialization  in  tissue.  The  data  are 
contained  in  a  recent  paper  by  Buchsbaum  in  the  Journal  oj 
Experimental  ^oology.  He  and  his  co-workers  were  trying  to  develop 
a  planarian  tissue  culture  inethod,  and  succeeded  in  making  a 
pleasantly  simple  medium  in  which  explants  were  grown.  They 
observed  that  a  small  explant  occasionally  proliferated  as  a  sheet, 
reached  a  certain  size,  folded  back  on  itself,  apparently  dedifferenti- 
ated, and  developed  into  a  planarian.  This  rather  unexpected 
observation  suggests  that,  at  least  at  this  level,  the  loss  of  plasticity 
with  specialization  is  reversible. 

More  directly  relevant  to  our  major  concern  here  is  the  recent 
paper  by  Sporn  and  Dingman  in  the  Journal  of  Psychiatric  Research 
in  which  8-azaguanine  was  used  to  interfere  with  RNA  synthesis, 
and  a  significant  decrease  in  the  rate  of  maze  learning  was  ob- 
served. I  would  also  like  to  mention  the  on-going  thesis  work  of 
Eugene  Sachs,  in  our  laboratory,  which  may  provide  additional 
insight  into  aspects  of  information  storage. 

Some  time  ago,  in  collaboration  with  Wenzel  and  Tschirgi,  we 
observed  that  small  intraventricular  injections  of  electrolytes  seri- 
ously interfered  with  the  performance  of  some  previously  estab- 
lished conditioned  responses.  Mr.  Sachs  has  investigated  the  effects 
of  small  alterations  in  central  potassium  or  calcium  on  learning 
and  performance  by  making  intraventricular  injections  before 
each  training  session.  Control  groups  are  first  trained,  and  then 
receive  an  appropriate  number  of  central  injections.  Sachs'  results 
indicate  that  animals  perform  conditioned  responses  best  under 
conditions  of  central  electrolyte  concentration  like  those  present 
during  training  and  poorly  under  other  conditions,  including 
normal  cerebrospinal  fluid  concentration.  Control  groups  that 
receive  the  injections  after  training  show  no  evidence  of  accom- 
modation effects.  In  these  animals,  central  injection  causes  per- 


Summary  and  Ge?ieral  Discussion  369 

formance  deterioration,  while  in  the  annnals  in  which  these  changes 
were  present  during  learning,  performance  continues  perfectly. 
Certain  chemical  changes  seem  to  facilitate  learning,  while  others 
slow  it.  These  groups  showed  differential  sensitivity  to  drugs  many 
months  after  training,  indicating  that  the  effects  of  the  small  elec- 
trolyte shifts  are  long  lasting.  These  various  findings  show  that 
very  small  local  electrolyte  shifts  seem  capable  of  affecting  the 
long-term  storage  of  an  experience  in  such  a  way  that  readout 
is  optimal  when  the  electrolyte  microenvironment  of  the  readout 
mechanism  resembles  the  situation  during  the  initial  experience. 

I  would  like  to  add  just  one  thing.  We  have  replicated  the 
cannibalism  experiments  of  McConnell  using  a  blind  procedure. 
It  seems  to  me  that  the  most  striking  evidence  in  favor  of  a  sequence 
specificity  model  comes  from  such  studies. 

Gerard:  Let  me  answer  your  second  question  about  the  aza- 
guanine  findings  first.  That  paper  appeared  while  our  own  experi- 
ments were  in  progress  and  were  coming  out  the  same  way.  As 
to  your  rounded-up  planaria,  I  seem  to  remember  that  Child  and 
Hyman  got  smaller  segments  to  regenerate,  but  this  is  an  unim- 
portant detail.  I  am  not  quite  sure  what  you  are  asking  of  me. 
Maybe  another  question  will  help:  When  the  cells  reorganize 
inside  such  a  sheath  or  coating  and  are  all  mixed  up,  if  you  have 
previously  trained  them,  do  they  remember? 

John:  That  is  one  reason  why  we  are  doing  tissue  culture  ex- 
periments. I  do  not  know  yet. 

Gerard:  Regarding  your  electrolyte  shift,  this  strikes  me  as 
exactly  what  would  be  expected,  on  the  following  argument. 
Small  shifts  in  the  calcium-potassium  ratio  produce  large  changes 
in  the  neuron  thresholds;  high  potassium  lowers  threshold,  high 
calcium  raises  it.  If  you  have  done  your  conditioning  under  one 
set  of  thresholds  of  the  neuron  group,  tlien  the  engram  set  up 
would  be  congruent  with  that  distribution  of  neuron  thresholds 
in  that  neuron  population.  Having  once  established  the  pattern, 
which  is  more  difficult  with  calcium  and  easier  with  potassium, 
you  would  need  the  same  balance  of  thresholds  that  then  existed 
in  order  to  re-evoke  the  engram  in  a  given  assembly  of  cells, 
because  various  cell  thresholds  do  not  change  exactly  in  proportion 
to  the  ion  ratio.  If  learnins;  was  under  the  normal  ion  ratio,  then 


370  Information  Storage  and  Neural  Control 

any  shift  in  ion  balance  would  disturb  it.  It  seems  to  me  this  is 
exactly  what  one  would  expect. 

John:  That  is  one  possible  explanation.  Yet  experiences  learned 
under  normal  circumstances  may  be  retrieved  in  situations  in 
which  it  is  unreasonable  to  argue  that  the  configuration  of  excita- 
bility of  neuron  populations  is  quite  as  it  was  during  the  experience. 
An  alternative  to  your  suggestion  might  be  that  the  altered 
electrolyte  surround  directly  affects  storage  mechanisms  on  the 
molecular  level. 

Gerard:  It  is  a  matter  of  how  it  has  shifted.  These  are  probably 
very  big  shifts,  even  with  small  amounts  of  electrolytes.  You 
remember  the  work  Ochs  did  with  the  Bures  potassium  technique. 
It  is  a  nice  way  of  locating  the  engram,  besides  the  split-brain 
technique.  He  had  rats  learn  a  performance  with  one  hemisphere 
inhibited  with  high  potassium  chloride.  He  removed  this  and  the 
animals  behaved  perfectly  well.  But  sometime  later,  when  he 
blocked  the  other  hemisphere  with  potassium  chloride,  with  the 
first  still  ticking  away  happily,  the  rats  had  no  knowledge  of  what 
they  had  learned.  The  engram  was  in  only  the  part  of  the  brain 
which  was  active  during  learning.  It  is  this  kind  of  an  effect,  I 
think,  that  you  are  dealing  with. 

Let  me  discuss  your  last  question.  Maybe  we  should  not  go 
into  it  because  this  whole  planaria  business,  while  fascinating,  is 
a  bit  off  the  line  of  the  discussion.  You  may  not  know,  though, 
that  your  student  Corning  turned  up  with  Jim  McConnell  and 
reported  the  RNAse  results,  but  could  not  interpret  them.  My  inter- 
pretation, and  I  think  the  one  you  have  used,  seemed  reasonable. 

Let  me  remind  the  group  of  the  basic  experiment.  A  flatworm  is 
trained,  cut  in  two  pieces,  and  the  head  allowed  to  regenerate  a 
tail  and  the  tail  a  head.  Both  new  worms  remember,  as  McConnell 
demonstrated.  Dr.  John  and  his  group  showed,  further,  that  if  each 
of  these  two  parts  is  regenerated  in  RNAse,  the  head  worm  still  re- 
members but  the  tail  worm  does  not.  One  can  explain  this  in 
terms  of  the  fact  that  the  head  worm  has  more  organized  structural 
units  in  it  to  begin  with  and  does  not  have  to  re-create  many  neu- 
rons. Now,  would  something  of  this  kind  apply  to  your  question? 

John:  I  am  sorry.  I  am  talking  about  the  cannibalism  studies. 
One  group  of  planaria  is  fed  shredded,  trained  worms.  Another 


Summary  and  General  Discussion  371 

group  of  planaria  is  fed  shredded,  naive  worms.  On  subsequent 
training,  the  group  which  ate  trained  worms  was  found  to  acquire 
that  conchtioned  response  significantly  more  rapidly  than  the 
group  which  ate  naive  worms.  This  experiment  was  run  blind 
in  our  laboratory,  and  the  results  confirmed  previous  reports  by 
McConnell's  group.  One  is  probably  justified  in  assuming  the 
absence  of  enzymes  in  the  planarian  gut,  which  would  degrade 
macromolecules.  The  reason  I  refer  to  this  work  is  to  ask  what 
sort  of  mechanism  you  would  suggest  to  account  for  these  results. 

Gerard:  WeU,  it  is  even  more  unbelievable  than  the  earlier 
stuff,  but  I  still  tliink  that  what  I  was  saying  was  relevant  to  your 
question.  I  would  have  to  assume  that  these  informed  molecules 
are  not  completely  degraded  in  being  digested  and  absorbed,  and 
so  supply  templates  on  which  the  organized  learning  can  be  based, 
just  as  for  the  tail  regrowing  a  head  with  its  neurons. 

I  wonder  if  we  should  not  let  some  of  tlie  other  people  get  in 
before  pushing  this  one  point. 

Morrell:  I  had  hoped  to  get  a  specific  comment  on  the  plausi- 
bility of  my  suggestion  for  chemical  "protection."  I  wonder 
whether  a  possible  mechanism  for  preservation  of  an  imposed 
shift  in  charge  distribution  might  be  the  bonding  of  the  charged 
moiety  to  phospholipid.  Conceivably  such  bonding  might  not 
only  protect  this  given  molecular  rearrangement,  but  also  fix  it 
to  sites  within  the  membrane  where  influences  on  synaptic  trans- 
mission might  be  expected.  There  is  some  evidence  by  Tobias 
which  indicates  that  axons  treated  with  proteases  continue  to 
conduct  action  potentials  for  many  hours,  while  treatment  with 
lipase  rapidly  abolishes  conduction.  Tobias  (personal  communi- 
cation) has  now  found  that  similar  treatment  with  ribonuclease 
also  impairs  the  capacity  of  the  axon  to  generate  action  potentials. 
Moreover,  there  is  some  preliminary  evidence  from  Dr.  Herzenberg 
(personal  communication)  to  the  effect  that  the  DNA-RNA  speci- 
fication system  may  not  only  regulate  protein  synthesis  but  also 
influence  molecules  containing  phospholipid.  In  fact,  these  lipid 
molecules  are  antigenic  and  thus  conceivably  could  provide  a 
chemical  mechanism  for  cell  recognition. 

Gerard:  I  think  that  is  fine  to  have  on  the  record.  I  had  rather 
not  push  it,  although  I  must  say  that  I  heard  recently  that  Tobias' 


372  Information  Storage  and  Neural  Control 

finding,  which  I  have  also  quoted  with  enthusiasm,  is  under 
question  as  to  whether  the  hpase  at  the  pH  and  ionic  strength 
used  was  acting  on  hpids  or  exhibiting  its  other,  venomlike,  action. 
So  this  may  not  hold. 

Saul  Kit  (Houston,  Texas) :  Dr.  John's  question  is  a  very 
complicated  one.  I  believe  I  would  have  to  discuss  it  with  him  to 
understand  fully  all  of  its  implications.  I  think  we  should  be  very 
careful  in  extrapolating  from  the  molecular  biology  level  to  the 
neurophysiology  frame  of  reference.  I  should  prefer,  therefore,  to 
let  Dr.   Gerard's  answer  stand. 

Gregory  Bateson  (Palo  Alto,  California) :  This  is  changing  the 
subject  somewhat,  but  going  back  to  what  Dr.  Gerard  said  about 
evolution  and  the  relations  between  Lamarckian  and  Darwinian 
theories,  there  are  some  rather  peculiar  problems  in  the  economics 
of  communication  within  the  organism  which  indicate,  at  first 
glance,  that  neither  the  Lamarckian  nor  the  Darwinian  system 
will  work.  Let  me  put  it  this  way.  We  have  an  organism.  We 
describe  it  at  any  given  time  or  over  any  given  finite  time  in  terms 
of  all  necessary  variables  to  define  all  possible  states — Vi,  V2,  .  .  . , 
V„ — perhaps  many  thousands  of  variables.  Any  one  of  these  has 
a  finite  set  of  values.  If  the  organism  exceeds  any  of  these  finite 
thresholds,  it  dies. 

Now  consider  a  pre-girafTe  which  has  the  good  fortune  to  get 
the  mutant  "long  neck"  as  an  item  in  the  genotypic  corpus  of 
genes.  That  genotypic  system  is  not  going  to  tell  the  heart  of  the 
giraffe  that  it  now  has  to  enlarge  in  order  to  supply  the  head  with 
blood.  It  is  not  going  to  deal  with  the  new  problems  of  the  inter- 
vertebral disks.  It  is  not  going  to  solve  all  sorts  of  other  new 
somatic  problems  which,  in  fact,  the  happy  giraffe,  the  lucky 
giraffe,  is  going  to  have  to  deal  with  at  the  somatic  level.  The 
giraffe  is  going  to  have  to  occupy  servo-circuits  within  its  soma 
to  modify  the  size  of  the  heart,  and  so  on.  By  doing  so,  it  has 
reduced  the  finite  set  of  possible  states  of  its  organism. 

Later,  this  pre-giraffe  is  lucky  enough  to  get  another  externally 
adaptive  mutation — let  us  say  big  feet,  which  it  needs  for  kicking  lions. 
It  is  now  again  limited  to  a  subset  of  its  possibilities;  and  if  it  has  to  deal 
with  both  mutations  simultaneously,  it  is  limited  to  that  overlapping 
subset  of  possibilities  which  is  compatible  with  both  mutations. 


Summary  and  General  Discussion  373 

You  see  that  very  soon  a  sequence  of  externally  adaptive  muta- 
tions of  this  sort  is  going  to  lead  to  a  nonviable  giraffe.  It  is  using 
up  its  somatic  flexibility  with  every  adaptive  mutation  that  it  gets. 
The  only  way  it  can  regain  flexibility  is  by  getting  those  mutations 
which  will  enlarge  its  heart  or  do  whatever  is  necessary  to  cope 
with  the  externally  adaptive  changes.  It  has  to  shift  some  of  its 
acc|uired  characteristics  from  the  somatic  servo-systems  to  the 
soldered-in  genotypic  systems. 

The  system  can  only  work  if  there  is  a  comparatively  large 
number  of  mutations  which  will  simulate  a  Lamarckian  process, 
and  evidently  God  set  it  up  this  way  to  deceive  the  Russians. 

Now,  let  us  look  at  the  other  side  of  the  picture.  Suppose  the 
system  were  set  up  on  Lamarckian  lines.  The  genotype  would 
then  have  to  pick  up  from  the  soma  (and  it  is  difficult  enough  to 
imagine  it  picking  up  anything)  those  particular  acquired  charac- 
teristics which  are  the  essential  ones.  But  the  enlarged  heart  is 
not  just  an  enlarged  heart.  It  is  one  item  in  a  general  shift  in  value 
all  around  servo-circuits  to  enlarge  that  heart.  All  those  values 
at  other  points  around  those  circuits  are  going  to  be  picked  up 
in  a  Lamarckian  system,  passed  on  by  inheritance,  and  soldered 
into  the  genotype.  In  fact,  a  Lamarckian  system  will  very  rapidly 
gum  up  the  works  by  decreasing  the  somatic  flexibility  just  as 
badly  as  the  Darwinian  system. 

We  face,  therefore,  an  economics  of  communicational  pathways. 
Evolution  will  only  work  if  you  have  one  system  (the  genotype) 
relatively  independent  of  the  other  (the  somatic),  with  natural 
selection  playing  on  the  whole  thing.  You  have  to  have  a  digital 
genotype,  soldered  in,  with  random  changes,  and  you  have  to 
have  a  system  (the  soma)  of  analogue  operations.  The  soma  is 
being,  so  to  speak,  a  trial  model  to  test  the  genotype.  The  hen  is 
the  egg's  way  of  finding  out  if  it  was  a  good  egg  or  not. 

The  whole  economics  of  the  system  depends  upon  keeping  the 
soma  and  the  genotype  separate.  If  you  are  right  in  saying  that 
cultural  evolution  is  something  much  more  like  a  Lamarckian 
system,  I  think  we  may  look  forward  to  considerable  chaos  in 
the  culture.  The  genotype  is  the  analogue  of  a  legislator.  He  can 
only  afford  to  make  those  changes  which  affirm  changes  that  have 
already  occurred  at  the  somatic  or  popular  level.  If  we  live  in  a 


374  Information  Storage  and  Neural  Control 

Lamarckian  system  in  which  the  lower  levels  are  maximally  able 
to  affect  the  higher  ones,  then  perhaps  we  are  headed  for  chaos. 

Gerard:  I  am  sorry  we  are  talking  about  the  giraffe.  It  seems 
a  camel  would  be  more  appropriate.  As  you  know,  the  definition 
of  a  camel  is  an  animal  made  by  a  committee.  This  seems  to  be 
the  problem  you  are  bothered  about.  I  also  think  that,  in  a  sense, 
it  should  be  a  camel,  because  I  let  his  head  under  the  tent  and 
you  have  brought  the  whole  animal  in.  The  issues  you  are  raising 
are  really  not  too  close  to  the  basic  one  of  the  fixation  of  the  experi- 
ence as  I  was  trying  to  discuss  it.  Let  me  simply  say  this  in  response 
to  these  important  considerations. 

As  you  know,  this  difficulty — the  fact  that  there  must  be  multiple 
changes  that  interact  with  each  other — has  been  recognized  by 
evolutionary  theorists  for  a  long  time.  One  of  the  earliest  criticisms 
of  natural  selection  was  that  it  could  explain  the  survival  of  the 
fittest  but  not  the  arrival  of  the  fittest.  I  think  this  is  partly  what 
you  are  raising.  I  am  certainly  no  expert  in  the  field  of  evolution, 
but  I  have  been  in  close  touch  with  many  of  the  experts  in  this 
field  over  many  years.  They  remain  an  absolutely  solid  phalanx 
on  selection  as  an  adequate  and  satisfactory  mechanism  for 
evolution,  without  bringing  in  Lamarckianism.  Waddington  and 
Dobzhansky  have  recognized  very  clearly  the  fact  that  natural 
selection  has  favored  mutable  genes,  which  is  a  bit  in  your  direction. 

Hyman  Olken  (Livermore,  California):  I  have  one  question 
I  would  like  to  ask  Dr.  Morrell.  Bottley  pointed  out  that  if  you 
increase  the  frequency  of  light  pulses  toward  a  certain  value,  you 
get  increased  response;  then  if  you  increase  beyond  that  frequency, 
the  response  decreases.  Would  that  have  any  effect  on  the  results 
that  you  pointed  out  yesterday  where  you  tested  the  memory  of 
certain  frequencies  and  recovered  other  ones? 

Morrell:  Well,  it  would  have  an  influence  on  the  detectability 
of  any  frequency  in  the  system  with  which  we  were  working.  You 
could  see  from  the  illustrations  that  frequencies  beyond  seven,  say, 
would  gradually  fill  in  the  interval,  and  you  could  not  possibly 
count  a  frequency;  therefore,  it  would  be  undetectable  with  these 
methods. 

Max  E.  Valentinuzzi  (Atlanta,  Georgia):  I  think  that  this 
is  the  appropriate  moment  to  bring  up  three  questions  which  have 


Summary  and  General  Discussion  375 

not  been  answered  as  yet.  They  are  related  to  the  amount  of 
information  necessary  to  transmit  or  to  organize  one  unit  of 
information.  As  you  know,  it  is  not  possible  to  transmit  information 
if  there  is  not  a  previous  amount  of  information  available  as  a 
storage  unit.  So,  the  first  question  is:  How  many  units  of  infor- 
mation do  we  need  as  a  minimum  to  store  one  unit  of  information? 
The  second  question  is:  How  much  energy  is  necessary  to  organize 
one  unit  of  information?  The  third  question  is:  How  much  energy 
is  necessary  to  transmit  from  one  point  to  another  the  same  unit 
of  information? 

Warren  S.  McCulloch  (Cambridge,  Massachusetts):  Do  the 
first  two  questions  amount  to  how  much  information  you  have  to 
have  to  make  another  unit  of  information?  This  is  one  of  the  nasty 
cjuestions  that  is  puzzling  us  at  the  present  time.  There  is  a  way 
of  approaching  it,  but  no  one  is  happy  about  it.  You  cannot  say 
in  a  simple  way,  "How  much  for  unit?";  but  you  can  ask — and 
it  is  the  famous  question  put  by  John  von  Neumann — "How 
much  of  a  computing  machine  do  you  have  to  have  for  that 
computing  machine  to  make  more?"  This  is  the  same  question; 
you  have  the  problem  of  the  generation  of  a  computer,  and  it 
does  not  matter  whether  you  make  it  formally  or  make  it  in  hard- 
ware. The  actual  problem  is  that  of  starting  with  no  form.  This 
means  starting  from  noise,  and  from  noise  it  is  hard  to  get  anything, 
to  generate  any  form.  The  answer  is  that  nobody  knows  how  much 
information  you  need. 

Gerard:  What  about  the  second  question  on  the  energy  for 
transmitting  the  unit  of  information? 

McCulloch:  With  regard  to  the  last  question,  as  small  an 
amount  of  energy  as  you  can  get  in  one  packet  can  carry  one  bit. 
The  limit  is  strictly  that  of  the  physics. 

Kit:  I  wonder  if  this  question  is  not  too  general.  Should  we 
not  be  thinking  about  the  kind  of  information  that  we  are  storing, 
transmitting,  and  replicating?  I  think  estimates  could  be  made 
of  the  amount  of  energy  needed  to  replicate  a  DNxA  molecule. 
Also,  one  can  measure  the  amount  of  energy  consumed  by  a 
bacterial  cell  during  the  replication  of  the  DNA  of  a  phage.  This 
measured  value  will  be  greater  than  the  amount  needed  for  phage 
DNA  synthesis  and   presumably  will   be   an   upper  limit   of  the 


376  Informatwn  Storage  and  Neural  Control 

amount  of  energy  needed.  However,  I  feel  diat  if  we  investigate 
another  information  system,  the  amount  of  energy  required  to 
make  anotlier  unit  of  information  might  be  very  different. 

McCulloch:  Light  does  not  come  in  packets  smaller  than  a 
single  photon,  and  from  Bowman's  figures  one  photon  can  excite. 
That  is  the  lowest  figure  that  anybody  has  and  the  lowest  anyone 
will  ever  have. 

Gerard:  Time  has  gone  on.  It  is  now  my  privilege  and  pleasure, 
since  I  am  acting  as  moderator  at  the  moment,  to  thank  the 
organizers  of  this  symposium,  the  Houston  Neurological  Society 
and  Baylor  University,  the  various  local  people  who  have  been 
kind  to  us,  and,  above  all,  the  speakers  who  have  given  us  such 
interesting  material.  We  are  adjourned. 


APPENDIX  A 

INTRODUCTION 

Michael  H.  Arbib 


"A 


LOGICAL  Calculus  of  the  Ideas  Immanent  in  Nervous 
Activity"  by  Warren  S.  McCulloch  and  Walter  Pitts  is  the  classic 
paper  on  neurophysiological  automata  theory  and  still  merits 
reading  today,  almost  twenty  years  after  its  publication.  Section  I. 
which  gives  the  neurophysiological  basis  for  the  model,  is  still  valid 
in  all  its  essentials  and  remains  the  most  readable  discussion  of  this 
basis.  Section  II,  on  the  theory  of  nets  without  circles,  and  the  dis- 
cussion of  Section  IV  are  equally  excellent. 

However,  Section  III,  the  theory  of  nets  with  circles,  was  only 
intended  as  a  sketchy  account.  It  was  presented  in  Carnap's 
notation,  which  was  not  apt  for  the  task  at  hand,  and  is  incomplete, 
hard  to  read,  and  contains  many  errors.  Hence,  for  this  part  of 
the  theory,  we  advise  the  reader  to  turn  to  more  recent  publications. 
The  theory  of  nets  with  circles  was  first  fully  worked  out  by 
Kleene  (1)  and  has  since  been  given  an  elegant  re-presentation 
by  Copi,  Elgot,  and  Wright  (2).  The  assertions  of  McCulloch  and 
Pitts  concerning  the  connection  between  the  neural  nets  and  Turing 
machines  [Turing  (3)]  have  been  fully  worked  out  by  Arbib  (4). 

REFERENCES 

\.  Kleene,  S.  C:  Representation  of  Events  in  Nerve  Nets  and  Finite 
Automata.  In  Automata  Studies,  ed.  by  C.  E.  Shannon  and  J.  Mc- 
Carthy, Princeton,  Princeton  University  Press,  1956,  p.  3. 

2.  Copi,  I.  M.,  Elgot,  C.  C,  and  Wright,  J.  B.:  Reahzation  of  events  by 

logical  nets.  J.  Assn.  Computing  Mchy.,  5.- 181-1 96,  1958. 

3.  Turing,  A.  M.:  On  computable  numbers,  with  an  application  to  the 

Entscheidungs-problem.    Proc.    London    Math.    Soc.    (2)    ^i.-230-265, 
1936;  with  a  correction,  ibid.,  43:544-546,  1947. 

4.  Arbib,   M.:   Turing  machines,   finite  automata  and   neural  nets.  J. 

Assn.   Computing  Mchy.,   8:461-415,    1961. 

377 


A  LOGICAL  CALCULUS  OF  THE 
IDEAS  IMMANENT  IN  NERVOUS  ACTIVITY* 

Warren  S.  McCulloch  and  Walter  H.  Pitts 

Because  of  the  "all-or-none"  character  of  nervous  activity, 
neural  events  and  the  relations  among  them  can  be  treated  by 
means  of  propositional  logic.  It  is  found  that  the  behavior  of 
everv  net  can  l:)e  described  in  these  terms,  with  the  addition  of 
more  complicated  logical  means  for  nets  containing  circles;  and 
that  for  any  logical  expression  satisfying  certain  conditions,  one 
can  find  a  net  behax'ing  in  the  fashion  it  describes.  It  is  shown 
that  many  particular  choices  among  possible  neurophysiological 
assumptions  are  equivalent,  in  the  sense  that  for  every  net  be- 
having under  one  assumption,  there  exists  another  net  which 
behaves  under  the  other  and  gives  the  same  results,  although 
perhaps  not  in  the  same  time.  Various  applications  of  the  calculus 
are  discussed. 


T. 


INTRODUCTION 


HEORETIClAL  neurophysiology  rests  on  certain  cardinal  as- 
sumptions. The  nervous  system  is  a  net  of  neurons,  each  having  a 
soma  and  an  axon.  Their  adjunctions,  or  synapses,  are  always  be- 
tween the  axon  of  one  neuron  and  the  soma  of  another.  At  any  in- 
stant a  neuron  has  some  threshold,  which  excitation  must  exceed  to 
initiate  an  impulse.  This,  except  for  the  fact  and  the  time  of  its 
occurrence,  is  determined  by  the  neuron,  not  by  the  excitation. 
From  the  point  of  excitation  the  impulse  is  propagated  to  all  parts 
of  the  neuron.  The  velocity  along  the  axon  varies  directly  with  its 
diameter,  from  less  than  one  meter  per  second  in  thin  axons, 
which  are  usually  short,  to  more  than  150  ixieters  per  second  in 
thick  axons,  which  are  usually  long.  The  time  for  axonal  conduc- 
tion is  consequently  of  little  iinportance  in  determining  the  tiine 


*Reprinted  from  The  Bulletin  of  Mathematical  Biophysics,  5:115-133.    1943,  with 
permission  of  the  Editor,  N.  Rashevsky. 

379 


380  Information  Storage  and  Neural  Control 

of  arrival  of  impulses  at  points  unequally  remote  from  the  same 
source.  Excitation  across  synapses  occurs  predominantly  from 
axonal  terminations  to  somata.  It  is  still  a  moot  point  whether 
this  depends  upon  irreciprocity  of  individual  synapses  or  merely 
upon  prevalent  anatomical  configurations.  To  suppose  the  latter 
requires  no  hypothesis  ad  hoc  and  explains  known  exceptions,  but 
any  assumption  as  to  cause  is  compatible  with  the  calculus  to 
come.  No  case  is  known  in  which  excitation  through  a  single  syn- 
apse has  elicited  a  nervous  impulse  in  any  neuron,  whereas  any 
neuron  may  be  excited  by  impulses  arriving  at  a  sufficient  number 
of  neighboring  synapses  within  the  period  of  latent  addition,  which 
lasts  less  than  one  quarter  of  a  millisecond.  Observed  temporal 
summation  of  impulses  at  greater  intervals  is  impossible  for  single 
neurons  and  empirically  depends  upon  structural  properties  of  the 
net.  Between  the  arrival  of  impulses  upon  a  neuron  and  its  own 
propagated  impulse  there  is  a  synaptic  delay  of  more  than  half 
a  millisecond.  During  the  first  part  of  the  nervous  impulse  the 
neuron  is  absolutely  refractory  to  any  stimulation.  Thereafter  its 
excitability  returns  rapidly,  in  some  cases  reaching  a  value  above 
normal  from  which  it  sinks  again  to  a  subnormal  value,  whence 
it  returns  slowly  to  normal.  Frequent  activity  augments  this  sub- 
normality.  Such  specificity  as  is  possessed  by  nervous  impulses 
depends  solely  upon  their  time  and  place  and  not  on  any  other 
specificity  of  nervous  energies.  Of  late  only  inhibition  has  been 
seriously  adduced  to  contravene  this  thesis.  Inhibition  is  the  ter- 
mination or  prevention  of  the  activity  of  one  group  of  neurons  by 
concurrent  or  antecedent  activity  of  a  second  group.  Until  recently 
this  could  be  explained  on  the  supposition  that  previous  activity 
of  neurons  of  the  second  group  might  so  raise  the  thresholds  of 
internuncial  neurons  that  they  could  no  longer  be  excited  by 
neurons  of  the  first  group,  whereas  the  impulses  of  the  first  group 
must  sum  with  the  impulses  of  these  internuncials  to  excite  the 
now  inhibited  neurons.  Today,  some  inhibitions  have  been  shown 
to  consume  less  than  one  millisecond.  This  excludes  internuncials 
and  requires  synapses  through  which  impulses  inhibit  that  neuron 
which  is  being  stimulated  by  impulses  through  other  synapses. 
As  yet  experiment  has  not  shown  whether  the  refractoriness  is 
relative  or  absolute.  We  will  assume  the  latter  and  demonstrate 


.1  Logical  Calculus  of  the  Ideas  Immanent  in  Jservous  Activity  381 

that  tlie  difference  is  immaterial  to  our  argument.  Either  variety 
of  refractoriness  can  be  accounted  for  in  eitlier  of  two  ways.  The 
"inhibitory  synapse"  may  be  of  such  a  kind  as  to  produce  a  sub- 
stance whicii  raises  the  tlireshold  of  the  neuron,  or  it  may  be  so 
placed  that  the  local  chsturbance  prockiced  by  its  excitation 
opposes  the  alteration  induced  by  tlie  otlierwise  excitatory  syn- 
apses. Inasmuch  as  position  is  already  known  to  have  such  effects 
in  the  case  of  electrical  stimulation,  the  first  hypothesis  is  to  be 
excluded  unless  and  until  it  be  substantiated,  for  the  second 
involves  no  new  hypothesis.  We  have,  then,  two  explanations  of 
inhibition  based  on  the  same  general  premises,  differing  only  in 
the  assumed  nervous  nets  and,  consecjuently,  in  the  time  required 
for  inhibition.  Hereafter  we  shall  refer  to  such  nerv'ous  nets  as 
equivalent  in  the  extended  sense.  Since  we  are  concerned  with  properties 
of  nets  which  are  invariant  under  equivalence,  we  may  make  the 
physical  assumptions  which  are  most  convenient  for  the  calculus. 
Many  years  ago  one  of  us,  by  considerations  impertinent  to 
this  argument,  was  led  to  conceive  of  the  I'esponse  of  any  neuron 
as  factually  equivalent  to  a  proposition  which  proposed  its  ade- 
quate stimulus.  He  therefore  attempted  to  record  the  behavior  of 
complicated  nets  in  the  notation  of  the  symbolic  logic  of  proposi- 
tions. The  "all-or-none"  law  of  nervous  activity  is  sufficient  to 
insure  that  the  activity  of  any  neuron  may  be  represented  as  a 
proposition.  Physiological  relations  existing  among  nervous  activ- 
ities correspond,  of  course,  to  relations  among  the  propositions; 
and  the  utility  of  the  representation  depends  upon  the  identity 
of  these  relations  with  those  of  the  logic  of  propositions.  To  each 
reaction  of  any  neuron  there  is  a  corresponding  assertion  of  a 
simple  proposition.  This,  in  turn,  implies  either  some  other  simple 
proposition  or  the  disjunction  or  the  conjunction,  with  or  without 
negation,  of  similar  propositions,  according  to  the  configuration 
of  the  synapses  upon  and  the  threshold  of  the  neuron  in  question. 
Two  difficulties  appeared.  The  first  concerns  facilitation  and  ex- 
tinction, in  which  antecedent  activity  temporarily  alters  responsive- 
ness to  subsequent  stimulation  of  one  and  the  same  part  of  the 
net.  The  second  concerns  learning,  in  which  activities  concurrent 
at  some  previous  time  have  altered  the  net  permanently,  so  that 
a  stimulus  which  would  previously  have  been  inadequate  is  now 


382  Information  Storage  and  Neural  Control 

adequate.  But  for  nets  undergoing  both  alterations,  we  can  sub- 
stitute equivalent  fictitious  nets  composed  of  neurons  whose  con- 
nections and  thresholds  are  unaltered.  But  one  point  must  be 
made  clear:  neither  of  us  conceives  the  formal  equivalence  to  be 
a  factual  explanation.  Per  contra! — we  regard  facilitation  and 
extinction  as  dependent  upon  continuous  changes  in  threshold 
related  to  electrical  and  chemical  variables,  such  as  after-potentials 
and  ionic  concentrations;  and  learning  as  an  enduring  change 
which  can  survive  sleep,  anaesthesia,  convulsions  and  coma.  The 
importance  of  the  formal  equivalence  lies  in  this:  that  the  altera- 
tions actually  underlying  facilitation,  extinction  and  learning  in 
no  way  affect  the  conclusions  which  follow  from  the  formal  treat- 
ment of  the  activity  of  nervous  nets,  and  the  relations  of  the 
corresponding  propositions  remain  those  of  the  logic  of  propositions. 
The  nervous  system  contains  many  circular  paths,  whose  ac- 
tivity so  regenerates  the  excitation  of  any  participant  neuron  that 
reference  to  time  past  becomes  indefinite,  although  it  still  implies 
that  afferent  activity  has  realized  one  of  a  certain  class  of  con- 
figurations over  time.  Precise  specification  of  these  implications 
by  means  of  recursive  functions,  and  determination  of  those  that 
can  be  embodied  in  the  activity  of  nervous  nets,  completes  the 
theory. 

THE  THEORY:  NETS  WITHOUT  CIRCLES 

We  shall  make  the  following  physical  assumptions  for  our  cal- 
culus. 

1.  The  activity  of  the  neuron  is  an  ^'all-or-none"  process. 

2.  A  certain  fixed  number  of  synapses  must  be  excited  within 
the  period  of  latent  addition  in  order  to  excite  a  neuron  at  any 
time,  and  this  number  is  independent  of  previous  activity  and 
position  on  the  neuron. 

3.  The  only  significant  delay  within  the  nervous  system  is  syn- 
aptic delay. 

4.  The  activity  of  any  inhibitory  synapse  absolutely  prevents 
excitation  of  the  neuron  at  that  time. 

5.  The  structure  of  the  net  does  not  change  with  time. 


A  Logical  Calculus  of  the  Ideas  Immanent  in  jYervous  Activity  383 

To  present  the  theory,  the  most  appropriate  symbolism  is  that 
of  Language  II  of  R.  Carnap  (1938),  augmented  with  various 
notations  drawn  from  B.  Russell  and  A.  N.  Whitehead  (1927) 
including  the  Pnncipia  conventions  for  dots.  Typographical  neces- 
sity, however,  will  compel  us  to  use  the  upright  'E'  for  the  existen- 
tial operator  instead  of  the  inverted,  and  an  arrow  ('-^')  for 
implication  instead  of  the  horseshoe.  We  shall  also  use  the  Carnap 
syntactical  notations,  but  print  them  in  boldface  rather  than 
German  type;  and  we  shall  introduce  a  functor  S,  whose  value 
for  a  property  P  is  the  property  which  holds  of  a  number  when  P 
holds  of  its  predecessor;  it  is  defined  by  'S{P)  (/) .  =  .  P(A'.v)  .  /  =  .v')'; 
the  brackets  around  its  argument  will  often  be  omitted,  in  which 
case  this  is  understood  to  be  the  nearest  predicate-expression  [Pr] 
on  the  right.  Moreover,  we  shall  write  S-Pr  for  S{S{Pr)),  etc. 

The  neurons  of  a  given  net  '^  may  be  assigned  designations 
'^I'j  '^2',  .  .  .  ,  'c„'.  This  done,  we  shall  denote  the  property  of  a 
number,  that  a  neuron  c,  fires  at  a  time  which  is  that  number  of 
synaptic  delays  from  the  origin  of  time,  by  ^A^'  with  the  numeral 
i  as  subscript,  so  that  N ,{t)  asserts  that  c,  fires  at  the  time  t.  N,  is 
called  the  action  of  c,.  We  shall  sometimes  regard  the  subscripted 
numeral  of  '  N'  as  if  it  belonged  to  the  object-language,  and  were 
in  a  place  for  a  functoral  argument,  so  that  it  might  be  replaced 
by  a  number-variable  [z]  and  quantified;  this  enables  us  to  abbre- 
viate long  but  finite  disjunctions  and  conjunctions  by  the  use  of 
an  operator.  We  shall  employ  this  locution  quite  generally  for 
sequences  of  Pr\  it  may  be  secured  formally  by  an  obvious  dis- 
junctive definition.  The  predicates  '.Vi',  '.V^',  .  .  .  ,  comprise  the 
syntactical  class  '  N\ 

Let  us  define  the  peripheral  afferents  of  V)!  as  the  neurons  of  ^^I 
with  no  axons  synapsing  upon  them.  Let  N,,  . .  .  ^  N^  denote  the 
actions  of  such  neurons  and  A^,,+i,  N,^,,  . .  .  ,  N„  those  of  the  rest. 
Then  a  solution  of  VX  will  be  a  class  of  sentences  of  the  form  S- 
A^p+i  (21)  .  ^  .  Pr,  {N,,  N,,  ...  ,  N„  2i),  where  Pr,  contains  no 
free  variable  save  Zi  and  no  descriptive  symbols  save  the  A''  in  the 
argument  [Arg],  and  possibly  some  constant  sentences  [sa];  and 
such  that  each  S,  is  true  of  VX.  Conversely,  given  a  Pvi  {^i,  ^2 
^p\,  Zi,  s),  containing  no  free  variable  save  those  in  its  Arg,  we 
shall  say  that  it  is  realizable  in  the  narrow  sense  if  there  exists  a  net  9l 


384  Information  Storage  and  Neural  Control 

and  a  series  of  A^,  in  it  such  that  M  (zi)  .  =  .  Pr^  (M,  A^o,  ••  •  , 
Zi,  sai)  is  true  of  it,  where  sax  has  the  form  A^(0).  We  shall  call  it 
realizable  in  the  extended  sense,  or  simply  realizable,  if  for  some  n  S"{Pri) 
ipi,  •  ...  pp.  Zu  s)  is  realizable  in  the  above  sense.  Cp,  is  here  the 
realizing  neuron.  We  shall  say  of  two  laws  of  nervous  excitation 
which  are  such  that  every  S  which  is  realizable  in  either  sense 
upon  one  supposition  is  also  realizable,  perhaps  by  a  different 
net,   upon   the   other,   that   they   aie   equivalent   assumptions,   in 

that  sense. 

The  following  theorems  about  realizability  all  refer  to  the  ex- 
tended   sense.    In    some    cases,    sharper    theorems    about    narrow 
realizability  can  be  obtained;   but   in   addition   to  greater  com- 
plication in  statement  this  were  of  little  practical  value,  since  our 
present  neurophysiological  knowledge  determines  the  law  of  ex- 
citation   only    to    extended    equivalence,    and    the    more    precise 
theorems  differ  according  to  which  possible  assumption  we  make. 
Our  less  precise  theorems,  however,  are  invariant  under  equiva- 
lence, and  are  still  sufficient  for  all  purposes  in  which  the  exact 
time  for  impulses  to  pass  through  the  whole  net  is  not  crucial. 
Our  central  problems  may  now  be  stated  exactly:  first,  to  find 
an  effective  method  of  obtaining  a  set  of  computable  S  constituting 
a  solution  of  any  given  net;  and  second,  to  characterize  the  class 
of  realizable    S   in    an    effective   fashion.    Materially   stated,    the 
problems  are  to  calculate  the  behavior  of  any  net,  and  to  find  a 
net  which  will  behave  in  a  specified  way,  when  such  a  net  exists. 
A  net  will  be  called  cyrlic  if  it  contains  a  circle:  i.e.,  if  there 
exists  a  chain  c„  C/+i  ,  ...  of  neurons  on  it,  each  member  of  the 
chain  synapsing  upon  the  next,  with  the  same  beginning  and  end. 
If  a  set  of  its  neurons  Ci  ,  c-i  ,  . .  .  ,  Cp  is  such  that  its  removal  from 
VX  leaves  it  without  circles,  and  no  smaller  class  of  neurons  has  this 
property,  the  set  is  called  a  cj>clic  set,  and  its  cardinality  is  the 
order  o/vX.  In  an  important  sense,  as  we  shall  see,  the  order  of  a 
net  is  an  index  of  the  complexity  of  its  behavior.  In  particular, 
nets   of  zero   order   have   especially   simple   properties;   we   shall 
discuss  them  first. 

Let  us  define  a  temporal  propositional  expression  (a  TPE),  desig- 
nating a  temporal  propositional  function  {TPF),  by  the  following 
recursion: 


A^, 


A  Logical  Calculus  of  the  Ideas  Immanent  in  Nervous  Activity  385 

1.  A^p^  [zi]  is  a  TPE,  where  Pi  is  a  predicate-variable. 

2.  If  Si  and  So  are  TPE  containing"  the  same  free  individual 
variable,  so  are  SS\,  SivSo,  Si.S->  and  S,.  ^-^  S2. 

3.  Nothing  else  is  a  TPE. 

Theorem  I 

Every  net  of  order  0  can  be  solved  in  terms  of  temporal  propositional 
expressions. 

Let  Ci  be  any  neuron  of  V^l  with  a  threshold  6,  >  0,  and  let  Cn, 
Ci2,  .  ••  ,  (',p  have  respectively  //,i,  '?,2,  ••  •  ,  n,p  excitatory  synapses 
upon  it.  Let  Cj],  r,2,  •  •  •  ,  ^'jy  have  inhibitory  synapses  upon  it. 
Let  Ki  be  the  set  of  the  subclasses  of  \n,i,  n,2,  ••  •  ,  fi,p\  such  that 
the  sum  of  their  members  exceeds  6,.  We  shall  then  be  able  to 
write,  in  accordance  with  the  assumptions  mentioned  above, 

where  the  'E'  ^'^^  'H'  are  syntactical  symbols  for  disjunctions 
and  conjunctions  which  are  finite  in  each  case.  Since  an  expression 
of  this  form  can  be  written  for  each  C;  which  is  not  a  peripheral 
afferent,  we  can,  by  substituting  the  corresponding  expression  in 
(1)  for  each  A''^,,,  or  A'',,-  whose  neuron  is  not  a  peripheral  afferent, 
and  repeating  the  process  on  the  result,  ultimately  come  to  an 
expression  for  A^,  in  terms  solely  of  peripherally  afferent  A^,  since 
^^l  is  without  circles.  Moreover,  this  expression  will  be  a  TPE, 
since  obviously  (1)  is;  and  it  follows  immediately  from  the  definition 
that  the  result  of  substituting"  a  TPE  for  a  constituent  p{z)  in  a 
TPE  is  also  one. 

Theorem  II 

Every  TPE  is  realizable  by  a  net  of  order  zero. 

The  functor  .9  obviously  coi"nmutes  with  disjunction,  conjunction, 
and  negation.  It  is  obvious  that  the  result  of  substituting  any  S,, 
realizable  in  the  narrow  sense  (i.n.s.),  for  the  p{z)  in  a  realizable 
expression  Si  is  itself  realizable  i.n.s.;  one  constructs  the  realizing 
net  by  replacing  the  peripheral  afferents  in  the  net  for  Si  by  the 
realizing"  neurons  in   the  nets  for   the    Si.   The   one   neuron   net 


386  Information  Storage  and  Neural  Control 

realizes  p\{z\)  i.n.s.,  and  Figure  1-a  sliows  a  net  tliat  realizes 
Spi{zi)  and  hence  SS-i,  i.n.s.,  if  So  can  be  realized  i.n.s.  Now  if 
So  and  S3  are  realizable  then  S"'S-2.  and  S"Sz  are  realizable  i.n.s., 
for  suitable  m  and  n.  Hence  so  are  S^'^'^So  and  ^''""'""Sa.  Now  the 
nets  of  Figures  lb,  c  and  d  respectively  realize  S{pi{zi)\ p2{z\)), 
S{pi{zi)  .  p2{zx)),  and  S\pi{z,)  .  ~  poiz,))  i.n.s.  Hence  S'-+"+'  (SiV 
S2),  ^"'+"+1  (Si  .  So),  and  ^''«+"+i  (Si  .  ~  So)  are  realizable  i.n.s. 
Therefore  Si  v  SoSi  .  SoSi  .  ~  So  are  realizable  if  Si  and  So  are. 
By  complete  induction,  all  TPE  are  realizable.  In  this  way  all 
nets  may  be  regarded  as  built  out  of  the  fundamental  elements 
of  Figures  la,  b,  c,  d,  precisely  as  the  temporal  propositional  ex- 
pressions are  generated  out  of  the  operations  of  precession,  dis- 
junction, conjunction,  and  conjoined  negation.  In  particular, 
corresponding"  to  any  description  of  state,  or  distribution  of  the 
values  true  and  false  for  the  actions  of  all  the  neurons  of  a  net  save 
that  which  makes  them  all  false,  a  single  neuron  is  constructible 
whose  firing  is  a  necessary  and  sufficient  condition  for  the  validity 
of  that  description.  Moreover,  there  is  always  an  indefinite  number 
of  topologically  different  nets  realizing  any  TPE. 

Theorem  III 

Let  there  be  given  a  complex  sentence  Si  built  up  in  any  manner  out 
of  elementary  sentences  of  the  form  p(zi  —  zz)  where  zz  is  any  numeral, 
by  ary  of  the  propositional  connections:  negation,  disfunction,  conjunction, 
implication,  and  equivalence.  Then  Si  is  a  TPE  and  only  if  it  is  false 
when  its  constituent  p(zi  —  zz)  are  all  assumed  false — i.e.,  replaced 
by  false  sentences — or  that  the  last  line  in  its  truth-table  contains  an 
'F', — or  there  is  no  term  in  its  Hilbert  disjunctive  normal  form  com- 
posed exclusively  of  negated  terms. 

These  latter  three  conditions  are  of  course  equivalent  (Hilbert 
and  Ackermann,  1938).  We  see  by  induction  that  the  first  of  them 
is  necessary,  since  p{zi  —  zz)  becomes  false  when  it  is  replaced 
by  a  false  sentence,  and  Si  v  So,  Si  .  S2  and  Si  .  ~  S2  are  all 
false  if  both  their  constituents  are.  We  see  that  the  last  condition 
is  sufficient  by  remarking  that  a  disjunction  is  a  TPE  when  its 
constituents  are,  and  that  any  term 

Si  .  So  .  .  .  .  Sm  .  -^  S,„+i  .  '^  .  .  .  .  -^  s„ 
can  be  written  as 


A  Logical  Calculus  of  the  Ideas  Immanent  in  Nervous  Activity  387 

(Si    .    So    ...     .    S„0    .    ~    {Sm+xV   S,n  +  lV    .     .     .     .V   S„), 

which  is  clearly  a    TPE. 

The  method  of  the  last  theorems  does  in  fact  provide  a  very 
convenient  and  workable  procedure  for  constructing  nervous  nets 
to  order,  for  those  cases  where  there  is  no  reference  to  events 
indefinitely  far  in  the  past  in  the  specification  of  the  conditions. 
By  way  of  example,  we  may  consider  the  case  of  heat  produced 
by  a  transient  cooling. 

If  a  cold  object  is  held  to  the  skin  for  a  moment  and  removed, 
a  sensation  of  heat  will  be  felt;  if  it  is  applied  for  a  longer  time,  the 
sensation  will  be  only  of  cold,  with  no  preliminary  warmth,  how- 
ever transient.  It  is  known  that  one  cutaneous  receptor  is  affected 
by  heat,  and  another  by  cold.  If  we  let  Ni  and  A^2  be  the  actions 
of  the  respective  receptors  and  N?.  and  A^4  of  neurons  whose 
activity  implies  a  sensation  of  heat  and  cold,  our  requirements 
may  be  written  as 

N^{t)  :  =  :  A'i(/-1)  .  v  .  N^.{t-^)  .  ^N~,{t-2) 

Ndt)  .  =  .No(t-2)  .No(t-l) 

where  we  suppose  for  simplicity  that  the  required  persistence  in 
the  sensation  of  cold  is,  say,  two  synaptic  delays,  compared  with 
one  for  that  of  heat.  These  conditions  clearly  fall  under  Theorem 
III.  A  net  may  consequently  be  constructed  to  realize  them,  by 
the  method  of  Theorem  II.  We  begin  by  writing  them  in  a  fashion 
which  exhibits  them  as  built  out  of  their  constituents  by  the 
operations  realized  in  Figures  la,  b,  c,  d:  i.e.,  in  the  form 

N^(t)  .  ^  .  S{A\it)  V  S[{SN,{t))  >'^N,(t)]} 
N,(t)  .  ^  .  S{[SN,{t)]  .N,{t)]. 

First  we  construct  a  net  for  the  function  enclosed  in  the  greatest 
number  of  brackets  and  proceed  outward;  in  this  case  we  run  a 
net  of  the  form  shown  in  Figure  la  from  Co  to  some  neuron  r„,  say, 
so  that 

Nait)  .   =   .  SN,(t). 
Next  introduce  two  nets  of  the  forms   Ic  and   Id,  both  running 
from  Ca  and  c^,  and  ending  respectively  at  Ci  and  say  Cb.  Then 

A^4(0  .  =  .  S[NAt)  .  N,it)]  .  ^  .  S[(SN2(t))  .  N,(t)]. 


388  Information  Storage  and  Neural  Control 

Finally,  run  a  net  of  the  form  lb  from  C\  and  Cb  to  fs,  and  derive 

.¥3(0  .  ^  .  .S[.Vi(Ov.V,(0] 

.  ^  .  .StVi(0  v.S'[GSWo(0)  .  ~A'2(0]1. 

These  expressions  for  N z{t)  and  iV4(/)  are  the  ones  desired;  and 
the  realizing  net  in  toto  is  shown  in  Figure  le. 

This  illusion  makes  very  clear  the  dependence  of  the  correspond- 
ence between  perception  and  the  "external  world''  upon  the 
specific  structural  properties  of  the  intervening  nervous  net.  The 
same  illusion,  of  course,  could  also  have  been  produced  under 
various  other  assumptions  about  the  behavior  of  the  cutaneous 
receptors,  with  correspondingly  different  nets. 

We  shall  now  consider  some  theorems  of  equivalence:  i.e., 
theorems  which  demonstrate  the  essential  identity,  save  for  time, 
of  various  alternative  laws  of  nervous  excitation.  Let  us  first  dis- 
cuss the  case  of  relative  inhibition.  By  this  we  mean  the  supposition 
that  the  firing  of  an  inhibitory  synapse  does  not  absolutely  prevent 
the  firing  of  the  neuron,  but  merely  raises  its  threshold,  so  that 
a  greater  number  of  excitatory  synapses  must  fire  concurrently 
to  fire  it  than  would  otherwise  be  needed.  We  may  suppose,  losing 
no  generality,  that  the  increase  in  threshold  is  unity  for  the  firing 
of  each  such  synapse;  we  then  have  the  theorem: 

Theorem  IV 

Relative  and  absolute  inhibition  are  equivalent  in  the  extended  sense. 

We  may  write  out  a  law  of  nervous  excitation  after  the  fashion 
of  (1),  but  employing  the  assumption  of  relative  inliibition  instead; 
inspection  then  shows  that  this  expression  is  a  TPE.  An  example 
of  the  replacement  of  relative  inhibition  by  absolute  is  given  by 
Figure  If.  The  reverse  replacement  is  even  easier;  we  give  the 
inhibitory  axons  afferent  to  c,  any  sufficiently  large  number  of 
inhibitory  synapses  apiece. 

Second,  we  consider  the  case  of  extinction.  We  may  write  this 
in  the  forni  of  a  variation  in  the  threshold  6,  after  the  neuron  Ct 
has  fired;  to  the  nearest  integer — and  only  to  this  approximation 
is  the  variation  in  threshold  significant  in  natural  forms  of  excita- 
tion— this  may  be  written  as  a  sequence  di  +  bj  for  j  synaptic 


A  Logical  Calculus  of  the  Ideas  Immanent  in  Nervous  Activity  389 

delays  after  firing,  where  bj  =  0  for  /  large  enough,  say  7  =  M  or 
greater.  We  may  then  state 

Theorem  V 

Extinction  is  equivalent  to  absolute  inhibition. 

For,  assuming  relative  inhibition  to  hold  for  the  moment,  we 
need  merely  run  M  circuits  U\,  U'2,  . .  .  'hi  containing  respectively 
1,  2,  ...  ,  A/  neurons,  such  that  the  firing  of  each  link  in  any  is 
sufficient  to  fire  the  next,  from  the  neuron  c,  back  to  it,  where 
the  end  of  the  circuit  Wj  has  just  b,-  inhibitory  synapses  upon  c,. 
It  is  evident  that  this  will  produce  the  desired  results.  The  reverse 
substitution  may  be  accomplished  by  the  diagram  of  Figure  Ig. 
From  the  transitivity  of  replacement,  we  infer  the  theorem.  To 
this  group  of  theorems  also  belongs  the  well-known 

Theorem  VI 

Facilitation  and  temporal  summation  may  be  replaced  by  spatial  sum- 
mation. 

This  is  obvious:  one  need  merely  introduce  a  suitable  secjuence 
of  delaying  chains,  of  increasing  numbers  of  synapses,  between  the 
exciting  cell  and  the  neuron  whereon  temporal  summation  is 
desired  to  hold.  The  assumption  of  spatial  summation  will  then 
give  the  required  results.  See  e.g.  Figure  Ih.  This  procedure  had 
application  in  showing  that  the  observed  temporal  summation  in 
gross  nets  does  not  imply  such  a  mechanism  in  the  interaction  of 
individual  neurons. 

The  phenomena  of  learning,  which  arc  of  a  character  persisting 
over  most  physiological  changes  in  nervous  activity,  seem  to  re- 
quire the  possibility  of  permanent  alterations  in  the  structure  of 
nets.  The  simplest  such  alteration  is  the  formation  of  new  synapses 
or  equivalent  local  depressions  of  threshold.  We  suppose  that  some 
axonal  terminations  cannot  at  first  excite  the  succeeding  neuron; 
but  if  at  any  time  the  neuron  fires,  and  the  axonal  terminations 
are  simultaneously  excited,  they  become  synapses  of  the  ordinary 
kind,  henceforth  capable  of  exciting  the  neuron.  The  loss  of  an 
inhibitory  synapse  gives  an  entirely  equivalent  result.  We  shall 
then  have 


390  Information  Storage  and  Neural  Control 

Theorem  VII 

Alterable  synapses  can  be  replaced  by  circles. 

This  is  accomplished  by  the  method  of  Figure  li.  It  is  also  to 
be  remarked  tliat  a  neuron  which  becomes  and  remains  spon- 
taneously active  can  likewise  be  replaced  by  a  circle,  which  is  set 
into  activity  by  a  peripheral  afferent  when  the  activity  commences, 
and  inhibited  by  one  when  it  ceases. 

THE  THEORY:   NETS  WITH  CIRCLES 

The  treatment  of  nets  which  do  not  satisfy  our  previous  assump- 
tion of  freedom  from  circles  is  very  much  more  difficult  than  that 
case.  This  is  largely  a  consequence  of  the  possibility  that  activity 
may  be  set  up  in  a  circuit  and  continue  reverberating  around  it 
foi  an  indefinite  period  of  time,  so  that  the  realizable  Pr  may 
involve  reference  to  past  events  of  an  indefinite  degree  of  remote- 
ness. Consider  such  a  net  VX,  say  of  order  /;,  and  let  Ci,  c-2,  . .  .  ,  r^  be 
a  cyclic  set  of  neurons  of  ^^l.  It  is  first  of  all  clear  from  the  definition 
that  every  N^  of  ^^\  can  be  expressed  as  a  TPE,  of  M,  A^2,  . .  .  ,  Np 
and  the  absolute  afferents;  the  solution  of  v)l  involves  then  only 
the  determination  of  expressions  for  the  cyclic  set.  This  clone,  we 
shall  derive  a  set  of  expressions  [A\: 

Ndz,)  .  ^  .  Pr.[S""  M(zi),  *S"''-^  N,(z,),  ...  ,  S"''^  N,{z,)],         (2) 

where  Pr ,  also  involves  peripheral  afferents.  Now  if  n  is  the  least 
common  multiple  of  the  n,„  we  shall,  by  substituting  their  equiva- 
lents according  to  (2)  in  (3)  for  the  A^„  and  repeating  this 
process  often  enough  on  the  result,  obtain  S  of  the  form 

N,{z,)  .  ^  .  Pr,[S"N,{zr),  S"N,(zi),  ...  ,  S" Np(z,)].  (3) 

These  expressions  may  be  written  in  the  Hilbert  disjunctive  nor- 
mal form  as 

N,{z,)  .  =  .  E  S„ n  '?"  A^. (zi )  n  ~  *^"  ^i(2i),  for  suitable  ^  (-i) 

where  S„  is  a  TPE  of  the  absolute  afferents  of  V^I.  There  exist 
some  2"  different  sentences  formed  out  of  the  pN,  by  conjoining 
to  the  conjunction  of  some  set  of  them  the  conjunction  of  the 


A  Logical  Calculus  of  the  Ideas  Immanent  in  Nervous  Activity  391 

negations  of  the  rest.  Denumerating  these  by  A'i(zi),  ^'2(21),  .  .., 
X-2p{zi),  we  may,  by  use  of  the  expressions  (4),  arrive  at  an  equi- 
pollent set  of  equations  of  the  form 

X,{z,)  .  ^  .ZPruiz,)  .  S^Xjiz,).  (5) 

Now   we   import   the   subscripted    numerals   i,j   into   the  object- 
language:  i.e.,  define  Pri  and  Pr2  such  that  Pri(zzi,Zi)  .   =  .   X,{zi) 
and  Prj(zzi,zz2,Zi)  .   =  .  Pr,j{zi)   are  provable  whenever  zzi  and 
ZZ2  denote  i  and  /  respectively. 
Then  we  may  rewrite  (5)  as 

(zi)zzp  :  Pri(zi,  Z3) 
.  =  .  {EZ'{)zZp  .  Pr-iizi.  Zo,  z-i  -  zzn)  .  Priizo,  Zz  -  zZn)  (6) 

where  zz^  denotes  n  and  zZp  denotes  2'\  By  repeated  substitution 
we  arrive  at  an  expression 

(zi)zZp  :  Pri(zi,  zz„zzo)  .  =  .  {Ez-z)zZp  {Ezz)zZp  .  .  .  {Ez„)zZp 

Pr-zizi,  z-2,  ZZn  {zzo  —  1))  .  Pr-i{z2,Zz,zZn  {zz2  -  1)) (7) 

Pr2(z„_i,z„,0)  .  Pri(Zn,0),   for   any   numeral   ZZ2   which  denotes  s. 

This  is  easily  shown  by  induction  to  be  equipollent  to 

{zi)zzp  :  .  Pri{Zi,zZnZZ2)  :  =  :  (Ef)  (Z2)  zzo  —  l/(zoZZ„) 

^  ZZp  .  fiZZnZZ2)    =  Zi   .    Pr2{f(,ZZn   (Z2  +   1)),  (8) 

f(zz,a-z))  .  PrAf {0),0) 
and  since  this  is  the  case  for  all  ZZ2,  it  is  also  true  that 

(Z4)  {z,)zzp  :  Pn{z,,z,)  .  =  .  (Ef)  (Z2)  (Z4  -  1)  ./(Z2) 

^  ZZp  .  /(Z4)  =  Zi/(Z4)  -  zi  .  Pro[/(z2  +  1),/(Z2),  Z2]  .  (9) 

Pri[/(res  (Z4.  zZn)),  res  (Z4,  zz,,)], 

where  zz„  denotes  n,  res   {r,s)  is  the  residue  of  /•  mod  s  and  zZp 
denotes  2''.  This  may  be  written  in  a  less  exact  way  as 

N^t)  .  ^  .  (Ecf>)  ix)t  -  1  .  <^(.r)  ^  2'  .  0(0  =  i  . 

P[0(.f+  l),0(.r)  ..V,(o^  (0)], 

where  a  and  t  are  also  assumed  divisible  by  n,  and  Pr2  denotes  P. 
From  the  preceding  remarks  we  shall  have 


392  Information  Storage  and  Neural  Control 

Theorem  VIII 

The  expression  (9)  for  neurons  of  the  cyclic  set  oj  a  net  S'X  together 
with  certain  TPE  expressing  the  actions  oJ  other  neurons  in  terms  oj 
them,  constitute  a  solution  of  V)I. 

Consider  now  the  question  of  the  reahzabihty  of  a  set  of  S,.  A 
first  necessary  condition,  demonstrable  by  an  easy  induction,  is  that 

(z.2)zi  .  pAz-2)  ^  p,{z,)  .^.Si^  sMj  (10) 

should  be  true,  with  similar  statements  for  the  other  free  p  in  Si'. 
i.e.,  no  nervous  net  can  take  account  of  future  peripheral  afferents. 
Any  S,  satisfying  this  requirement  can  be  replaced  by  an  equi- 
pollent S  of  the  form 

{Ef)  (z,)zy  {z,)zz,r.hPr,„, 

:f{Zr,Z,,Zs     =   1  .   ^   ./>.3(Z2)  (11) 

where  zZp  denotes  p,  by  defining 

Pr„,i  =  /[(zi)  {Z2)zi{zs)zzp  :  .  f(zu  z-i,  Zs)  =  0  .  v  .  /(zi,  Zo,  Zs) 
=  1  :/(zi,  Zo,  Z3)  =  1  .  =  .  /),3(z,)  :  -^  :  S,]. 

Consider  now  these  series  of  classes  a,,  for  which 

N ,{V)  :  =  :  {E<\>)  {.v)t(^m)q  :  4>ecxi  :ISf„,{x)  .  =  .  <i>{t,  x,  m)  =  1. 

[/  =  ry +  !,•••  ,M]  (12) 

holds  for  some  net.  These  will  be  called  prehensible  classes.  Let  us 
define  the  Boolean  ring  generated  by  a  class  of  classes  k  as  the 
aggregate  of  the  classes  which  can  be  formed  from  members  of  k 
by  repeated   application   of  the  logical   operations;   i.e.,   we  put 

-^  aeX  :  a,  ^eX  .  — >  .   —  a,  a  .  (3,  aW  jSeX]. 

We  shall  also  define 

^(k)  .  =  .  (R(k)  -  t'p'  -  "'V', 

f-i\e(K)     =p  X[(a,  /3)  :  atK  -^  ae\  .  ^  .   —  a,  a  .  (3,  aV  (3,  S  "aeX 

and 

G{'\>,t)  =  i[{m)  .  cf>{t  -\-  l,t,  m)  =   '!^(m)]. 


A  Logical  Calculus  of  the  Ideas  Immanent  in  Nervous  Activity  393 

The  class  !-iv,,(/c)  is  formed  from  k  in  analogy  with  H\(>'),  but  by 
repeated  apphcation  not  only  of  the  logical  operations  but  also 
of  that  which  replaces  a  class  of  properties  P  e  a  by  S{P)  e  S  ^^  a. 
We  shall  then  have  the 

Lemma 

Priipu  Pi.  • .  .  ,  p.,.  Zi)  is  a   TPE  if  and  only  if 

(Zl)     (pu    ...    ,  pra)     {Ep„,  +  i)    :  />,„+!   e  ir^^eilpl,  p2,    •  •  •    ,  P,n]  ) 

A„+i(zi)  =  PuiPuPi,  ...  ,A»,Zl)  (13) 

is  true;  and  it  is  a  TPE  not  involving  \S"  if  and  only  if  this  holds 
when  '<-R,.'  is  replaced  by  'f-R',  and  we  then  obtain 

Theorem  IX 

A  series  of  classes  ai,  a-^,  ...  a,  is  a  series  of  f)rehensible  classes  if  and 
only  if 

(Em)  (En)  (p)n(i)  ('V)  :  .  i.r)ni';^ix)  =  Ov  •b{x    =  1  :^  :  (E^) 
{Ey)m  .  'M^)  =  0  .  fSeiillyiiEi)  .  y  =  a,))  .  v  .  {x)m  . 
^(.r)  =  0  .  l3efk[yaE,)  .  y  =  «,)]  :  (0  (0)  :  ^ea.  .  (14) 

'i4>,  nt  +  p)  .  ^  .  (Ef)  .  fef3  .  {w)m{.v.)t  -  1  . 

(t){n{t  +  1)  +  p,  nx  +  p,  iv)  =  f(nt  +  p,  nx  +  p,  iv). 

The  proof  here  follows  directly  from  the  lemma.  The  condition 
is  necessary,  since  every  net  for  which  an  expression  of  the  form 
(4)  can  be  written  obviously  verifies  it,  the  t];'s  being  the  charac- 
teristic functions  of  the  S„  and  the  (3  for  each  -^  being  the  class 
whose  designation  has  the  form  JJ  ^r,  J  J  PTj,  where    Pr,,  denotes 

I'-i  J-4i,-, 

a,,  for  all  k.  Conversely,  we  may  write  an  expression  of  the  form 
(4)  for  a  net  VX  fulfilling  prehensible  classes  satisfying  (14)  by  putting 
for  the  Pra  Pr  denoting  the  ']j's  and  a  Pr,  written  in  the  analogue 
for  classes  of  the  disjunctive  normal  form,  and  denoting  the  a 
corresponding  to  that  '4^,  conjoined  to  it.  Since  every  S  of  the  form 
(4)  is  clearly  realizable,  we  have  the  theorem. 

It  is  of  some  interest  to  consider  the  extent  to  which  we  can 
by  knowledge  of  the  present  determine  the  whole  past  of  various 
special  nets:  i.e.,  when  we  may  construct  a  net  the  firing  of  the 
cyclic  set  of  whose  neurons  requires  the  peripheral  afferents  to 


394  Information  Storage  and  Neural  Control 

have  had  a  set  of  past  values  specified  by  given  functions  0,.  In 
this  case  the  classes  a,  of  the  last  theorem  reduced  to  unit  classes; 
and  the  condition  may  be  transformed  into 

{Em,  n)  {v)n{i,  <];)  {Ej)  :  .  {x)m  :  '\>{x)  =  0  .  w  ,  '^(x)  =  I  : 
^i€j(({;,  nt  +  p)  :  -^  :  (:w)m(x)t  —  1  .  (f)i{n{t  +  1) 
+  p,  nx  -\-  'p,w)  =  4>j{nt  +  p,  nx  +  p,  w)  :  . 
(:u,  v)  (w)m  .  4>iix>'{u  +  1)  -\-  p,  nn  +  p,  w) 
=  (t)i{n(v  +  1)  -]r  p,nv  -{-  p,w). 

On  account  of  limitations  of  space,  we  have  presented  the  above 
argument  very  sketchily;  we  propose  to  expand  it  and  certain  of 
its  implications  in  a  further  publication. 

The  condition  of  the  last  theorem  is  fairly  simple  in  principle, 
though  not  in  detail;  its  application  to  practical  cases  would, 
however,  require  the  exploration  of  some  2-"  classes  of  functions, 
namely  the  members  of  fjv(|ai,  •••  ,  «..j).  Since  each  of  these  is 
a  possible  ^  of  Theorem  IX,  this  result  cannot  be  sharpened.  But 
we  may  obtain  a  sufficient  condition  for  the  realizability  of  an  S 
which  is  very  easily  applicable  and  probably  covers  most  practical 
purposes.  This  is  given  by 

Theorem  X 

Let  us  define  a  set  of  X"  of  S  by  the  following  recursion: 

1.  Any  TPE  and  any  TPE  whose  arguments  have  been  re- 
placed by  members  of  K  belong  to  K; 

2.  If  Pri{zi)  is  a  member  of  K,  then  (22)21  •  Pri{zo),  (£22)2:1  . 
Pviiz-i),  and  C,„rXzi)  •  *  belong  to  it,  where  C,„„  denotes  the  property 
of  being  congruent  to  m  modulo  n,  m  <  n. 

3.  The  set   K  has  no  further  members. 
Then  every  member  of  K  is  realizable. 

For,  if  Pr\{zi)  is  realizable,  nervous  nets  for  which 

A^,(2i)  .  =  .  Pry{z,)  .  SN,{zi) 
iV,(zi)  .  ^  .  Pn(z,)vSN,fzr) 

are  the  expressions   of  equation    (4),   realize    (22)21  •  Priiz-z)   and 


A  Logical  Calculus  of  the  Ideas  Immanent  in  Nervous  Activity  395 

{E Zi)Zi  .  Priiz^j  respectively;  and  a  simple  circuit,  a,  C2,  ...  ,  c,., 
of  ?2  links,  each  sufficient  to  excite  the  next,  gives  an  expression 

A^„(zi)  .  =  .M(0)  .  C.„ 

for  the  last  form.  By  induction  we  derive  the  theorem. 

One  more  thing  is  to  be  remarked  in  conclusion.  It  is  easily 
shown:  first,  that  every  net,  if  furnished  with  a  tape,  scanners 
connected  to  afferents,  and  suitable  efferents  to  perform  the 
necessary  inotor-operations,  can  compute  only  such  numbers  as 
can  a  Turing  machine;  second,  that  each  of  the  latter  numbers 
can  be  computed  by  such  a  net;  and  that  nets  with  circles  can  be 
computed  by  such  a  net;  and  that  nets  with  circles  can  compute, 
without  scanners  and  a  tape,  some  of  the  numbers  the  machine 
can,  but  no  otliers,  and  not  all  of  them.  This  is  of  interest  as 
affording  a  psychological  justification  of  the  Turing  definition  of 
computability  and  its  equivalents,  Clhurch's  X  —  definability  and 
Kleene's  primitive  recursiveness:  If  any  number  can  be  computed 
by  an  organism,  it  is  computable  by  these  definitions,  and  con- 
versely. 

CONSEQUENCES 

Causality,  which  requires  description  of  states  and  a  law  of 
necessary  connection  relating  them,  has  appeared  in  several  forms 
in  several  sciences,  but  never,  except  in  statistics,  has  it  been  as 
irreciprocal  as  in  this  theory.  Specification  for  any  one  time  of 
afferent  stimulation  and  of  the  activity  of  all  constituent  neurons, 
each  an  "all-or-none''  affair,  determines  the  state.  Specification 
of  the  nervous  net  provides  the  law  of  necessary  connection  whereby 
one  can  compute  from  the  description  of  any  state  that  of  the 
succeeding  state,  but  the  inclusion  of  disjunctive  relations  prevents 
complete  determination  of  the  one  before.  Moreover,  the  regen- 
erative activity  of  constituent  circles  renders  reference  indefinite 
as  to  time  past.  Thus  our  knowledge  of  the  world,  including 
ourselves,  is  incomplete  as  to  space  and  indefinite  as  to  time. 
This  ignorance,  implicit  in  all  our  brains,  is  the  counterpart  of 
the  abstraction  which  renders  our  knowledge  useful.  The  role  of 
brains  in  determining  the  epistemic  relations  of  our  theories  to  our 


396 


Information  Storage  and  Neural  Control 


-<1 ^ 


<^ 


i7 


<■ 
< 


^> 


^f- 


FlOURE  1 


A  Logical  Calculus  of  the  Ideas  Immanent  in  Nervous  Activity  397 

observations  and  of  these  to  the  facts  is  all  too  clear,  for  it  is  ap- 
parent that  every  idea  and  every  sensation  is  realized  by  activity 
within  that  net,  and  by  no  such  activity  are  the  actual  afferents 
fully  determined. 

There  is  no  theory  we  may  hold  and  no  observation  we  can  make 
that  will  retain  so  much  as  its  old  defective  reference  to  the  facts 
if  the  net  be  altered.  Tinnitus,  paraesthesias,  hallucinations,  de- 
lusions, confusions  and  disorientations  intervene.  Thus  empiry 
confirms  that  if  our  nets  are  undefined,  our  facts  are  undefined, 
and  to  the  "real''  we  can  attribute  not  so  much  as  one  quality 
or  "form."  With  determination  of  the  net,  the  unknowable  object 
of  knowledge,  the  "thing  in  itself,''  ceases  to  be  unknowable. 

To  psychology,  however  defined,  specification  of  the  net  would 
contribute  all  that  could  be  achieved  in  that  field — even  if  the 
analysis  were  pushed  to  ultimate  psychic  units  or  "psychons,"  for 
a  psychon  can  be  no  less  than  the  activity  of  a  single  neuron. 
Since  that  activity  is  inherently  propositional.  all  psychic  events 
have  an  intentional,  or  "semiotic,"  character.  The  "all-or-none" 
law  of  these  activities,  and  the  conformity  of  their  relations  to 
those   of  the   logic   of  propositions,   insure   that   the   relations   of 

-^  EXPRESSION  FOR  THE  FIGURES 

In  the  figure  the  neuron  cv  is  always  marked  with  the  numeral  i  upon  the 
body  of  the  cell,  and  the  corresponding  action  is  denoted  by  W  with  i  as  sub- 
script, as  in  the  text. 

Figure  la  N-i(t)  .  =  .  A^i(/  -  1) 

Figure  lb  A^3(0  •  =  •  A^i^'  -  1)  V  N2(t  -  1) 

Figure  Ic  A^3(0  .  =  .  A^i(/  -  D  •  NM  -  1) 

Figure  Id  Nsit)  .  =  .  N,(t  -  1)  .  ^  N-2(t  -  1) 

Figure  le  .V,(0  :  =   :  .V,(/  -  1)  .  V  .  ,V2(/  -  3)  .  -  A^,(/  -  2) 

N,(t)  .  =  .  N-At  -  2)  .  N-zit  -  1) 

Figure  If  N^{t)  :  =  :  ~  A^i(;  -  1)  .  N ■i{t  -  \)vN^{t  -  1)  .  V  .  yVi(t  -  1)  • 

N-At  -  1)  .  iVsO  -  1) 
is!, it)  :  =  :  -  7Vi(<  -  2)  .  N At  -  2)  v  N -M  -  2)  .  v  .  NAt  -  2)  . 

NAt  -  2)  .  ^At  -  2) 
Figure  Ig     A^3(0  .  =  .  NAt  -  2)  .  ~  iV,(<  -  3) 
Figure  Ih     iVsCO  .  =  .  A^,(/  -  1)  .  NAt  -  2) 
Figure  li      .V3(0  :  =  :  Ar,(/  -  1)  .  V  .  ,V,(^  -  1)  .  {Ex)t  -  1  .  A^,(.v)  .  N Ax) 


398  Information  Storage  and  Neural  Control 

psychons  are  those  of  the  two-valued  logic  of  propositions.  Thus 
in  psychology,  introspective,  behavioristic  or  physiological,  the 
fundamental  relations  are  those  of  two-valued  logic. 

Hence  arise  constructional  solutions  of  holistic  problems  involving 
the  differentiated  continuum  of  sense  awareness  and  the  norma- 
tive, perfective  and  resolvent  properties  of  perception  and  execu- 
tion. From  the  irreciprocity  of  causality  it  follows  that  even  if  the 
net  be  known,  though  we  may  predict  future  from  present  activities, 
we  can  deduce  neither  afferent  from  central,  nor  central  from 
efferent,  nor  past  from  present  activities — conclusions  which  are 
reinforced  by  the  contradictory  testimony  of  eye-witnesses,  by  the 
difficulty  of  diagnosing  differentially  the  organically  diseased,  the 
hysteric  and  the  malingerer,  and  by  comparing  one's  own  mem- 
ories or  recollections  with  his  contemporaneous  records.  Moreover, 
systems  which  so  respond  to  the  difference  between  afferents  to 
a  regenerative  net  and  certain  activity  within  that  net,  as  to 
reduce  the  difference,  exhibit  purposive  behavior;  and  organisms 
are  known  to  possess  many  such  systems,  subserving  homeostasis, 
appetition  and  attention.  Thus  both  the  formal  and  the  final 
aspects  of  that  activity  which  we  are  wont  to  call  mental  are 
rigorously  deducible  from  present  neurophysiology.  The  psychi- 
atrist may  take  comfort  from  the  obvious  conclusion  concerning 
causality — that,  for  prognosis,  history  is  never  necessary.  He  can 
take  little  from  the  equally  valid  conclusion  that  his  observables 
are  explicable  only  in  terms  of  nervous  activities  which,  until 
recently,  have  been  beyond  his  ken.  The  crux  of  this  ignorance 
is  that  inference  from  any  sample  of  overt  behavior  to  nervous 
nets  is  not  unique,  whereas,  of  imaginable  nets,  only  one  in  fact 
exists,  and  may,  at  any  moment,  exhibit  some  unpredictable 
activity.  Certainly  for  the  psychiatrist  it  is  more  to  the  point  that 
in  such  systems  "Mind"  no  longer  "goes  more  ghostly  than  a 
ghost."  Instead,  diseased  mentality  can  be  understood  without  loss 
of  scope  or  rigor,  in  the  scientific  terms  of  neurophysiology.  For 
neurology,  the  theory  sharpens  the  distinction  between  nets  neces- 
sary or  merely  sufficient  for  given  activities,  and  so  clarifies  the 
relations  of  disturbed  structure  to  disturbed  function.  In  its  own 
domain  the  difference  between  equivalent  nets  and  nets  equivalent 
in  the  narrow  sense  indicates  the  appropriate  use  and  importance 


A  Logical  Calculus  of  the  Ideas  Immanent  in  Nervous  Activity  399 

of  temporal  studies  of  nervous  activity:  and  to  mathematical  bio- 
physics the  theory  contributes  a  tool  for  rigorous  symbolic  treat- 
ment of  known  nets  and  an  easy  method  of  constructing  hypo- 
thetical nets  of  required  properties. 

REFERENCES 

1.  Carnap,   R.:    The  Logical  Sjntax  of  Language.   New  York,   Harcourt. 

Brace  and  Company,  1938. 

2.  Hilbert,   D.,   und  Ackermann,   W.:   Grundiige  der    Theoretischen  Logik. 

Berlin,  J.   Springer,   1927. 

3.  Whitehead,  A.  N.,  and  Russell,  B.:  Principia  Mathematica.  Cambridge, 

Cambridge  University  Press,  1925. 


NAME  INDEX 


Abbott.  W.,  19,  170,  171,  353 

Abraham,  S.,  348 

Abt,J.  P.,  226 

Ackerman,  W.,  133,  138 

Ackermann,  \V.,  386,  399 

Adey,  VV.  R.,  240 

Aldrich,  A.,  295 

AUee,  W.  C,  170 

Alper,  T.,  116 

Apgar,  J.,  116 

Aposhian,  H.  V.,  135 

Arbib,  M.  H.,  295,  377 

Arduini,  A.,  226 

Arnon, D.  I., 148,  170 

Ashby,  W.  R..  142.  169 

Astrachan,  L.,  1 10,  111 ,  1 12,  1 19,  124, 

135 
Attneave,  F.,  173,  184 


Bach,  L.  M.  N.,  242 

Bachtold,J.  G.,  136 

Barlow,  H.  B..  327 

Barlow,  J.,  277 

Barnett,  L.,  115 

Basilic,  C,  71,  113,  119 

Bates,  J.  A.  V.,  241 

Bateson,  G..  25,  173,  184,  185.  186,  242, 

296,  330,  354,  355,  372 
Baumol,  ^V.  J..  171 
Bavelas,  A..  173,  181,  182 
Beadle,  G.  W..  59 
Beavers,  W.  R.,  186 
Beckwith,  W..  276 
Beers,  R.  F.,Jr.,  229 
Bell,  D.  A.,  16 
Bellman,  R.,  167,  171 
Belozersky,  A.  N.,  87,  106,  114 
Benzer,  S.,  71 
Berg,  P.,  71.  98,  115 
Bergold,  G.  H.,  114 
Bernard,  C.,  233 


Bidwell,  R.  G.  S.,  170 

Birdsall,  T.  G..  306,  327 

Bishop,  G.  H.,  226 

Blackman,  R.  B.,  348 

Block,  L.  N.,  in 

Blum,  M.,  285,  289,  292,  293,  295 

Blustein,  H..  21,  22,  184 

Boltzman,  L.,  5,  144 

Boyer.  G.  S..  137 

Branson,  H.  R.,  141,  169 

Brattgard,  S..  226 

Brazier,  M.  A.   B..   19,  226,  230,  24L 

242,  277.  355.  360 
Brenner,  S..  71,  114,  115,  135 
Brillouin.  L..  141.  142,  147,  169 
Britten,  R.J.,  135 
Broadbent,  D.  E.,  307,  308,  327 
Brown,  R.,  310,  327 
Bubel,  H.  C..  136 
Buchsbaum,  R.,  368 
Burch.  N.  R.,  24,  329,  348,  349,  355 
Burma,  D.  P.,  114 
Burns,  B.  D.,  226 
Burton,  K.,  101,  115 
Bush,  R.  R.,  49,  56 


Caceros,  C.  A.,  348 

Carnap,  R.,  377,  383,  399 

Chamberlin,  M.,  98,  115 

Chargaff,  E.,  115 

Cheng,  P.  v.,  115 

Cherry,  C,  17 

Childers,  H.  E.,  329,  348 

Chow,  K.  L.,  197,  198,  200,  225,  226 

Cohen,  G.  N.,  138 

Cohen,  S.  S.,  115,  125.  134.  135 

Copi,  I.  M..  377 

Cordes,  S.,  115 

Corley,  K.,  269 

Corning,  VV.  C,  276,  370 

Courtois.  G.,  216,  228 


401 


402 


Information  Storage  and  Neural  Control 


Cowan,  J.,  294,  295 
Craston,  D.  F.,  170 
Crawford,  E.  M.,  115 
Crawford,  L.  V.,  115 
Crick,  F.  H.  C,  60,  71,  74,  77, 
82, 83,  115,  119 


),  81, 


Daesch,  G.  E.,  137 

Darnell,  J.   E.,  Jr.,  74,   122,   123,   136, 

138,  139 
Davenport,  W.  F.,  241 
Davern,  C.  I.,  98,  115 
Davies,  D.  R.,  71 
Davison,  P.  F.,  115 
De  LaHaba,  G.  L.,  138 
Dean,  W.,  229 
Deininger,  R,  L.,  309,  327 
Deutsch,  J.  A.,  226 
Dewson,  J.,  197,  198,  200,  226 
Dickerson,  R.  E.,  71 
Dingman,  W.,  276,  368 
Dixon,  M.  K.,  137 
Dobzhansky,  T.,  374 
Doty,  P.,  87,  100,  115,  118 
Driesch,  H.  A.  E.,  358 
Dubbs,  D.  R.,  117 
Duda,  W.  L.,  55 
Dulbecco,  R.,  129,  136,  137 
Duncan,  C.  P.,  189,  196,  226 
Dunlop,  C.  W.,  240 
Dunn,  D.  B.,  115 


Echols,  H.,  59,  71,  72,  73,  74,  75,  121, 

353 
Edstrom,  J.,  115 
Edwards,  R.  J.,  348 
Eiduson,  S.,  276 
Elgot,  C.  C,  377 
Ellen,  P.,  276 
Emerson,  A.  E.,  170 
Epstein,  H.  T.,  115 
Essman,  W.  B.,  226 
Estes,  W.  K.,  49,  56 


Feinstein,  A.,  17 
Fields,  W.  S.,  353 
Finamore,  F.  J.,  116 
Finch,  J.  T.,  136 
Fitts,  P.  M.,  309,  327 
Flaks,J.  G.,  135 
Fogh,  J.,  133,  138 
Freeman,  G.,  137 
Freese,  E.,  83,  116 
Freifelder,  D.,  115 
Fresco,  J.  R.,  118 
Frey,  B.  A.,  136 
Frisch-Niggemeyer,  W.,  116 
Furth,J.  J.,  Ill,  116 

Gaarder,  T.,  170 

Gabor,  D.,  17,  294 

Gafford,  L.  G.,  118 

Garen,  A.,  71,  136 

Gebhardt,  L.  P.,  136 

Geiduschek,  E.  P.,  116 

Geller,  E.,  276 

Gerard,  R.  W.,  26,  189,  196,  226,  227, 
228,  305,  327,  353,  361,  367,  369, 
370,  371,  372,  374,  375,  376 

Gibbs,  W.,  144 

Gilbert,  W.,  116,  135 

Gillbricht,  M.,  170 

Gillies,  N.  E.,  116 

Ginsberg,  H.  S.,  137 

Glickman,  S.  E.,  276 

Goldman,  M.,  Ill,  116 

Goldman,  S.,  17,  348 

Goldring,  S.,  227 

Goldstein,  M.  H.,  241 

Gorman,  A.  L.  F.,  243 

Gran,  H.  H.,  170 

Gray,  E.  G.,  285 

Green,  M.,  137 

Gregory,  R.  T.,  27 

Grey  Walter,  VV.,  241 

Gros,  F.,  112,  116,  135,  136 

Grunberg-Manago,  M.,  67,  71 

Gumnit,  R.  J.,  227 


Fatt,  I.,  289 
Faulkner,  P.,  115 


Haibt,  L.  H.,  55 

Hall,  B.   D.,   Ill,   112,   116,   118,   119, 
124, 135 


Name  Index 


403 


Hall,  V.  E.,  361 

Halstead,  VV.  C,  211,  227 

Hamberger,  C.  A.,  227 

Hammer,  G.,  227 

Hart,  R.  G.,  71 

Hartley,  J.  W.,  138 

Hartley,  R.  V.  L.,  6 

Hartline,  H.  K.,  360 

Hayashi,  M.,  112,  116 

Hebb,  D.  O.,  37,  54,  55,  190,  227 

Hede,  R.,  115 

Helinski,  R.,  71 

Henderson,  K.,  134 

Hendrix,  C.  E.,  240 

Herriot,  R.,  121 

Hershey,  A.  13.,  134 

Hiatt,  H.,  116,  135 

Hilbert,  D.,  386,  399 

Hoagland,  M.  B.,  116 

Holland,  J.  H.,  55 

Holland,  J.  J.,  136 

Holley,  R.  W.,  103,  116 

Hollingworth,  B.  R.,  135 

Hooper,  L.,  138 

Home,  R.  W.,  132,  136 

Horowitz,  N.  H.,  59,  70 

Horvath,  W.J.,  314,  328 

Howes,  D.  VV.,  137 

Human,  M.  L.,  134 

Hurwitz,  J.,  Ill,  116 

Hutchinson,  G.  E.,  169 

Hyden,  H.,  211,  226,  227,  276 

Hyman,  L.  H.,  369 


lizuka,  R.,  211,  227 
Ingram,  V.  M.,  71 
Isaacs,  A.,  138 
Ivers,  R.  R.,  18 
Iwamura,  T.,  116 


Jacob,  F.,  69,  71,  114,  135,  136 

Jarvik,  M.  E.,  226 

Jasper,  H.  H.,  228,  277 

Jaynes,  E.  T.,  144,  170 

John,  E.  R.,20,  120,211,  227,243,247, 
249,  250,  260, 261, 262, 263, 267, 276, 
278,  282,  355,  364,  366,  369, 370, 372 


Joklik,  W.  K.,  136 
Jones,  O.  W.,  71 
Josse,  J.,  135 


Kaiser,  A.  D.,  136 

Katz,J.  J.,  211,  227,  289 

Kellaway,  P.,  23,  24 

Kcndrew,  J.  C.,  71 

Khinchin,  A.  I.,  17 

Killam,  K.  F.,  245,  246,  247,  248,  249, 

250,  259,  260,  261,  262,  263,  276 
Kiinura,  K.,  119 
Kirby,  K.  S.,  117 
Kit,  S.,  76,  117,  120,  122,  139,  353,362, 

372,  375 
Kleene,  S.  C.,  377,  395 
Kleinschmidt,  W.  J.,  117 
Klug,  A.,  136 
Kok,  I.  P.,  117 
Kornberg,  A.,  125,  135 
Kornberg,  S.  R.,  135 
Korolkova,  T.  A.,  276 
Kozloff,  L.  M.,  134 
Kraft,  M.  S.,  228 
Kreps,  E.,  211,  228,  276 
Krey,J.,  170 
Kristiansen,  K.,  216,  228 
Kroger,  H.,  114 
Kurland,  C.  G.,  116,  135 
Kurtz,  H.,  138 


Lawley,  P.  D.,  83,  117 
Lciman,  A.  L.,  243,  267,  276 
Lengyel,  P.,  71,  113,  117,  119 
Lenneberg,  E.,  310,  327 
Leuchtenbergcr,  C.,  137 
Levine,  M.,  136 
Levinthal,  C.,  70,  115 
Levintow,  L.,  136,  138 
Levy,  H.  B.,  138 
Liberson,  W.  T.,  264,  268,  276 
Libet,  B.,  227 
Lichtenstein,  J.,  135 
Lindegren,  C.  C.,  118 
Lindegren,  G.,  118 
Lindeman,  R.  L.,  169 
Lindsay,  R.  K.,  34,  353 


404 


Information  Storage  and  Neural  Control 


Linschitz,  H.,  141,  169 

Lipmann,  F.,  1 38 

Littlefield,  J.  \V.,  120 

Livanov,    M.   N..   245,   246,   264,   268, 

276,  277 
Lockart,  R.  Z.,  136 
Loeb,  T.,  117 

Lorente  de  No,  R.,  332,  348 
Loucks,  R.  B.,  272,  277 
Luria,  S.  E.,  134,  136 
Lute,  M.,  134 
Lwoff,  A.,  137 
Lwoff,  M.,  137 


MacArthur,  R.,  143,  169 

MacFadyen,  A.,  169 

Magasanik,  B.,  117 

Majkowski,  J..  252,  253,  277 

Makhinko,  V.  I..  117 

Maling,  B.,  71 

Mandelbrot,  B.,  43,  44,  45,  46,  56 

Marcus,  P.  I.,  137 

Margalef,  D.  R.,  169 

Marmur,  J.,  87,  115,  118 

Martin,  E.  M.,  115 

Martin,  R.  G.,  71 

Matthaei,  J.  H.,  64,  71,  112,  117,  134, 

139 
Mayor,  H.D.,  17,  18,74,  121,  122,  171, 

172 
McConnell,  J.,  369,  370,  371 
McConnell,  W.,  170 
McCulloch,  W.  S.,  36,  38,  55,  283,  296, 

297,  298,  306,  375,  376,  377,  379 
McGlothlen,  M.,  72,  121 
McLaren,  L.  C,  136 
McQuillen,  K.,  135 
Meier,  R.  L.,  324,  325,  328 
Merrill,  S.  H.,  116 

Meselson,  M.,71,96,  114,  117,  118,  135 
Miller,  G.  A.,  52,  53,  56,  355,  357 
Miller,  J.  G.,  301 
Minagawa,  T. ,  117 
Minckler,  S.,  118 
Minsky,  M.,  292 
Monod,  J.,  69,  71 
Moore,  H.  F.,  118 
Morganstern,  O.,  170 


Morrell,  F.,  73,  189,  228,  244,  245,  246, 
248, 277, 278, 282, 355, 364, 365, 366, 
367,  371,  374 

Moses,  L.,  225 

Mostellar,  F.,  49,  56 

Mountcastle,  V.  B.,  237,  241 

Mowbray,  G.  H.,  328 

Muller,J..  231,  233,  234 

Munier,  R.,  138 

Myers,  J.,  116 


Nagington,  J.,  132,  136 

Naitoh,  P.,  225 

Nakamoto,  T.,  1 11,  116,  120 

Nathan,  P.  W.,  224,  229 

Nathans,  D.,  138 

Neff,  W.  D.,  268,  277 

Newton,  A.,  137 

Nieder,  P.  C.,  277 

Nirenberg,  M.  W.,  64,  67.  71,  74,  75. 

112,  117,  134,  139 
Nomura,  M.,  110,  111,  118 
Nyquist,  H.,  6 


Obrist,  VV.  D.,  228 

Ochoa,  S.,  67,71,  74,  75,  112,  113,  114, 

117,  119 
Odum,  H.  T.,  170 
Oesterreich,  R.  E.,  277 
Ogur,  M.,  118 
OXeary,  J.  L.,  226,  227 
Olken,  H.,  374 
Onesto,  N.,  295 
Opalskii,  A.  F.,  117 
Osawa,  S.,  118 


Park,  O.,  170 

Park,  T.,  170 

Patrick,  B.  S.,  25 

Patten,  B.  C.,  140,  169,  171,  172,  356 

Pautlcr,  E.,  297 

PhiUips,  D.  C.,71 

Pigon,  A.,  227 

Pitts,  W.  H.,  36,  55,  284,  377,  379 

Piatt,  J.  R.,  305,  327 

Polyakov,  K.  L.,  245,  277 


Name  Index 


405 


Prange,  E.,  292 
Pribram,  K.  H.,  228,  229 
Puck,  T.  T.,  137 


Quastler,  H.,  308,  309,  327 


Rabinovvitch,  E.  I.,  170 
Rabson,  A.,  138 
Randall,  C.  G.,  118 
Ransmeier,  R.  E.,  228 
Rapoport,  A.,  46.  56,  314,  328 
Rashevsky,  N.,  167,  171,  379 
Reymond,  D.  B.,  231 
Rhoades,  M.  V.,  328 
Rich,  A.,  110,  118 
Rikli,  A.  E.,  348 
Ris,  H.,  71 

Risebrough,  R.  VV.,  116,  135 
Roberts,  L.,  228 
Roberts,  R.  B.,  135 
Rochester,  N..  37,  38,  55 
Rogers,  S.,  138 
Roitbak,  A.  E,  197,  229 
Roizman,  B.,  138 
Rolfe,  R.,  96,  118 
Root,  W.  L.,  241 
Rose,  VV.  R.,  138 
Rosenblatt,  P.,  54,  56 
Ross,  G.,  228 
Ross,  R.  W.,  137 
Rothman,  F.,  71 
Rothschild,  367 
Rothstcin,  J.,  170 
Rowland,  V.,  192,  229 
Rubin,  H.,  129,  137,  138 
Rusinov,  V.  S..  192.  229 
Russell,  B.,  383,  399 
Russell,  W.  R..  224,  229 
Ryther.J.  H.,  170 


Sachs,  E.,  267,  276,  368 

Sachs,  L.,  137 

Saltzbcrg,  B.,  5,  17,  18,  19,  20,  21,  22, 

23,  24,  25,  26,  296,  330,  348 
Salzman,  N.  P.,  118,  128,  136,  137,  138 
Sandler.  B.,  228 


Schafer,  \V.,  1 18,  131,  138 

Schaffer,  F.  E.,  118,  136 

Schildkraut,   C.   E.,   87,   96,    101,    102, 

110,  118 
Schlessinger,  D.,  135 
Schmidt,  K.  P.,  170 
Schrodinger,  E..  140,  141,  142,  147,  169 
Schuster,  H.,  74,  118 
Schwartz,  M.,  17 
Schwerdt,  C.  E.,  118,  136 
Sebring,  E.  D.,  136 
Sclfridge,  O.,  292 
Sevring,  E.  D.,  138 
Shannon,  C.  E.,  5,  13,  14,  15,  17,  52, 

142,  144,  168,  169, 230, 240, 241, 318 
Shapiro,  A.,  18,  19,  23,  73 
Sherrington,  C.  S..  51,  231,  241,  360 
Shipton,  H.  W.,  21.  241.  349 
Shore,  V.  C.,  71 
Sibatani,  A..  119 
Siebert,  W.  M.,  241 
Siminovitch,  L..  135,  136 
Simon,  E.  H.,  119 
Simon,  H.  A.,  41.  42,  46,  56 
Sines,  J.  O.,  277 
Sinsheimer,  R.  E..  119,  135 
Smith,  J.  D.,  115 
Smith,  K.  M.,  137 
Snedecor,  G.  \V.,  170 
Sokolov,  E.  N.,  233,  241 
Spahr,  P.  F.,  115,  135 
Speyer,J.  F.,  71.  113,  117,  119 
Spiegelman,  S.,  1 1 1.  1 12,  1 16,  1 18,  1 19, 

124, 135 
Spirin,  A.  S.,  87,  106,  114,  119 
Spoor,  VV.  A.,  348 
Sporn,  M.  B.,  276,  368 
Stahl,  F.  VV.,  117 
Stamm,  J.  S.,  229 
Steinberg,  C.  A.,  346.  348 
Steiner,  R.  F.,  229 
Stephenson,  M.  E.,  103,  119 
Stern,  J.  A.,  248,  277 
Stevens,  A.,  119 
Stevens,  S.  S.,  306,  307,  327 
Stoker,  M.  G.  P.,  137 
Storck,  R.,  112,  119 
Strandberg,  B.  E.,  71 
Strauss,  B.,  117 


406 


Information  Storage  and  Neural  Control 


Streisinger,  G.,  135 
Stuart,  D.  C.,Jr.,  133,  138 
Stumpers,  F.  L.  H.  M,,  17 
Sueoka,  N.,  99,  115,  119 
Suwa,  N.,  227 
Sved,  S.,  115 
Sverdrup,  H.  U.,  171 
Swets,J.  A.,  306,  327 
Szent-Gyorgi,  A.,  147,  148,  170 
Szilard,  L.,  5 


Takahashi,  T.,  119 
Takahata,  N.,  227 
Takeda,  T.,  227 
Taketomo,  Y.,  185 
Tamm,  I.,  137 
Tanabe,  M.,  227 
Tanner,  W.  P.,  Jr.,  306,  327 
Tatum,  E.  L.,  59 
Temin,  H.  M.,  129,  137,  138 
Tessman,  I.,  97,  119 
Thomas,  R.  S.,  119 
Thompson,  R.,  229 
Thoren,  M.,  138 
Thrall,  R.  M.,  169 
Tissieres,  A.,  135 
Tobias,  J.  M.,  120,  229,  371 
Travers,  P.  L.,  180,  184 
Tribus,  M.,  144,  170 
Tschirgi,  R.  D.,  211,  368 
Tukey.J.  W.,  348 
Turing,  A.  M.,  36,  55,  377 


Volkin,  E.,  110,  111,  112.  116,  119,  124, 

135 
von  Foerster,  H.,  27"^, 
von  Neumann,  J.,  5,  36,  55,  170,  284, 

285,  375 


Wagner,  B.,  117 

Warner,  R.  C,  114 

Watson,  J.  D.,  60,  71,  77,  80,  81,  116, 

119,  120,  135,  139 
Watts-Tobin,  R.  J.,  115 
Weaver,  W.,  17 
Weill,  J.  D.,  114 
Weiner,  M.  F.,  19,  186 
Weir,  H.  F.,  297,  298 
Weiss,  M.,  243,  252,  254,  255,  256,  277 
Weiss,  S.  B.,  Ill,  116,  120 
Welch,  A.  J.,  348 
Wenzel,  B.  M.,  211,  368 
Wheelock,  E.  F.,  137 
Whitehead,  A.  N.,  360,  383,  399 
Whitfield,  I.  C.,  234,  241 
Wiener,  N.,  6,  169,  171,  230,  241 
Winocour,  E.,  137 
Winograd,  S.,  294,  295 
Wittman,  H.  G.,  71 
Woese,  C.  R.,  120 
Woodward,  P.  M.,  9,  17 
Work,  T.  S.,  115 
Wright,  J.  B.,  377 
Wyatt,  G.  R.,  135 


Ulett,  G.  A.,  277 


Valentine,  R.  C.,  115 
Valentinuzzi,  M.  E.,  24,  374 
Vallentyne,  J.  R.,  170 
van  Leeuwuen,  W.  S.,  349 
Vendrely,  R.,  119 
Verbeek,  L.,  294,  295 
Verveen,  B.,  289,  295 
Vinograd,  J.,  117 
Vladimirov,  G.  E.,  229 
Vogt,  M.,  129,  137 


Yamana,  K.,  119 
Yanofsky,  C.,  71 
Yarmolinsky,  M.  B.,  138 
Yngve,  v.,  46,  47,  48,  49,  52,  56 


Zamecnik,  P.  C.,  103,  119 
Zimmerman,  J.  B.,  135 
Zimmerman,  T.,  138 
Zinder,  N.  D.,  117,  136 
Zipf,  G.  K.,  39,  40,  43,  44,  56 
Zubkoff,  P.  L.,  116 
Zuckermann,  E.,  266,  277 


SUBJECT  INDEX 


A 

Afferents 

interaction  of,  285-289,  296 
peripheral,  383,  385,  390-393 

After  discharge,  199,  218,  382 

All-or-none  law,  381,  397 

Amino  acids,  62-67,  72,  103 

Amnesia,  retrograde,  189,  224,  366 

Assimilated  rhythms,  247,  248 

Attention,  48,  233,  361 

Auditory  mechanisms,  266,  285,  307 

Automata  theory,  284,  377 

Axons,  190,  284,  289,  293-347,  371,  379, 


B 

Bacteria,  59,  102,  106,  139 

DNA  from,  102 

metabolism  of,  97 

mutant  strains,  63 
Bacteriophages,  94,  110,  123,  124,  127 

T-even,  79,  85 

T-4  mutants,  82 

nitrous  acid  induced  mutants,  97 
Base  pairing,  62,  65 

specific,  79 
Behavior 

differential  conditioned,  274 

disturbed,  179 

ordered,  306,  355 

purposive,  398 
Binary  representation  in  computers,  27- 

30 
Binary  units,  6,  9,  10 
Biomass,  144,  149,  151,  155,  164 
Brain 

information  storage  in,  56 

information  transfer  ir,  240 

number  of  neurons  in,  361 


c 

Calculation 

error-free,  292,  293 
Calculus,  logical,  55,  284,  377 
Cannibalism  experiments,  369 
Channel 

Capacity,  12,  15,  142,  240,  295,  308, 
311,  325 

communication,  7,  52 

correction,  15,  142 

length  of,  308 

noise,  240 

overloaded,  31 1 
Chlorophyll,  150,  155,  157 
Coding 

genetic,  61,  76,  82,  112,  353 

in  nerve  cells,  246 

in  nervous  system,  233 

in  time  domain,  330 

of  language,  43-48 

spatio-temporal  patterns  of,  279-282 
Coincidence  analysis,  243,  278,  340 
Communication 

accuracy  in,  26 

channels,  12,  54,  142,  240 

economics  of,  25,  372 

pathological  alteration,  180 

systems,  12,  15,  25,  242,  373 

theory,  12,  17,  142,  230,  308 
Communities 

adaptability  of,  166 

bioenergetics  of,  140,  147 

complexity  of,  1 43 

diversity  of,  162 

energy  balance  in,  163 

stability  of,  143 

trophodynamics,  140,  147 
Computation,  error-free,  290 
Computers 

averaging  by,  233 

coding  in,  31,  36,  37 


407 


408 


Information  Storage  and  Neural  Control 


general  purpose,  345 

generation  of,  375 

simulation  of  brain,  38,  51 
Conditioning 

avoidance  response,  249 

differential  to  central  stimulation,  268 
Correction  of  errors,  14,  15 


D 

Decision  making,  34,  49,  167,  243,  304, 

310 
Decoding,  10,  295,  303,  309 
Deoxyribonucleic  acids  (DNA) 

amount  per  cell,  84,  85 

as  genetic  material,  60 

average  composition,  85-93 

base  composition,  93 

base  sequence  of,  62,  66,  70,  72,  80,  88 

distribution  of,  98 

equilibrium  sedimentation  of,  95 

formation  of  hybrid  molecules,  101 

genetic  information  in,  60,  66,  222 

heterogeneity  of  composition,  96,  99, 
112^ 

molecular  size  of,  85,  96,  97,  101,  122 

non-overlapping  bands,  97 

phage  0X1,  74,  98,  111 

primer,  65,  109,  111 

replication,  80,  363 

structure,  60,  77,  80,    81,    101,    109, 
121 

synthesis,  76,  80,  109,  121,  125,  375 
Dependency,  177 
Deterministic  models,  230,  236 
Discrimination,  178,  243,  259,  264 
Dominance,  177 


Error 

correction  of,  7,  13 

frequency  of,  14 

in  performance,  259-264 

of  commission,  261 

of  omission,  260 

probability  of,  290.  293,  295 
Exchange,  interpersonal,  175 
Expectation,  37,  185,  354 
Experience,  fixation  of,  354,  359,  363, 

374 
Extinction,  153,  381,  388 


Filter,  22,  278,  297,  340,  346,  348 
Filtering,  311,  314,  319,  324 
Fixation,  189,  355,  365-367 
Frequency  analysis,  23,  26,  329,  340 


Galvanic  skin  response  (GSR),  338,  348 
Generalization,  248,  258,  272,  278 
Genes,  59-63,  69,  72,  121,  126,289,353, 
359,  374 
as    determinants    of    protein    struc- 
ture, 61 
mutation  of,  72,  126 
suppressor,  72 
Genetic  coding,  61,  64,  70,  114 


H 

Habituation,  192,  233,  264 
Homeostasis,  142,  232,  398 


E 
Electrocardiogram  (EKG),  330,  343- 

348 
Electroencephalogram  (EEG),    21,    23, 

190, 207, 252, 278, 330-340,  348 
Energy  balance,  163,  165 
Energy  gains  and  losses,  140,  149,  155, 

166,  171,  356 
Entropy,  5,  15,  25,  141,  147,  149,  240 
Environmental  influences,  355,  356 
Equivocation,  13,  14,  15,  19,  336 


Information 
capacity,  9 
content,  9 

coding,  in  brain,  268 
flow,  route  of,  305 
genetic,  59,  70,  76,  103,  108,  122,  123. 

189 
input  overload,  311 
measure,  5,  6,  7,  240 
overload,  311,  314,  325 
overload  testing,  316 
transmission  of,  25,  124,  304,  310,  362 


Subject  Index 


409 


Information  processing 

essential  subsystems,  302 

in  computers,  31,  32,  33 

in  human  brain,  240,  301 

in  time  domain,  329-348 

models  of,  35,  38,  41 ,  54,  230,  236,  239 

subsystems  research,  306 
Information  storage,  8-10 

fixation  of  experience,  363,  366 

in  nerve  cells,  189 

long-term,  244,  369 

mechanisins,  192 

short-term,  195-197,  210,  211 
Information  theory,  5,  240,  295 

in  ecology,  140-149 

in  neurophysiology,  230-239 
Inhibition,  235,  268,  284,  341,  360,  381, 

388 
Inputs,  53,  235,  287,  293,  302-308,  311, 

314,  325,  365 
Interaction 

group,  341,  343 

of  afferents,  285,  296 

virus  and  cell,  129 
Interference,  18,  22,  244 


Language,  38,  303,  305,  362,  383 

coding,  43-48 

information  in,  20 

redundancy  of,  26 

statistical  properties  of,  42 
Learning,  173,  181,  283,  303,  310,  357, 
366,  381,  389 

levels  of,  174,  177,  183,  185,  190,  355, 
359 

process  of,  177,  181 ,  232,  354,  359,  371 

theory  of,  56,  173,  301,  310 
Logic 

of  propositions,  379,  381,  397 

probabilistic,  55,  284,  293 


M 

Machine,  computing,  see  Computers 
Machine,  Turing,  36,  284,  377,  395 
Malleability,  of  processing  system,  355, 
360,  362 


Memory,   37,   52,   225,  243,   255,   280, 
304,  310,  359,  363,  374 

cellular,  201,  212 

enduring,  189,  365 

functional,  201 

recent,  189 

retention  of,  189 
Message,   12,   15,    19,  25,  301-305 

{see  also  Binary  units) 
Messenger,  see  RNA 
Metacommunications,  186,  354 
Metalanguage,  186 

Models,  information  processing,  see  In- 
formation processing 
Mutations,  61,  74,  80 

chemically  induced,  81 

externally  adaptive,  372 

genetic,  63,  363 

suppressor,  72 


N 

Natural  selection,  167,  373 
Nets,  see  Neuron  nets 
Neuron  nets 

anastomotic,  283-295 

logically  stable,  287 

with  circles,  390 

without  circles,  382 
Neurons 

input.  290,  294 

internuncial,  380 

logical  functions  of,  286 

output,  284,  292,  297 

spontaneously  active,  236,  390 

storage,  235 

threshold  of,  289,  314,  381 
Noise,  13,  18,  21,  22,  171,  283,  289,  296, 
305,  330,  375 

errors  induced  by,  26,  142 

fluctuation  in,  309 

high  frequency,  340 

random,  12,  18,  22 
Nucleotides 

composition,  94,  95 

sequence,  83,  101 

triplets,  82,  97,  114 
Numbers,  binary,  27,  29,  30 


410 


Information  Storage  and  Neural  Control 


O 

Omission,  260,  311,  317,  319,  324,  325 
Order 

functional,  355 

structural,  355 
Outputs,   38,   142,   233,   288,   292-297, 

302-308,  311,  318,  326,  356 
Overload,  see  Information  overload 


Perception,  283,  284,  304 
Performance 

erroneous,  269,  275 

principle  factors  limiting,  309 

under  overload  conditions,  320,  323 
Period  analysis,  189,  329,  340-348,  359 
Photosynthesis,  154-156,  164,  165 
Planaria,  368 

cannabalism  studies,  370 
■  Poliovirus 

biosynthesis  of,  132 

multiplication  of,  133 

properties  of,  127 
Prediction,  174,  181 
ProbabiHstic  models,  230,  233 
Probability,   10,  11,  19,  20,  25,  41-43, 

143,  145,  326 
Problem  solving,  35 
Protein 

specificity,  59,  68 

structure,  62,  70,  72 

synthesis,  59,  64-68,  70,  108,  124,  371 

viral,  129-132 
Purines,  60,  61,  76,  79 
Pyrimidine,  60,  61,  76,  79,  103 

Q 

Queuing,  311,  317,  321,  324 

R 

Random  processes,  232 

Receiver,  6,  25,  303 

Redundancy,  7,  14,  25,  26,  84,  122,  173, 

181,  283 
Reinforcement,  175-177,  259,  269,272 
Replication 

phage,  130 

virus,  123,  132 


Response,  175-177,  197 

behaviorally  appropriate,  257 
conditioned,  190,  211,  245,  248,  253, 

263,  317,  371 
differentiated,  271 
generalization  of,  248,  258 
graded,  235 
labeled  potentials,  245 

Reverberation,  37,  280,  282 

Ribonucleic  acid  (RNA) 

and  information  storage,  120 
and  memory,  244,  280 
base  sequence,  66,  73,  222 
cellular  concentration  after  stimula- 
tion, 211,  221 
general  characteristics,  103 
informational,  see  messenger 
messenger,  64-67,  74,  76,  80,  98,  108- 

114,  124,  363 
ribosomal,  64,  105,  124,  363 
synthesis.  111,  121,  125,  368 
total  cellular,  105 
transfer,  64-66,  72,  103,  112,  131 
virus,  74,  103,  107,  114,  128,  132,  139 

Ribosomes,  64,  139 


Shannon's  Theorem  10,  142,  144,  148 
Signals,  11,  13,  23,  233,  283,  289,  290, 
305,  310 

electroencephalographic,  330 

electrocardiographic,  343,  347 

meaningful,  237 

random,  18 

reconstituted,  335 

synchronous,  284 
Signals-in-noise,  230,  232 
Specificity 

genetic,  62 

protein,  59,  62,  68 
Stimuli,  176,  238 

concurrent    peripheral    and    central, 
264 

conditioned,  211,  248,  252,  258,  264- 
266,  271,  275 

flicker,  264 

peripheral,  252,  266,  269 

photic,  265,  269,  271,  275 

tracer,  245,  271 


Subject  Index 


411 


Storage  {see  also  Information  storage) 

capacity,  8-11,  375 

mechanisms,  210,  370 

long-term,  244,  308,  369 

short-term,  195,  197 
Symbols,  14,  52,  53,  362 
Synapses,  37,  234,  313,  365,  379,  382, 
388,  390 

alterable,  390 

delay  across,  380,  382,  387 

excitatory,  381,  385,  388 

inhibitory,  382,  385,  388 

irreciprocal,  360 

reciprocal,  360 
System 

auditory,  192,  306 

biological,  17,  18,  57,  305 

communication,  8,  184 

Darwinian,  372,  373 

dissolution  of,  325 

equivocation  of,  16 

genotypic,  372,  373 

homogeneous,  357 

Lamarckian,  372,  374 

non-linear,  234 

protein  synthesizing,  65 

output  of,  38 

permanently  altered,  357 

random,  11,  25 

receiving,  20 

static,  8 

visual,  306 


Theorems  I  to  X 

McCulloch  and  Pitts,  382-394 
Theory 

behavior,  38,  53 

mentalistic,  39 

stochastic,  41,  43 
Threshold  [| 

changes  in,  382  "8 

differential,  265 

neuron,  272,  289,  369 

occlusion,  265,  270,  273 

variation  in,  388 
Time  domain,  329-348 
Trophodynamics,  140,  141 
Turing  machine,  see  Machine,  Turing 


u 

Uncertainty,  5,  17,  143,  145,  314 

V 

Virus  action 

formation  of  precursor  molecules,  1 30 

fowl  plaque  virus,  131 

on  cell  synthesis,  126 

poliovirus  biosynthesis,  132 
Virus  replication,  123,  133 


w 

Watson-Crick  model  for  DNA,  60,  77