Skip to main content

Full text of "...Mathematical biophysics of the central nervous system"

See other formats


Marine  Biological  Laboratory 

7_,  I  33  Pt  A  R   Y 

AUG  1  4  1947 

WOODS  HOLE.  MASS. 

Mathematical  Biophysics  Monograph  Series,  No.  1 


MATHEMATICAL  BIOPHYSICS  OF 
THE  CENTRAL  NERVOUS  SYSTEM 


By 
ALSTON  S.  HOUSEHOLDER 

AND 

HERBERT  D.  LANDAHL 


SttJV-ULTPS 


The  Principia  Press,  Inc. 
bloomington,  indiana 


COPYRIGHT  1945  BY 
THE  PRINCIPIA  PRESS 


COMPOSED  AND  PRINTED   BY    THE    DENTAN    PRINTING    COMPANY 
COLORADO  SPRINGS,  COLORADO 


PREFACE 

Since  the  proposal  of  the  two-factor  dynamical  model  of  neural 
activity  by  Professor  Rashevsky  in  his  book,  Mathematical  Biophysics, 
a  great  deal  of  work  has  been  done  in  the  formal  development  of  the 
theory  as  well  as  in  the  applications  to  specific  psychological  problems. 
It  is  only  natural  that  these  developments  by  various  authors  over  a 
period  of  time  will  be  lacking  somewhat  in  coherence  and  continuity. 

With  this  in  view,  it  seems  appropriate  to  pause  at  the  present 
time  to  review  the  work  which  has  been  done  so  far  to  explain  sys- 
tematically the  techniques,  to  summarize  and  describe  such  structures 
as  have  been  devised  and  the  applications  made,  to  suggest  promising 
directions  for  future  development,  to  indicate  types  of  experimental 
data  needed  for  adequate  checks,  and  also  to  present  new  material  not 
published  elsewhere.  It  is  the  hope  that  this  perspective  of  past 
achievements  and,  still  more,  of  future  prospects,  will  be  of  benefit  to 
those  theorists  and  experimenters  alike  who  are  interested  in  con- 
tributing to  the  understanding  of  some  of  the  mechanisms  which  un- 
derlie psychological  processes. 

While  perhaps  the  names  which  most  commonly  run  throughout 
the  monograph  are  those  of  the  authors  themselves,  this  is  largely 
only  a  reflection  of  their  dominating  interests  in  its  preparation;  and 
neither  the  monograph  itself  nor  the  papers  of  the  authors  and  of 
many  others  herein  referred  to  would  ever  have  seen  the  light  of  day 
had  not  the  general  procedures  and  fundamental  postulates  been  pre- 
viously developed  by  Professor  Rashevsky.  For  this,  and  for  other 
reasons  too  abundant  to  enumerate,  the  authors  owe  to  him  their 
foremost  debt  of  gratitude. 

Their  thanks  are  due  also  to  Dr.  Warren  S.  MeCulloch,  Dr.  Ger- 
hardt  von  Bonin,  and  Dr.  Ralph  E.  Williamson  for  many  helpful  sug- 
gestions made  during  the  preparation  of  the  manuscript,  and  to  Mr. 
Clarence  Pontius  for  preparing  all  the  original  drawings ;  to  Mrs.  Gor- 
don Ferguson  and  Miss  Helen  De  Young  for  typing  of  the  manuscript, 
to  Miss  Gloria  Robinson  for  final  preparation  of  the  manuscript, 
proofreading,  and  preparation  of  the  index.  For  help  with  the  latter 
thanks  are  also  due  to  Mr.  Richard  Runge. 

The  authors  also  wish  to  thank  the  Editor  of  The  Bulletin  of 
Mathematical  Biophysics  for  permission  to  reproduce  Figures  1  and 
2  of  chapter  vi ;  Figure  2  of  chapter  vii ;  Figures  3,  4,  9,  and  10  of 
chapter  ix ;  Figures  2,  3,  and  4  of  chapter  xi,  Figure  1  of  chapter  xiii 
and  Figure  1  of  chapter  xiv.  To  the  Editors  of  Psychometrika,  their 
thanks  are  due  for  permission  to  reproduce  Figures  1,  6,  7,  and  8  of 
chapter  ix,  and  to  The  University  of  Chicago  Press  for  permission  to 

iii 


reproduce  Figures  2  and  5  of  chapter  ix,  from  Rashevsky's  Advances 

and  Applications  of  Mathematical  Biology,  and  Figure  1  of  chapter 

xi  from  his  Mathematical  Biophysics. 

Finally,  the  authors  wish  to  express  their  gratitude  to  the  Prin- 

cipia  Press  and  to  the  Dentan  Printing  Company  for  their  unfailing 

efforts  involved  in  publishing  the  book. 

Alston  S.  Householder 
Herbert  D.  Landahl 

Chicago,  Illinois 

October,  19U 


IV 


TABLE  OF  CONTENTS 

PAGE 

Introduction vii 

PART  ONE 

CHAPTER 

I         Trans-synaptic  Dynamics 1 

II.  Chains  of  Neurons  in  Steady-State  Activity    -    -  7 

III.  Parallel,  Interconnected  Neurons 13 

IV.  The  Dynamics  of  Simple  Circuits 22 

V.  The  General  Neural  Net 30 

PART  TWO 

VI.  Single  Synapse:  Two  Neurons 37 

VII.  Single  Synapse  :  Several  Neurons 49 

VIII.  Fluctuations  of  the  Threshold 53 

IX.  Psychological  Discrimination 56 

X.  Multidimensional  Psychophysical  Analysis    -    -  74 

XI.  Conditioning    ------- 80 

XII.  A  Theory  of  Color-Vision 90 

XIII.  Some  Aspects  of  Stereopsis 94 

PART  THREE 

XIV.  The  Boolean  Algebra  of  Neural  Nets    -    -    -    -  103 

XV.  A  Statistical  Interpretation Ill 

Conclusion 114 

Literature 116 

Index 119 

61131 


INTRODUCTION 

This  monograph  is  directed  toward  the  explanation  of  behavior  by- 
means  of  testable  hypotheses  concerning  the  neural  structures  which 
mediate  this  behavior.  We  use  the  word  behavior,  for  lack  of  a  better 
term,  in  a  very  broad  sense  to  cover  any  form  of  response  to  the  en- 
vironment, internal  or  external,  whether  it  is  acting  or  only  perceiv- 
ing, and  whether  the  response  occurs  immediately  or  after  long  delay, 
providing  only  the  response  is  governed  by  nervous  activity  initiated 
by  occurrences  in  the  environment.  We  are  seeking  to  develop  a 
theory  of  the  nervous  system  as  the  determiner  of  behavior. 

Data  of  anatomy  and  physiology  are  altogether  inadequate — un- 
less in  the  case  of  a  simple  spinal  reflex — for  tracing  in  detail  the 
progress  of  a  nervous  impulse  from  its  inception  at  a  receptor, 
through  its  ramified  course  in  the  nervous  system,  to  its  termination 
at  the  effector.  We  know  pretty  well  where  the  fibers  lead  from  the 
retina;  we  known  even  in  some  detail  where  the  different  retinal 
areas  are  mapped  on  the  cortex ;  we  know  a  good  deal  about  the  inter- 
action of  one  region  of  the  cortex  upon  another,  and  we  know  many 
details  about  the  functioning  of  neural  units.  But  how  the  neural 
units  are  combined  in  the  visual  area  to  enable  the  organism  to  locate 
an  object  seen  and  to  act  accordingly,  how  the  nervous  discharges 
from  the  two  retinas  are  shunted  this  way  and  that  to  combine  and 
emerge  at  the  appropriate  effectors,  is  not  explained  by  existing  ex- 
perimental and  observational  technique.  For  solving  such  problems 
it  is  necessary  to  create  testable  hypotheses,  to  be  revised,  replaced, 
or  expanded  according  to  the  outcome  of  the  tests. 

The  task  of  developing  this  theory  is  three-fold.  First,  an  ideal- 
ized model  of  the  elementary  units  must  be  constructed  in  terms  of 
postulates  governing  their  individual  behavior  and  their  interactions. 
The  model  must  be  simple  enough  to  permit  conceptual  manipula- 
tion. The  units  we  have  designated  neurons.  Our  neurons  are  defined 
by  the  hypotheses  we  impose  upon  them,  and  it  may  turn  out  that  not 
the  single  neuron  of  the  physiologist  and  anatomist,  but  some  re- 
curring complex  of  these  is  most  properly  to  be  regarded  as  its  pro- 
totype. The  junction  of  neurons  we  refer  to  as  a  synapse,  and  where 
the  impulses  from  two  or  more  neurons  are  able  to  summate  in  pro- 
ducing a  response  in  one  efferent  neuron,  or  in  each  of  several,  we 
have  also,  briefly,  referred  to  the  set  of  these  junctions  as  constitut- 
ing a  single  synapse.  Such  usage,  in  harmony  with  that  of  Rashevsky 

vii 


MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

(1938,  1942) ,  seems  to  simplify  the  terminology,  since  for  us  the  junc- 
tion is  primarily  of  dynamical,  not  anatomical,  significance.  We  de- 
vote chapter  i  to  the  development  of  the  postulates  of  our  system,  and 
to  the  elaboration  of  a  few  elementary  consequences. 

The  second  stage  in  the  development  of  the  theory  consists  in 
the  investigation  of  the  properties  of  complexes  of  specified  structure, 
these  properties  being  deduced  from  the  postulated  properties  of  the 
units  and  from  their  interrelations  in  the  structure.  The  bulk  of  Part 
1  deals  in  a  purely  abstract  manner  with  a  number  of  different  struc- 
tures which  are  in  some  sense  typical  of  those  required  by  the  appli- 
cations, and  considers  the  general  problem  of  the  reciprocal  determi- 
nations of  the  structural  form  and  the  dynamics. 

The  final  stage  is  the  comparison  of  prediction  with  fact.  A  neu- 
ral complex  of  particular  structure  is  assumed  to  link  the  stimulus 
to  the  response  in  a  given  class  of  cases.  From  this  assumption, 
a  certain  quantitative  functional  relation  between  stimulus  and  re- 
sponse can  be  deduced.  To  the  extent  to  which  experience  verifies  the 
prediction,  we  have  confidence  in  our  initial  assumption  and  are  jus- 
tified in  extending  the  range  of  our  predictions.  In  general  the  func- 
tional relations  involve  variables  and  parameters,  each  capable  of 
assuming  values  over  a  certain  range,  so  that  any  such  relation  yields 
predictions  well  beyond  the  actual  range  of  verification. 

Whether  or  not  verification  occurs  over  some  range,  there  must 
somewhere  occur  a  failure.  This  is  because  both  our  units  and  our 
structures  are  of  necessity  over-simplified.  But  the  failure  is  itself 
instructive,  for  the  trend  of  the  deviations  can  yield  insight  into  the 
nature  of  the  complications  required  for  increasing  the  realism  and 
extending  the  range  of  applicability  of  our  model.  This  is  the  theme 
of  Part  II,  in  which  deductions  made  on  the  basis  of  special  struc- 
tures are  compared  with  data,  where  data  are  available.  Unfortunate- 
ly, even  when  data  of  a  kind  are  available,  these  are  not  always  well 
adapted  to  our  special  purpose.  The  test  of  a  specific  theory  gen- 
erally requires  the  imposition  of  specific  conditions  upon  the  conduct 
of  the  experiment,  and  when  the  theory  is  not  available  to  the  experi- 
menter it  is  largely  chance  if  these  conditions  are  satisfied.  Hence 
some  comparisons  can  be  made  only  in  the  light  of  special  assump- 
tions, and  too  often  no  quantitative  comparison  at  all  is  possible.  It 
is  our  hope  in  publishing  this  monograph  that  more  experiments  will 
be  planned  to  make  these  tests. 

In  Part  III  we  present  the  basis  for  an  alternative  development, 
as  laid  down  quite  recently  by  McCulloch  and  Pitts  (1943).  The  neu- 
ronal dynamics  as  postulated  by  these  authors  is  much  more  realistic, 
but  the  deductions  from  them  of  laws  of  learning  and  conditioning, 

viii 


INTRODUCTION 

of  response-times,  and  of  discrimination,  remains  largely  a  program 
for  the  future.  Their  laws  are  temporally  microscopic,  as  opposed  to 
those  of  Part  I,  which  are  temporally  macroscopic.  It  is  therefore 
to  be  hoped  that  the  macroscopic  laws  can  be  deduced  from  the 
microscopic  ones  as  approximations  valid  at  least  for  certain  com- 
monly occurring  neural  complexes,  and  some  steps  in  this  direction 
are  outlined  in  the  concluding  chapter. 

In  general  we  have  sought,  within  the  available  space,  to  sum- 
marize and  systematize  the  most  important  methods  and  results  to 
date.  We  have  passed  lightly  over  most  of  the  results  already  reported 
in  Rashevsky's  "Advances  and  Applications  of  Mathematical  Biology," 
and  we  have  omitted  all  reference  to  Rashevsky's  aesthetic  theory.  On 
the  other  hand,  many  pages  of  Part  I  have  been  devoted  to  formal  dis- 
cussions making  no  immediate  contact  with  experience.  While  those 
whose  interest  lies  only  in  the  applications  may  wish  to  skip  this  ma- 
terial, the  theoretically  minded  will  recognize  in  these  pages  the 
groundwork  for  the  further  elaboration  of  what  we  hope  will  become 
a  comprehensive  and  unified  theory  of  the  operation  of  the  central 
nervous  system. 


IX 


PART  ONE 


TRANS-SYNAPTIC  DYNAMICS 

The  performance  of  any  overt  act,  by  any  but  the  most  primitive 
of  organisms,  is  accomplished  by  the  contraction  and  relaxation  of 
groups  of  specialized  cells  called  muscles.  Normally  these  contractions 
and  relaxations,  by  whatever  mechanisms  they  may  be  effected,  are  at 
least  initiated  by  prior  events  occurring  at  the  junctions  with  these 
muscles  of  certain  other  specialized  cells  called  neurons.  Whatever 
the  nature  of  these  prior  events,  and  by  whatever  mechanism  they 
are  effected,  they  are  themselves  initiated  by  a  sequence  of  still  prior 
events  in  the  neurons  themselves,  and  these  in  turn  by  yet  earlier 
events  at  the  junctions  of  other  neurons  with  these.  Thus  regressing, 
step  by  step,  we  conclude  that  apart  from  possible  cycles,  pools  of 
perpetual  activity,  the  whole  sequence  was  started  by  an  initial  set 
of  events  at  the  points  of  origin  of  an  initial  set  of  neurons.  And 
finally  this  ultimate  set  of  initial  events — the  set,  or  any  member  of 
the  set,  according  to  convenience,  being  called  a  stimulus — was 
brought  about  by  or  consisted  in  some  physical  or  physiological  occur- 
rence in  the  environment  or  within  the  organism. 

Doubtless  there  are  often  and  perhaps  always  countless  other 
accompanying  events  occurring  within  the  organism  and  interacting 
to  a  greater  or  lesser  degree  with  those  events  here  mentioned,  but 
no  scientific  theory  can  account  for  everything,  and  still  less  for 
everything  all  at  once.  We  wish,  therefore,  to  define  our  schematic 
reacting  organism  as  one  consisting  solely  of  receptors  (sense-or- 
gans) ,  effectors  (muscles) ,  and  a  connecting  set  of  neurons,  the  whole 
and  the  parts  being  affected  by  the  physical  or  physiological  environ- 
ment only  insofar  as  this  acts  as  a  stimulus  via  the  receptors.  We 
wish  to  consider  to  what  extent  behavior  can  be  accounted  for  in 
terms  of  such  a  model.  In  undertaking  such  an  inquiry,  we  freely  and 
expressly  acknowledge  that  much  is  left  out,  and  we  emphatically 
refuse  to  make  any  claim  in  advance  as  to  the  range  of  the  behavior 
that  can  be  so  accounted  for.  This  is  an  empirical  question  to  be  ex- 
perimentally decided.  But  a  hypothesis  cannot  even  be  refuted  until 
it  is  clearly  formulated. 

The  structure  of  a  neuron  is  fairly  complicated  and  its  behavior 
is  hardly  less  so.  Consequently,  to  make  progress  the  neurons,  too, 
must  be  schematized.  Structurally  there  is  a  cell  body  and  two  or  more 


2       MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

threadlike  processes,  but  the  terminations  of  these  processes  are  of 
two  sorts.  A  termination  of  the  one  sort  we  shall  call  an  origin,  one 
of  the  other  sort,  a  terminus.  When  appropriate  stimulation  of  suffici- 
ently high  degree  is  applied  at  an  origin,  there  is  conducted  along  the 
neuron  a  "nervous  impulse"  all  the  way  to  the  various  termini.  This 
nervous  impulse,  arriving  at  a  terminus,  may  contribute  to  the  stimu- 
lation of  any  neuron  which  has  an  origin  at  the  same  place. 

Doubtless  in  all  strictness  the  impulse  does  not  simply  jump  from 
neuron  to  neuron  but  passes  by  way  of  some  intermediary  process 
set  up  in  the  synapse  which  is  the  junction  between  the  two  neurons. 
From  our  point  of  view,  it  is  largely  a  matter  of  convenience  whether 
we  postulate  such  an  additional  process  or  not. 

The  nervous  impulse  manifests  itself  as  a  localized  change  in 
electric  potential,  its  duration  at  any  point  is  about  half  a  millisecond, 
and  it  is  transmitted  at  a  rate  that,  though  low  in  some  neurons,  in 
others  may  equal  or  exceed  10*  cm  sec-1.  Moreover,  in  physiological 
stimulation,  if  the  stimulation  is  maintained,  the  impulses  are  repeat- 
ed and  may  reach  a  frequency  which  is  of  the  order  of  102  sec-1.  The 
more  intense  the  stimulation,  the  more  frequent  the  impulses,  but 
there  is  an  upper  limit  for  any  given  neuron  which  varies  somewhat 
from  neuron  to  neuron.  When  we  have  occasion  to  take  account  of 
it,  we  shall  suppose  this  upper  limit  to  be  a  fixed  characteristic  of 
the  neuron. 

McCulloch  and  Pitts  (1943)  have  developed  a  theory  of  the 
"quantized"  dynamics  of  the  neuron  which  takes  account  of  the  in- 
dividual impulses  and  we  shall  return  to  this  later  at  the  end  of  this 
monograph.  For  the  present,  however,  we  shall  schematize  further 
by  doing  some  statistical  averaging  and  by  fixing  our  attention  upon 
the  synapse  rather  than  upon  the  neuron  itself.  We  shall  choose  the 
alternative  of  supposing  that  the  impulses  of  the  afferent  neurons  are 
not  the  immediate  stimuli  for  the  efferent  neuron,  but  that  these  im- 
pulses start  or  maintain  at  the  synapse  an  intermediate  process  which 
is  the  immediate  stimulus.  To  have  a  concrete  picture,  one  may  imag- 
ine that  some  chemical  substance  is  released  by  the  impulses  and  dis- 
sipated or  destroyed  as  a  monomolecular  breakdown.  However,  it  is 
by  no  means  implied  that  this  is  the  case,  and  furthermore,  we  shall 
not  speak  in  such  terms  but  shall  speak  merely  of  an  "excitatory 
state,"  and  denote  the  state  or  its  intensity  by  s .  More  briefly  we 
shall  speak  of  the  excitation  e. 

The  amount  by  which  the  impulses  increase  s  in  unit  time  is  pre- 
sumably proportional  to  the  frequency  of  these  impulses,  and  the  fac- 
tor of  proportionality  is  taken  to  be  a  characteristic  of  the  fiber.  We 
make  the  simplest  assumption  as  to  the  rate  of  dissipation  of  e  and 


TRANS-SYNAPTIC  DYNAMICS  3 

assume  it  to  be  representable  by  the  term  ae  .  If  we  then  take  cuf> , 
proportional  to  the  frequency,  to  represent  this  rate  of  increase  of  e 
by  the  impulses,  we  obtain  the  equation  (Rashevsky,  1938) 

de/dt  =  a(<j>  —  e)  (1) 

which  we  assume  to  describe  the  development  of  e  .  We  take  e  to  be 
a  measure  of  the  stimulus  acting  upon  any  neuron  which  originates 
at  the  synapse  in  question.  Note  that  by  equation  (1)  we  pass,  in  a 
sense,  directly  from  origin  to  terminus  of  the  neuron,  compressing 
into  the  function  6  our  only  reference  to  the  intra- neuronal  dynamics. 
When  <f>  =  0  the  impulses  have  zero  frequency,  i.e.  do  not  occur,  and 
we  shall  say  the  neuron  is  at  rest.  Nevertheless,  £  is  not  necessarily 
zero,  and  in  fact,  after  the  neuron  has  been  active,  e  vanishes  asymp- 
totically only  according  to  equation  (1)  in  which  </>  =  0  . 

Now  $  is  proportional  to  the  frequency  of  the  generating  im- 
pulses, and  this  is,  as  implied,  an  increasing  function  of  the  applied 
stimulus  with,  however,  a  finite  asymptotic  value.  Hence  we  may 
write 

$=<j>(S),  (2) 

where  S  is  a  measure  of  the  applied  stimulus.  However,  in  order  for 
the  impulses  to  occur,  S  must  exceed  a  certain  minimal  value,  called 
the  threshold,  which  is  characteristic  of  the  neuron  in  question  and 
which  we  shall  denote  by  h .  Hence  <£  (S)  is  zero  for  S  =  h ,  and  for 
S  >  h ,  </>  (<S)  is  positive,  is  monotonically  increasing  but  with  a  de- 
creasing slope,  and  approaches  a  finite  asymptotic  value  for  large 
values  of  5 . 

Relatively  simple  analytic  functions  possessing  these  properties 
for  S  >  h  are  the  following  (Rashevsky,  1938;  in  this  connection  cf. 
Hartline  and  Graham,  1932,  and  Matthews,  1933) : 

<£  =  4>o[l-e-Q<s-»>],  (3) 

«/>0  (S-h)d  +  h 

</>  = log ,  (4) 

log<5  S 

where  6  is  small  and  </>0  is  the  asymptotic  value  of  </> .  For  not  too  large 
values  of  S  either  function  may  be  approximated  by  an  expression  of 
the  form 

<l>  =  a(S-h),  (5) 

and  the  second  by 

4>  =  fi\og(S/h).  (6) 

In  any  case,  for  S  ^  h  ,  <f>  =  0  . 


4      MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

Now  ^  is  a  function  of  S ,  and  S  may  well  be  a  function  of  t ,  in 
which  case  $  is  a  function  of  t .  The  complete  solution  of  equation 
(1)  is  then  given  by 

e  =  e-at[e0  +  a(teaTct>(T)dT],  (7) 

J  0 

where  £0  is  the  initial  value  of  e  .  When,  in  particular,  S ,  and  there- 
fore also  </> ,  is  constant  with  respect  to  time,  this  becomes 

e  =  e-**  e0  +  <£(1  -  e-at).  (8) 

Thus  £  approaches  the  value  <f>  asymptotically,  the  approach  being  in 
all  cases  monotonic,  and  either  increasing  or  decreasing  according  to 
whether  </>  exceeds  or  is  exceeded  by  e.0 . 

If  we  were  now  to  introduce  an  assumption  to  relate  the  muscu- 
lar contraction  with  the  applied  e  ,  we  should  have  a  system  of  for- 
mulae to  be  evaluated  sequentially  along  any  neural  pathway  from 
receptor  to  effector,  for  relating  the  time  and  the  intensity  of  the  re- 
sponse to  the  temporal  form  of  the  stimulus.  But  this  would  obvious- 
ly provide  only  a  very  incomplete  picture.  A  given  stimulus  not  only 
leads  to  the  contraction  of  one  set  of  muscles ;  it  leads  also  to  the  re- 
laxation of  the  antagonistic  muscles.  Any  effective  movement  involves 
both  components,  of  contraction  and  the  inhibition  of  contraction. 
Thus  we  are  inevitably  led  to  extend  our  picture  to  include  the  phe- 
nomenon of  inhibition. 

There  are  many  ways  in  which  such  a  phenomenon  could  be  in- 
troduced into  our  schematic  picture,  but  the  simplest  way  seems  to  be 
to  suppose  that  at  least  some  neurons  have  the  property  of  creating, 
as  the  result  of  their  activity,  an  inhibitory  state  of  intensity  j , 
(briefly,  an  inhibition  j) ,  antagonistic  to  the  excitatory  state  e  ,  and  to 
suppose  that  the  production  of  j  follows  the  same  formal  law  as  that 
for  e: 

dj/dt  =  b{y>-j).  (9) 

The  function  y>  is  of  the  same  type  as  <£  and  it  is  only  as  a  matter  of 
convenience  that  we  introduce  a  separate  symbol. 

Rashevsky  (1938,  1940)  commonly  assumes  that  in  general  the 
activity  of  any  neuron  leads  to  the  production  of  both  s  and  j ,  al- 
though for  particular  neurons,  the  one  or  the  other  may  be  negligible 
in  amount.  Evidently  we  may  always  replace  a  single  neuron  devel- 
oping both  e  and  j  by  a  pair,  one  developing  £  alone  and  one  j  alone. 
It  is  useful,  however,  to  consider  some  of  the  characteristics  of  a 
"mixed"  neuron  of  the  Rashevsky  type. 

Since  £  and  j  are  antagonistic,  we  are  now  supposing  that 

<r  =  s-j  (10) 


TRANS-SYNAPTIC  DYNAMICS  5 

is  the  measure  of  the  effective  stimulus  acting  upon  any  neuron  origi- 
nating at  the  given  neuron's  terminus.  If  at  any  moment  the  o-  due 
to  the  activity  of  a  particular  neuron  is  positive,  we  shall  speak  of 
the  neuron  as  "exciting"  or  having  an  "exciting  effect"  at  that  mo- 
ment, without,  however,  meaning  to  imply  thereby  that  it  then  ex- 
cites any  neuron.  It  will  do  this  provided  only  that  a  exceeds  the 
threshold  of  a  neuron  suitably  placed.  Neither  do  we  imply  that  the 
neuron  which  produced  the  a  is  at  this  moment  acting,  though  it 
must  have  been  acting  in  the  very  recent  past  if  a  is  still  appreciable. 
Likewise,  we  shall  speak  of  the  neuron  as  "inhibiting"  or  having  an 
"inhibiting  effect"  whenever  its  a  is  negative.  Consider  the  case  of  a 
constant  5 ,  so  that  <j>  and  y  are  themselves  constant.  Asymptotically 
£  and  j  approach  </>  and  \p ,  respectively,  so  that  the  neuron  is  asymp- 
totically exciting  or  asymptotically  inhibiting  according  to  the  rela- 
tive magnitudes  of  <p  and  y> .  However,  the  initial  rates  of  increase  of 
£  and  j  are  equal  to  a<}>  and  to  by ,  respectively,  so  that  an  asymptoti- 
cally exciting  neuron — for  which  cf>  >  ip — would  be  momentarily  in- 
hibiting in  case  by>  >  a<j> ,  and  vice  versa.  Thus  the  transient  and  the 
asymptotic  effects  of  such  a  neuron  would  be  quite  opposite. 

Furthermore,  suppose,  for  definiteness,  that  the  neuron  is  asymp- 
totically inhibiting,  \p  >  </> ,  and  consider  the  effect  following  the  cessa- 
tion of  its  own  stimulus,  when  the  neuron,  as  a  result,  comes  to  rest. 
We  suppose  for  simplicity  that  the  constant  stimulus  is  maintained 
long  enough  for  the  asymptotic  state  to  be  reached.  Then,  on  re- 
moval of  the  stimulus,  <£  and  y.>  both  drop  to  zero  so  that  £  and  j  de- 
cline exponentially  to  zero.  If  b  >  a ,  the  decline  of  j  is  more  rapid 
than  that  of  £  and  a  transient  exciting  effect  may,  and  in  fact  always 
does,  ensue  while  the  neuron  is  thus  at  rest. 

To  summarize  all  possible  cases  of  this  sort:  A  neuron  is 
a)  Asymptotically  exciting  whenever 

4>  >  ip . 


/* 

K 

f^  . —  ~"": 

\  \ 

// 

'            J 

\  \ 

/  / 

\  \ 

/  / 

\  \ 

/  / 
1  / 

a<fc 

\  \ 

1  / 

1  / 

a<f><ib 

? 

1  / 

\   >^ 

1/ 

if 

\  \^^ 

1/ 

•v     — ^_ 

Figure  1 


6      MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

Furthermore,  when 

i)     a  <  b  ,  cuj>  >  by  it  is  always  exciting  in  activity  and  transiently 

exciting  at  rest; 
ii)     a  <  b  ,  04  <  by  it  is  transiently  inhibiting  in  activity,  tran- 
siently exciting  at  rest  (Figure  1)  ; 


Figure  2 


iii)     a  >  b  ,  a<j>  >  by  it  is  always  exciting  in  activity,  transiently 
inhibiting  at  rest  (Figure  2). 
The  case  a  >  b  ,  cuj>  <  by  is  inconsistent  with  <£  >  y> . 
b)     Asymptotically  inhibiting  whenever 


Furthermore,  when 


<f>  <xp 


i)     a  >  b  ,  cuj)  <  by  it  is  always  inhibiting  in  activity  and  tran- 
siently inhibiting  at  rest; 
ii)     a  >  b  ,  cuf>  >  by  it  is  transiently  exciting  in  activity,  transient- 
ly inhibiting  at  rest; 
iii)     a  <  b  ,  a$  <  by  it  is  always  inhibiting  in  activity,  transiently 
exciting  at  rest. 

The  case  a  <  b  ,  a</>  >  by  is  inconsistent  with  <£  <  \p  . 

We  have  tacitly  assumed  that  the  <j  at  any  synapse  is  affected  only 
by  neurons  terminating,  not  at  all  by  these  originating,  at  this  syn- 
apse. We  further  suppose  that  if  several  neurons  terminate  on  the 
same  neuron,  the  </s  of  all  of  them  combine  linearly.  It  is  now  ap- 
parent that  with  these  simple  assumptions,  results  of  considerable 
complexity  are  possible.  We  shall  attempt  first  of  all  to  explore  some 
of  these  complexities  in  the  abstract,  and  next  to  relate  a  few  of  them 
to  concrete  psychological  processes. 


II 

CHAINS  OF  NEURONS  IN  STEADY-STATE  ACTIVITY 

From  our  point  of  view,  a  receptor  provides  only  the  mediation 
between  certain  non-neural  events  and  the  occurrence  of  a  stimulus 
S  for  one  or  more  neurons,  and  an  effector  provides  the  mediation 
between  the  occurrence  of  an  excitation  a  produced  by  neurons  and 
certain  other  non-neural  events.  We  shall  not  investigate  these  medi- 
ations, but  shall  consider  only  the  problem  of  relating  a  to  S  for  any 
given  neural  structure.  Some  of  the  neural  structures  actually  to  be 
found  in  the  higher  forms  are  of  bewildering  complexity,  so  that 
merely  to  describe  them  from  observation  is  the  task  of  a  lifetime. 
We  can  hope  to  progress  with  our  problem  only  if  we  start  with  very 
simple,  hypothetical  structures. 

The  simplest  possible  connected  structure  of  more  than  one  neu- 
ron is  a  chain  of  neurons.  In  a  chain,  every  neuron  but  one  has  an 
origin  which  coincides  with  the  terminus  of  some  other.  Call  this 
one  neuron  N0 ,  its  origin  s0  and  its  terminus  st .  Let  2Va  be  the  neu- 
ron whose  origin  is  at  Si ,  let  &>  be  its  terminus,  and  so  sequentially, 
the  terminus  of  the  last  neuron  of  a  chain  of  n  neurons  being  sn  .  If  a 
stimulus  S  is  applied  to  N0  at  s0 ,  it  may  come  from  a  receptor  or  from 
a  neuron  or  neurons  not  in  the  chain.  If  Nn-X  develops  o-  at  sn ,  this 
may  act  upon  an  effector  or  upon  a  neuron  not  in  the  chain.  That  is 
immaterial.  Suppose  that  the  neurons  of  our  chain  are  all  of  the 
simple  excitatory  type.  Suppose,  further,  that  only  a  negligible  time 
is  required  for  the  o-(=  e)  produced  by  any  neuron  to  reach  its 
asymptotic  value  (/>  when  a  constant  stimulus  S  is  applied.  In  other 
words,  we  are  now  considering  the  chain  only  in  its  asymptotic  state 
after  stimulation  by  a  constant  stimulus.   Then  if 

u0  =  S 

is  the  total  stimulus  acting  at  s0  upon  N0 ,  it  follows  that 

o"i  —  </>o  (o"o) 

is  the  a  produced  by  N.0  at  sx ,  where  fo  is  the  ^-function  of  N0 .  If 
no  receptor  or  neuron  outside  the  chain  introduces  any  S  or  a  at  s1 , 
then 

(T2  =  <£i((Ti)    =  <£l[<£o(0"o)] 

is  the  o-  produced  by  Nx  at  s2 .  Thus  we  can  calculate  sequentially  all 
the  o-i . 


8      MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

Define  the  functions  <f>i,v(o)  by  recursion  formulas 

^>i,v+i(c7)    =  0i+v+i    [<f>i,v(<r)]    .  (1) 

Then 

Ci+v+i  —  </>j,v  \Oi)  ,  (2) 

so  that  each  function  <f>i>v  gives  the  a  produced  at  si+v+]  in  terms  of 
that  present  at  Si .  If  the  derivatives  exist,  then 

0'i,v+i(<r)  —  </>'i+v<-i[<£i,v(<x)]  </>'i+v[^);,v-i  (cr)]  •  •  •  <j>' i  (<r)  ,  (3) 

where  the  primes  denote  the  derivatives.  Hence  if,  as  we  suppose, 
each  <j>i(o)  is  monotonic,  so  is  each  4>'i,v(<r)  ;  and  if  each  <£'i(cr)  is  de- 
creasing, so  is  each  </>'i,v(a).  Further,  if  each  ■/>'<  vanishes  asymptoti- 
cally, so  does  each  <f>'itV .  Hence  these  "higher  order"  excitation-func- 
tions are  functions  of  the  same  type  as  the  ordinary  ones,  and  we 
may  always  replace  any  such  chain  of  neurons  by  a  single  one,  at 
least  if  we  are  interested  in  the  asymptotic  behavior  alone. 
If  every  </>'»  (a)  ^1  for  all  a  ,  then  the  sequence 

Co  ,  ffi  ,  (J?  ,  •  •  •  ,  trn 

is  decreasing.  In  fact,  if  hi  is  the  threshold  of  Nt ,  then 

<7i+l    <    CTj    ~~    Aij  , 

and  if  h  is  the  lowest  threshold  in  the  chain, 

<j%  *C  <r0       Irl  , 

so  that  the  number  of  neurons  in  the  chain  that  can  be  excited  is  not 
greater  than  the  greatest  integer  in  a.0/h  . 

If  the  <j>'i(a)  >  1  for  small  values  of  a  ,  this  is  not  necessarily  the 
case.  It  has  been  shown  (Householder,  1938a)  that  when  all  the  neu- 
rons are  identical,  and  the  chain  is  long,  the  a,  will  then  either  di- 
minish to  zero,  or  approach  a  certain  positive  limit  characteristic  of 
the  chain,  according  to  whether  <r.0  lies  below  or  above  a  certain  criti- 
cal value.  The  limit  and  the  critical  value  are  the  two  roots  of  the 
equation 

£'(<0=i. 

Note  that  it  is  legitimate  to  drop  the  subscript  when  all  the  <£*  are 
identical. 

If  all  of  the  ct's  are  within  the  range  that  permits  of  a  linear  ap- 
proximation to  the  ^  ,  it  is  easy  to  obtain  an  analytic  expression  for 
these  ctj  in  terms  of  a0  and  the  subscript  i  (Landahl,  1938b) .  We  have, 
in  fact  [chap,  i,  equation  (5)] 


CHAINS  OF  NEURONS  IN  STEADY-STATE  ACTIVITY  9 

o-i  =  a(tr0  ~  h) , 

<r2  =  a(<T!  —  h)  =a2<r0  —  ah(a  +  1), 

ui  =  a(ai_1  -  h)  =a*a0  -  aMa*-1  +  a*"2  +  •••  +  1), 
or 

(ah    \         ah 
«o+- )-- ,  a^l,  (4) 

1  —  a  /       1  —  a 

vi  =  o0  —  ih  ,     a  =  1  .  (5) 

Hence  if  a  =  1 ,  the  a;  decrease  in  arithmetic  progression  until,  for 
some  i  ,<nfkh ,  after  which  all  succeeding  <ri+v  are  zero.  Otherwise  the 
sequence 

<n  +  ah/(l  —  a) 

forms  a  geometric  progression.  The  progression,  and  hence  the  se- 
quence <Ji ,  increases  when 

a  >  1  ,  <t0  >  ah/ (a  —  1). 

When  the  second  inequality  is  reversed  but  the  first  holds,  the  pro- 
gression consists  of  negative  terms  which  increase  numerically  until 
some  <n  falls  below  threshold.  If  the  first  inequality  is  reversed,  the 
progression  is  decreasing.  Finally,  if  the  second  inequality  becomes 
an  equality,  then  every  a*  has  the  same  value. 

So  far  we  have  been  supposing  that  the  only  excitation  intro- 
duced from  the  outside — from  receptors,  that  is,  or  from  neurons 
which  are  not  members  of  the  chain— was  introduced  at  So  alone.  We 
have  further  supposed  all  the  neurons  in  the  chain  to  be  excitatory, 
that  is,  asymptotically  exciting  while  acting,  since  evidently,  in  such 
a  situation,  no  activity  could  occur  beyond  an  inhibiting  member  of 
the  chain.  We  turn  now  to  a  somewhat  more  general  situation  in 
which  the  chain  may  contain  neurons  which  are  asymptotically  in- 
hibiting while  acting,  and  in  which  outside  excitation  may  be  present 
at  any  or  all  the  Si .  Following  a  suggestion  made  by  Pitts  (1942b), 
we  now  represent  the  function  <j>  as  linear  until  the  stimulus  reaches 
a  certain  maximal  value,  and  constant  at  the  upper  limit  thereafter. 
This  representation,  though  no  doubt  less  accurate  than  the  functions 
(3)  and  (4)  of  chapter  i,  is  at  any  rate  a  fair  first  approximation, 
and  is  much  more  easily  handled.  Let  St  be  the  applied  stimulus  at 
Si ,  and  define  the  quantities 

£i  =  Si  —  hi ,    rji  =  £i  +  <Ti .  (6) 

Then  r\i  represents  the  excess  over  the  threshold  of  the  total  stimulus 


10   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

acting  upon  neuron  A/*  .  In  accordance  with  our  description  of  the 
0's,  therefore,  we  have 

CTi+i  =  ^>i(^i)  (7) 

where 

4>i  (v> )  —  0  when  rji  ^  0  , 

<f>i  (rji)  =  a»  rji  when  0  <  rji  <  H% ,  (8) 

</>«  (rji)  =  an  Hi  when  rji  ^  #;  . 

The  coefficient  a*  we  shall  call  the  activity-parameter  of  2V< ,  and  it 
may  be  positive,  for  an  excitatory,  or  negative,  for  an  inhibitory  neu- 
ron, but  is  not  zero.  Our  problem  is  the  following:  supposing  Si , 
52 ,  •  •  •  ,  Sn  fixed,  to  express  rjn  as  a  function  of  rj0  —  S0  —  K ,  and, 
more  generally,  to  express  rji+v  as  a  function  of  rji  when  Si+1 ,  •••  ,  S^v 
remain  fixed. 

Since  each  rji+1  varies  linearly  with  rjx  when  the  latter  occupies  a 
certain  restricted  range,  and  rjin  is  otherwise  constant,  it  is  at  once 
apparent  that  the  same  is  true  of  the  variation  of  any  iji+v  with  rji . 
Again,  since  any  &  may  be  so  large  that  rji  always  exceeds  Hi ,  or  so 
small  that  rji  is  always  negative,  it  is  evident  that  rjuv  may  remain 
constant  for  all  values  of  rji .  In  such  a  case  si+v  is  said  to  be  inac- 
cessible to  Si  (Pitts,  1942a)  ;  otherwise  it  is  accessible.  More  explic- 
itly, if,  as  rji  varies  over  all  values  from  —  oo  to  +  oo  while  Si+1 ,  ••• , 
Si+v  remain  fixed,  the  value  of  rji+v  remains  constant,  then  si+v  is  inac- 
cessible to  Si .  Clearly  si+1  is  always  accessible  to  Si .  If  si+v  is  inac- 
cessible to  Si ,  then  si+v  is  also  inaccessible  to  any  s      ,  and  further , 

i—v' 

any  s         is  inaccessible  to  Si .   Finally ,  it  is  clear  that  if  Si+V  is  acces- 

i+v+v' 

sible  to  Si ,  then  where  rj^v  varies  with  rji ,  it  decreases  if  there  is  an 
odd  number,  increases  if  an  even  number  of  inhibitory  neurons  (with 
negative  a's)  between  Si  and  Si+V . 

The  above  conditions  for  inaccessibility  may  be  phrased  thus: 

//  any  of  the  four  following  conditions  holds: 


a*  >  0  , 

S^  +  aiHi^O, 

ai  >0, 

£i+l  =  "t+1  > 

<Xi<0, 

li+i  ^  o , 

ai<0, 

Si+i     i     <Xj  Jtl  i  =  "i+i  , 

Mew  Si+2  *s  inaccessible  to  s% .  Otherwise  si+2  is  accessible  to  Si . 

These  conditions  are  also  sufficient  for  the  inaccessibility  of  any 
si+2+v  to  any  s.     ,  where  v  and  v  are  non-negative  integers,  but  the 

i-v' 

conditions  are  not  necessary. 

The  following  conditions  for  accessibility  are  somewhat  less  ob- 
vious. Let  us  say  of  a  neuron  AT*  whose  stimulus  exceeds  its  threshold 


CHAINS  OF  NEURONS  IN  STEADY-STATE  ACTIVITY  11 

by  more  than  Hi  that  Nt  is  in  a  state  of  maximal  activity.  Then  (cf. 
Pitts,  1942a) : 

A.  Let  there  be  an  odd  number  of  inhibitory  neurons  between  Si  and 
Sj .  Then  if,  for  given  Si+1 ,  •  •  •  ,  Sj ,  it  can  occur  that  both  N%  and  ZV, 
are  inactive,  or  that  both  are  in  maximal  activity,  sj+1  is  inaccessible 
to  Si ,  and  a  fortiori,  sn  to  s0 . 

B.  Let  there  be  an  even  number  of  inhibitory  neurons,  or  none,  be- 
tween Si  and  Sj .  Then  if,  for  given  Si+1 ,  •  •  •  ,  S, ■ ,  it  can  occur  that  Nj 
is  at  rest  while  Ni  is  in  maximal  activity,  or  that  Nj  is  in  maximal 
activity  while  Ni  is  at  rest,  s,+1  is  inaccessible  to  s;  ,  and  a  fortiori, 
sn  x>o  Sq  . 

Consider  case  A,  the  first  alternative.  Making  Ni  active  could 
decrease,  but  could  not  increase,  the  activity  of  Nj  and  hence  could 
not  initiate  activity  in  Nj .  Similarly  in  the  second  alternative,  dimin- 
ishing the  activity  of  Ni  could  increase,  but  could  not  decrease  that  of 
Ni ,  and  since  Nj  is  already  in  maximal  activity,  the  resulting  a  can 
not  be  further  increased.   Analogous  considerations  apply  to  case  B. 

Accessibility  is  denned  only  with  reference  to  a  given  distribu- 
tion of  the  stimulation,  supposedly  fixed,  applied  at  all  synapses  of 
the  chain  except  at  the  origin.  However,  if,  the  distribution  being 
fixed,  sn  is  accessible  to  s0 ,  then,  still  with  the  same  applied  distribu- 
tion, rjn  is  a  linear  function  of  t]0  when  ^  lies  between  certain  limits, 
and  elsewhere  rjn  is  constant.  Now  each  rji  defined  the  excess  of  the 
total  stimulation  at  Si  over  the  threshold  hi  of  Ni .   Hence  if 

Vi  =  rji  +  hi 

is  the  total  stimulation,  yi  is  a  linear  function  of  y0  when  y0  lies  be- 
tween suitable  limits.  This  is  the  property  postulated  of  single  neu- 
rons for  this  discussion.  To  get  an  explicit  representation,  let 

E  =  a,,-!  (Sn-!  —  /&„-i)  +  •  •  •  +  an_a  a„_2  •  •  •  at  (Sx  —  h^) 

-  a„_i  •  •  •  a0  h0  ,  (9) 

A  Ctn_i  0Cn_2  ■  •  •  (Xo  . 

Then  for  suitable  y'  and  y"  we  have 

yn  =  Sn  +  3  +  Ay'     when  yfl  =  V' , 

yn  =  Sn  +  S  +  Ay0     when  y'  <  y0  <  y" ,  (10) 

yn  =  Sn  +  S  +  Ay"     when  y0  ^  y" . 

The  quantities  E  ,  y  and  y"  depend  upon  the  applied  stimulation.  The 
A  does  not. 

With  reference  to  the  steady-state  dependence  of  <r  upon  S ,  the 
properties  of  a  chain  of  neurons  are  seen  to  be  very  similar  to  those 
of  a  single  neuron.  This  is  true  especially  in  the  case  when  a  constant 


12   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

S ,  coming,  perhaps,  from  self-exciting  circuits  (chap,  iv),  is  applied 
to  each  of  the  intermediate  neurons  of  the  chain,  and  if  the  linear 
representation  is  adequate  the  properties  of  the  chain  can  be  made  ex- 
actly the  same  as  those  of  an  individual  neuron.  Thus  it  is  legitimate  to 
speak  of  two  centers  as  connected  by  a  single  neuron  even  when  it 
would  be  more  plausible  on  anatomical  grounds  to  suppose  that  an 
entire  chain  is  required.  On  the  other  hand,  chains  can  exhibit  prop- 
erties quite  different  from  those  of  a  single  neuron,  since,  in  particu- 
lar, more  variables  enter,  as  is  clear  from  equations  (9)  and  (10). 

The  utility  of  the  notion  of  accessibility  will  become  more  appar- 
ent when  the  general  net  is  discussed  in  chapter  v,  but  it  is  perhaps 
sufficiently  evident  already  that  the  properties  of  very  complicated 
nets  might  in  special  cases  turn  out  to  be  very  simple  because  of  the 
inaccessibility  of  certain  centers  to  certain  others.  It  is  evident,  too, 
that  whereas  we  have  discussed  accessibility  only  in  connection  with 
the  linear  representation  of  <j>(S),  very  similar  results  must  hold 
in  general. 


Ill 

PARALLEL,  INTERCONNECTED  NEURONS 

Color-contrast  and  visual  illusions  of  shape  provide  well-known 
examples  of  the  almost  universal  interdependence  of  perceptions.  Phy- 
siologically no  stimulus  occurs  in  the  absence  of  all  others,  and  the 
response  to  any  stimulus  depends  in  part  upon  the  nature  of  the  back- 
ground against  which  it  is  presented.  One  may  wish  to  say,  indeed, 
that  the  true  stimulus  to  the  organism  is  the  whole  situation,  but 
since  we  cannot  discuss  any  whole  situation,  and  since  the  whole  situ- 
ation is  never  duplicated,  such  terminology  does  not  seem  to  serve 
any  useful  scientific  purpose. 

If  two  stimuli  differ  only  in  degree,  it  may  be  true  that  the  re- 
sponses which  they  evoke  differ  only  in  degree,  the  stronger  stimulus 
evoking  the  stronger  response.  But  in  many  instances  there  is  a  com- 
plete change  in  the  form  of  the  response,  and  in  others  it  is  the 
weaker  stimulus,  and  not  the  stronger,  which  brings  forth  the  strong- 
er response. 

In  our  schematic  reacting  organism,  such  phenomena  are  easily 
understood  in  terms  of  suitable  interconnections  between  parallel  neu- 
rons. We  are  reserving  Part  II  for  the  precise  formulations  neces- 
sary to  make  quantitative  predictions,  so  that  we  content  ourselves 
here  with  a  few  qualitative  results  to  indicate  in  general  how  this 
comes  about. 

In  the  barest  terms,  if  two  stimuli  which  differ  only  in  degree 
lead  to  responses  which  differ  in  form,  then  there  are  pathways — 
neural  chains  from  receptor  to  effector — which  can  be  traversed  when 
the  impulses  are  initiated  by  a  stimulus  within  a  given  range  of  in- 
tensities but  not  when  these  are  initiated  by  a  stimulus  lying  outside 
this  range  on  the  scale  of  intensities.  The  simplest  neural  mechanism 
having  this  property  consists  merely  of  two  neurons,  Ne  excitatory 
and  Ni  inhibitory,  having  a  common  origin  and  a  common  terminus 
(Landahl,  1939a).  Let  he  and  hi  be  the  thresholds  of  Ne  and  Ni ,  re- 
spectively, and  let  h  be  the  threshold  of  some  neuron  N  originating 
at  the  common  terminus  of  the  two  neurons.  Suppose 

hi>  he  ,y>(oo)  >  <p(  oo  )f(f>  (hi)  >h. 

Then  if  S  is  sufficiently  near  to  hi  in  value  (Figure  1),  asymptotically 

<r  =  </>(£)  ~y(S)  >h 

and  N  will  become  excited,  whereas  a  somewhat  larger  S  will  result 

13 


14    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


Figure  1 

in  a  negative  a  at  the  origin  of  N ,  and  a  somewhat  smaller  one  will 
yield  a  sub-threshold,  though  positive,  a .  There  will  be  some  range 
h',  h" ,  therefore,  which  contains  h, ,  and  within  which  S  must  lie  if 
transmission  is  to  occur. 

These  are  not  the  only  possible  relations  that  will  limit  the  range 
over  which  transmission  may  occur.  We  could  have,  for  example 
(Figure  2), 


-  v 


hi=  he 


Figure  2 


hi  =  he,     <j>' (hi)  >  y>'(hj),     y(oo)  >  </>(oo). 

Then  for  an  S  within  a  limited  range,  </>  exceeds  y> ,  which  is  all  that 
is  required  except  that  h  must  be  sufficiently  small.  Suppose,  then, 
one  has  a  number  of  such  sets,  Ni ,  Ne  and  N  ,  all  with  a  common 
origin,  and  each  possessing  a  characteristic  range.  If  these  ranges  do 
not  overlap,  and  if  each  set  is  connected  through  a  chain  to  a  par- 
ticular effector,  then  any  S  will  excite  only  the  effector  corresponding 
to  the  particular  range  on  which  S  lies. 

As  a  first  step  in  discussing  the  interaction  of  perceptions  (strict- 
ly the  interaction  of  the  transmitted  impulse),  and  as  a  kind  of  gen- 


PARALLEL,  INTERCONNECTED  NEURONS 


15 


eralization  of  the  mechanism  just  discussed,  consider  the  following 
(cf.  Rashevsky,  1938,  chap.  xxii).  The  neurons  N„  and  N22  are  ex- 
citatory, originating,  respectively,  at  s1  and  s2,  terminating,  respec- 
tively, at  s\  and  s'2 .  The  neurons  N12  and  AT.21  are  inhibitory,  originat- 
ing, respectively,  at  sx  and  s2 ,  terminating,  respectively,  at  s'2  and  s\  . 
Let  us  restrict  ourselves  here  to  a  range  of  intensities  over  which  the 
linear  approximations  to  the  functions  </>  and  y>  are  adequate.  Then, 
stimuli  Si  and  S2  being  applied  at  s3  and  s2 ,  we  have 


<*i  —  «n  (Si  —  /in)  +  a21  (S2  —  h21), 
a2  =  ai2  (Si  —  /i-i2)  +  a22  (S2  —  /l22) , 


(1) 


when  the  quantities  within  the  parentheses  are  all  positive.  When 
any  of  these  quantities  within  parentheses  is  negative,  however,  the 
term  is  deleted.  The  conditions  for  the  excitation  of  2v\  and  N2  are, 
respectively,  a1  >  hx  and  <t2  >  h? . 

A  geometric  representation  of  these  conditions  is  easily  obtained 
on  the  (Si  ,  S2) -plane.  The  graph  of  the  relation  <ti  =  hx  is  a  broken 
line  consisting  of  a  single  vertical  ray  of  abscissa  hxl  +  hjaxi  extend- 
ing downward  to  infinity  from  the  ordinate  hzX ,  and  a  ray  extending 
up  and  to  the  right  with  slope  -an/a21  .  Note  that  this  slope  is  posi- 
tive, since,  N21  being  inhibitory,  a2]  is  negative.  Then  the  region  de- 
fined by  o-i  >  hi  is  that  to  the  right  of  and  below  the  broken  line.  Like- 
wise the  region  defined  by  or2  >  Ju  consists  of  those  points  above  and 
to  the  left  of  a  certain  broken  line  which  consists  of  a  horizontal  ray 
extending  to  the  left,  and  a  ray  of  positive  slope.  If  these  regions 
overlap  (Figure  3),  it  is  possible  to  have  both  Nx  and  N2  acting  simul- 
taneously.  Otherwise  it  is  not  possible. 


Figure  3 


These  regions  necessarily  overlap  if  the  determinant  of  the  co- 
efficients in  equation  (1)  is  positive: 


16   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

an  a21 


O.X2.  tX22 


>0,  (2) 


for  then  the  line  o-i  =  hx  is  steeper  than  a2  =  h2 ,  so  that  for  large  Sx 
and  S2  the  magnitudes  can  be  so  related  that  both  the  inequalities, 
ax  >  hx  and  «r2  >  hn ,  are  satisfied. 

The  case  when  condition  (2)  fails  is  of  some  interest  (Rashev- 
sky,  1938,  chap.  xxii).  Now,  if  neither  corner  lies  in  the  other  re- 
gion, the  rays  do  not  intersect,  and  it  is  never  possible,  with  any  Si 
and  S2 ,  for  both  Nx  and  N2  to  be  excited  at  the  same  time.  In  fact, 
even  with  strong  stimuli,  the  point  (Sx ,  S2)  may  be  outside  both  re- 
gions and  neither  Nx  nor  2V2  is  excited.  On  the  other  hand,  if  either 
corner  does  lie  in  the  other  region,  the  rays  do  intersect,  and  there 
is  a  finite  region  of  overlap  as  illustrated  in  Figure  3.  Points  (Si ,  S2) 
beyond  the  intersection  and  between  the  two  rays  represent  pairs  of 
stimuli  which,  though  strong,  fail  to  excite  either  Nt  or  N2  .  The 
analytic  condition  for  this  is  the  simultaneous  fulfilment  of  the  two 
inequalities 

an  (h12  —  &n  —  fei/flu)  —  a21(/^i  —  h22  —  fh/a22)  >  0  , 

(3) 

—  <x12(h12  —  h22  —  ftj/on)  +  a22(h2l  —  /k,2  —  h>/a22)  >  0  , 

together  with  the  failure  of  equation  (2). 

The  situation  here  described  may  be  thought  of  as  that  of  two 
stimuli  competing  for  attention.  When  conditions  (3)  hold  and  (2) 
fails,  there  are  moderate  stimuli  which  lead  to  excitation  of  both 
2V"i  and  2V2  (awareness  of  both  stimuli) ,  while  with  more  intense  stim- 
uli, unless  one  is  sufficiently  great  as  compared  with  the  other,  each 
stimulus  prevents  the  response  appropriate  to  the  other  and  no  re- 
sponse occurs. 

For  the  special  case  in  which  the  mechanism  is  altogether  sym- 
metric (Landahl,  1938a;  cf.  also  chap,  ix)  we  may  set 

a11  =  a22  =  a,      —  a12  =  —  a21  =  P, 

(4) 

If  the  two  responses  are  incompatible  in  nature  the  parameters  might 
be  so  related  that  the  two  conditions  (3)  cannot  be  satisfied.  The 
failure  of  these  reduces  to  the  single  inequality 

h"^a(h'-h)  (5) 

which  holds  necessarily  in  case  we  have  h'  ^  h .  If  the  relation  (5) 
is  replaced  by  an  equation,  we  have  a  kind  of  discriminating  mechan- 


PARALLEL,  INTERCONNECTED  NEURONS  17 

ism  by  which  the  stronger  of  two  simultaneous  stimuli  elicits  its  ap- 
propriate  response  and   prevents   the   other   response.    If,   further, 

a1  =  S1-  S2+  (h'  -h), 

and  if,  finally,  h'  and  h  are  very  nearly  equal,  the  transmitted  stimu- 
lus approximates  the  absolute  value  of  the  difference  between  the 
two  stimuli. 

More  generally,  let  the  excitatory  neuron  NH  connect  s,  with 
s'i  (i==  1 ,  ••• ,  n)  (Figure  4),  let  the  inhibitory  neuron  Na(i  ¥=  j)  con- 

N11                                                                 N, 
S^^—^ , ^^.jt> 

So><L^ U- VT      » ^r^H  > ' 

2         — ><^^—""« 


*-""Va . ->3^^_?j3. 


Figure  4 

nect  St  with  s'j  and  let  a  ,  —  ft ,  h ,  li  and  ft"  be  the  activity  parameters 
and  the  thresholds  of  the  various  neurons.  If  all  neurons  of  the  first 
level  are  active, 

Vi  =  a(Si-h)  -/J2  (Sj-hf)  ,  (6) 

and  the  conditions  for  excitation  of  N\  (originating  at  s\)  is  oi  >  h". 
If  h  <  h',  then  for  values  of  the  Sj  between  h  and  h'  the  excitatory  but 
not  the  inhibitory  neurons  are  excited.  If,  further, 

a(h'  -h)  >h" ,  (7) 

then  for  values  of  the  Sj  near  h'  the  neurons  Ni  are  all  excited.  But 
with 

n  >  1  +  a/0  ,  (8) 

when  all  the  Si  are  equal,  the  <n  are  equal,  and  if 

P(n-l)h'  -  ah-h" 

S^ ,  (9) 

(w-1)  0-a 

none  of  the  Ni  responds,  though  for  somewhat  smaller  values  of  £ 
they  all  respond. 

With  the  same  relations  among  the  parameters,  suppose  the  Si 
are  not  all  equal,  but  that  each  exceeds  h' .  It  is  no  restriction  to  sup- 
pose that 


18   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

Evidently  if  some  but  not  all  the  Ni  are  responding,  there  must  be 
some  ra  ^  1  such  that  Nlf  N2 ,  ••• ,  Nm  are  responding,  Nm+1 ,  •  •  •  ,  Nn 
are  not.  Then  for  i  =  m,  am  as  given  by  equation  (6)  must  exceed, 
o-OT+i  fail  to  exceed  h" ,  and  if  S  represents  the  mean  of  all  the  Si  this 
leads  to  the  relation 

nfiS  +  ah-  (n-l)ph'  +  h" 

£>m  ->  "  =  Om+1*  (10) 

a  +  p 
In  particular  if 

Oi  ==  O2  — —  •  •  •  — —  o«i  — —  o    , 
&in+l  Orft+2  *  *  *  on  O      , 

then  these  relations  are  equivalent  to 

[a-/?(ra-l)]  S'  -  P(n-m)S"  >  ah-  (n-l)ph'  +  h" 

(11) 

^  [a  -  P(n-m-l)]  S"  -  fimS'  . 

In  either  case  the  m  stimuli  Si  produce  their  response  and  prevent 
the  occurrence  of  the  response  to  the  other  n—m  stimuli  (cf.  Rashev- 
sky,  1938,  chap,  xxii ;  Landahl,  1938a) . 

Receptors  in  the  skin  and  the  retina  are  far  too  numerous  to  be 
treated  by  the  simple  algebraic  methods  so  far  employed.  Here  we 
must  think  in  terms  of  statistical  distributions.  The  receptors,  or  at 
least  the  origins  of  the  neurons  to  be  discussed,  may  have  a  one-,  a 
two-,  or  a  three-dimensional  distribution.  According  to  the  case,  let 
the  letter  x  stand  for  the  coordinate,  the  coordinate-pair,  or  the  co- 
ordinate-triple of  the  origin  of  any  neuron.  Let  x'  represent  the  co- 
ordinate, the  coordinate-pair,  or  the  coordinate-triple  of  any  terminus 
of  one  of  these  neurons.  Running  from  the  region  x  ,  dx  (consisting 
of  points  whose  coordinates  fall  between  the  limits  x  and  x  +  dx)  to 
the  region  x',  dx'  may  be  excitatory  or  inhibitory  neurons,  or  both. 
If  we  consider  only  the  linear  representation  of  the  functions  <j>  and 
tp ,  each  neuron  is  characterized  by  the  two  parameters  a  and  h .  To 
consider  a  somewhat  more  general  type  of  structure  than  the  one  just 
discussed  for  the  discrete  case,  let 

N  (x  ,  x' ,  a  ,  h)  dx  dx'  do.  dh 

represent  the  number  of  neurons  originating  within  the  region  x  ,  dx  , 
terminating  in  x' ,  dx' ,  and  characterized  by  parameters  on  the  ranges 
a,  da  and  h ,  dh .  Then,  S(x)  being  the  stimulus-density  at  x ,  the 
a-density  which  results  from  these  neurons  alone  at  x'  is 

N(x  ,  x' ,  a  ,  h)  [S(x)  —  h]  dx  da  dh 


PARALLEL,  INTERCONNECTED  NEURONS  19 

provided  h  <  S (x) .  Hence  the  total  a-density,  obtained  by  summing 
over  the  entire  region  (x),  over  all  values  of  a  (positive  and  nega- 
tive), and  over  all  values  of  h  <  S(x)  is 

a(x')  =J(xJ^Jso^aN(x,x',a,h)[S(x)  -K]  dhdadx.    (12) 

Corresponding  expressions  can  be  derived,  of  course,  on  the  suppo- 
sition that  <f>  and  xp  are  non-linear  of  any  prescribed  form  (Rashev- 
sky,  1938,  chap.  xxii). 

Instead  of  writing  the  special  form  of  expression  (12)  for  the 
strict  analogue  of  the  discrete  case  considered  above,  let  us  suppose 
next  that  the  inhibitory  neuron  Na  ,  rather  than  passing  from  Si  to  s'j , 
passes  from  s\  to  s'j  (Rashevsky,  1938,  chap.  xxii).  The  net  a  at  any 
s'j  is  then  equal  to  the  a  produced  by  excitatory  neurons  terminating 
here  diminished  by  the  amount  of  inhibition  produced  by  the  inhibi- 
tory neurons  originating  at  the  other  s'(-  ,  whereas  it  is  this  net  a  at 
the  s'i  which  acts  as  the  stimulus  for  these  inhibitory  neurons.  We 
have  therefore  to  solve  an  integral  equation  in  order  to  determine  the 
net  a. 

In  the  continuous  case,  let  a(x)  be  the  gross  a-density  produced 
by  the  excitatory  neurons  terminating  in  the  region  x  ,  dx  .  For  the 
inhibitory  neurons,  let  —  p  =  a  represent  the  activity  parameter.  Let 
N(x'  y  x ,  /}  ,'h)  dx'  dx  dp  dh  represent  the  number  of  inhibitory  neu- 
rons which  run  from  the  region  x' ,  dx'  to  x ,  dx  ,  and  have  activity 
parameters  and  thresholds  limited  by  the  ranges  p ,  dp  and  h ,  dh . 
Then 

a(x)  =o(x)   — 

(13) 
f  x>  lo  IoiX,)  PN(X'  'X'P'W&ix')  ~  &]  dhdfidx'. 

This  is  an  integral  equation,  but  one  in  which  the  unknown  func- 
tion enters  as  one  of  the  limits  of  integration.  If  we  interchange 
orders  of  integration,  we  have  as  a  form  equivalent  to  equation  (13) 

a(x)  =7(x)  — 

(14) 
J"/00/  0N(x'  ,x,p,h)[<r(x')  -h]  dx'dfidh. 

0  0  <T(X')>h 

Then  if,  as  a  particular  case,  all  inhibitory  neurons  have  the  same  ft 
and  the  same  h  ,  this  becomes 

a(x)  =a(x)   -ft  J   N(X'  ,x)[<t(x')   -  K\  dx'  .  (15) 

It  is  easy  to  solve  this  equation  in  certain  special  cases.  Suppose 
N(x' ,  x)  is  independent  of  x'  and  x .  It  is  clear  that  the  integral  is 
then  independent  of  x  ,  and  we  may  write 


20   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

a(x)  =  a(x)  ~  XI  ,  (16) 

where 

I=f      [*(x')  -h]dx',k  =  pN.  (17) 

Now,  if  we  knew  the  limits  of  the  region  over  which  a  >  h ,  we  could 
substitute  expression  (16)  for  a  into  the  integral  in  expression  (17), 
integrate,  solve  for  / ,  and  finally  place  this  value  in  relation  (16)  to 
obtain  a .   Not  knowing  these  limits,  we  proceed  as  follows.   Since  o- 

and  a  differ  only  by  a  constant,  the  limits  of  the  region  can  be  defined 
by  the  equation 

«{x)=[i,  (18) 

for  a  suitable  ii .  Leaving  /u  for  the  moment  undetermined,  we  carry 
through  the  steps  as  outlined  except  that  the  range  of  the  integration 

is  to  be  defined  by  o-  >  n  .  We  first  obtain 

fa(x')dx'  -  hM(fx) 

/</*)  = (19) 

1  +  XM(ti) 

where  M([i)  is  the  measure  (length,  area,  or  volume,  according  to  the 

dimensionality)  of  the  region  o-  >  ii  .  But  since  a  >  ii  and  <r  >  h  define 
the  same  region,  it  follows  from  (16)  that 

it-XI(ii)=h.  (20) 

Hence  if  we  solve  this  equation  for  /a  ,  then  we  find,  by  equations  (16) 
and  (20),  that 

(x)  =7(x)  +h-  p.  (21) 


si 


It  is  evident  that  the  procedure  here  outlined  is  applicable,  with  obvi- 
ous modifications,  in  case  AT  is  a  function  of  x'  but  is  independent  of  x  . 
From  the  fact  that  a  and  cTdiff er  only  by  a  constant,  certain  prop- 
erties of  the  solution  are  at  once  apparent.  If  o-  anywhere  exceeds  h  , 
then  o-  must  somewhere  exceed  h  .  For  if  a  nowhere  exceeded  h ,  then 

I  =  0  ,  a  =  a ,  and  we  have  a  contradiction  in  the  fact  that  o-  itself 
somewhere  exceeds  h  .  Further,  o-  is  a  decreasing  function  of  X  .  For 
if  an  increase  in  X  led  to  an  increase  in  a,  then  by  relation  (17)  / 
would  increase  and  by  relation  (16)  a  would  decrease,  which  is  a  con- 
tradiction. To  suppose  that  a  decreases  as  h  increases  leads  likewise 
to  a  contradiction,  so  that  <r  is  an  increasing  function  of  h .  If  o-  is 
everywhere  increased  by  an  additive  constant,  a  also  is  increased  by 
an  additive  constant,  but  the  increase  of  o-  is  less  than  that  of  a  .  If  a 


PARALLEL,  INTERCONNECTED  NEURONS  21 

and  also  h  are  increased  by  the  same  multiplicative  factor  k ,  then  a 

also  is  increased  by  the  same  factor  k .  But  if  <r  alone,  and  not  h ,  is 
so  increased,  the  resulting  a  is  everywhere  less  than  k  times  the  origi- 
nal a- .  All  these  properties  are  intuitively  evident  from  the  character 
of  the  mechanism. 

In  general,  this  type  of  mechanism,  in  which  the  excitation  at  any 
point  is  dependent  upon  the  total  distribution  of  excitation,  is  highly 
suggestive  of  Gestalt-phenomena  (Rashevsky,  1938,  chap,  xxxii).  Ap- 
plications of  mechanisms  involving  parallel  interconnected  neurons 
will  be  made  in  chapter  ix  to  discriminal  sequences  of  stimulus  and 
response,  and  again  in  chapter  xii  to  the  perception  of  color. 


IV 

THE  DYNAMICS  OF  SIMPLE  CIRCUITS 

If  the  terminus  of  a  single  neuron  is  brought  into  coincidence 
with  its  origin,  or  the  final  terminus  of  a  chain  into  coincidence  with 
the  initial  origin,  the  result  is  a  simple  circuit.  Circuits  are  of  com- 
mon occurrence  in  the  central  nervous  system,  and,  in  fact,  Lorente 
de  No  (1933)  asserts  that  for  every  neuron  or  chain  of  neurons  pass- 
ing from  one  given  cell-complex  to  another  given  cell-complex,  there 
is  also  a  neuron  or  a  chain  of  neurons  passing  in  the  reverse  direc- 
tion. O'Leary  (1937)  notes  the  frequent  occurrence  of  circuits  in  the 
olfactory  cortex  of  the  mouse.  A  circuit  composed  of  only  excitatory 
fibers  may  have  the  effect  of  prolonging  a  state  of  activity  after  the 
withdrawal  of  the  stimulus,  of  enhancing  the  activity  due  to  a  pro- 
tracted but  weak  stimulus,  or,  perhaps,  of  providing  a  permanent 
reservoir  of  activity  through  perpetual  self-stimulation.  Thus  Kubie 
(1930)  has  discussed  their  possible  role  in  the  production  of  spon- 
taneous sensations  and  movements.  Prolongation  and  enhancement 
will  not,  of  course,  occur  when  one  member  of  the  circuit  is  inhibi- 
tory, but  besides  the  possible  modulating  effects  that  such  circuits 
might  have,  they  provide,  perhaps  less  obviously,  for  the  possibility 
of  regular  fluctuations  in  the  response  to  a  persistent,  constant  stim- 
ulus. 

Fluctuation,  prolongation  and  enhancement,  permanent  reser- 
voirs of  activity,  are  all  more  or  less  directly  observable  within  the 
central  nervous  system.  Whether  any  or  all  of  these  can  be  attributed 
to  mechanisms  of  precisely  this  type  is  a  question  to  be  decided  by 
the  comparison  of  experiment  with  theory.  We  proceed  therefore  to 
develop  some  of  the  consequences  to  be  expected  if  this  is  indeed  the 
case. 

The  simplest  circuit  is  that  formed  by  a  single  self-stimulating 
neuron  (Landahl  and  Householder,  1939)  of  the  simple  excitatory 
type.  The  total  stimulus  acting  upon  the  neuron  at  any  time  consists 
of  a  part  a  =  e  ,  due  to  the  activity  of  the  neuron  itself,  and  of  a  part 
S  coming  from  other  neurons  or  receptors.  If  we  may  disregard  the 
conduction  time — this  is  always  quite  small— let 

£  =  S-h,  (1) 

and  consider  ^  as  a  function  of  the  excess  of  stimulus  over  threshold, 
equation  ( 1 )  of  chapter  i  takes  the  form 

de/dt  =  a\_<f>(£  +  e)  -  e].  (2) 

22 


THE  DYNAMICS  OF  SIMPLE  CIRCUITS 


23 


If  the  outside  stimulus  is  constant,  the  only  case  we  shall  consider, 
this  differential  equation  can  be  solved  by  a  quadrature, 


t/  £1 


de 


</>(£  +  £>  -e 

to  obtain  t  as  a  function  of  e : 

t  =  T(e) 


=  at , 


(3) 


(4) 


We  must  then  solve  this  equation  for  £  as  a  function  of  t .  However, 
certain  properties  of  this  solution  are  obtainable  directly  from  a  con- 
sideration of  the  form  of  equation  (2). 

We  recall  that  for  £  +  £  positive,  </>  and  its  first  derivative  are 
positive,  with  the  derivative  decreasing  monotonically  to  zero.  For 
£  +  £  negative  $  is  identically  zero.  Suppose  first  that  <£'(0)  =  1 . 
Then,  since  £  is  always  non-negative,  the  equation 

£  =  <£(!  +  £)  (5) 

has  always  a  single  root  s0  which  may  be  zero  (Figure  1).  For  £  >  £0 


the  right  member  of  equation  (2)  is  negative  and  £  is  decreasing;  for 
£  <  £0  the  right  member  of  equation  (2)  is  positive  and  £  is  increas- 
ing. Hence  £  =  £0  represents  a  stable  equilibrium.  Whenever  £  ^  0  , 
£0  ==  0  .  Whenever  I  >  0  ,  then  £0  >  0  ,  and  enhancement  of  a  in  the 
amount  £„  results,  but  after  withdrawal  of  the  stimulus,  when  £  = 
—  h ,  the  neuron  comes  to  rest. 

If  (j>'  (0)  >  1 ,  equation  (5)  has  a  single  root  £  ==  e0  >  0  when  I  >  0  , 
the  single  root  s  =  0  when  I  <  0  and  numerically  large,  two  positive 
roots  besides  the  root  e  =  0  when  £  <  0  and  numerically  small  (Fig- 
ure 2),  and  one  positive  root  besides  the  root  £  =  0  when  £  =  0 .  If 


24    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


£<0 


4>'io)  >1 
Figure  2 


e  is  eliminated  from  equation  (5)  and 

*'U  +  e)=l  , 

and  the  resulting  equation  solved  for  £  =  £<, ,  then  £0  <  0  ,  and  for 
£o  <  I  <  0  we  have  the  case  for  which  equation  (5)  has  two  distinct 
positive  roots. 

Let  £0  represent  the  greatest  of  the  roots  (possibly  zero).  Then 
Co  represents  a  stable  equilibrium  of  equation  (2).  Since  we  can  have 
£0  >  0  even  for  £  <  0  (if  also  I  >  £0)  it  is  possible  for  the  activity  to 
persist  even  after  the  withdrawal  of  the  stimulus  when  £  =  —  h  , 
provided  —h  >  £0 ,  and  provided  the  initial  value  £  =  £i  at  the  time 
of  withdrawal  of  the  stimulus  exceeds  the  smaller,  unstable,  positive 
equilibrium  obtained  from  equation  (5)  when  £  —  —  h  . 

But  whatever  the  value  of  </>'(0),  if  bx  exceeds  the  threshold  h 
at  the  time  the  stimulus  is  withdrawn,  some  activity  will  continue 
for  a  time,  if  not  permanently.  In  order  to  account  for  learning  in 
terms  of  activated  circuits,  the  continuation  must  be  permanent  or 
nearly  so  (cf.  chap.  xi).  Very  likely  a  number  of  circuits  would  be 
involved  in  any  act  of  learning,  in  which  case  forgetting  could  be 
accounted  for  as  a  result  of  the  gradual  damping  out  of  one  after 
another  because  of  extraneous  inhibition.  In  order  to  determine  the 
period  of  the  continuation  where  it  is  not  permanent,  it  is  necessary 
to  know  something  about  how  the  applied  stimulus  S  disappears.  If 
S  suddenly  drops  to  zero,  then  the  time  required  for  the  activity  to 
die  out  is  given  by  equation  (3)  with  £  =  —  h  and  the  upper  limit  £ 
of  the  integration  equal  to  +h  .  But  if  S  is  itself  an  £  from  another 
neuron,  a  new  set  of  equations  must  be  written  down  and  solved. 

If  a  circuit  is  formed  by  a  single  inhibitory  neuron,  the  behavior 
is  described  by  the  equation 


THE  DYNAMICS  OF  SIMPLE  CIRCUITS 


25 


dj/dt  =  b  [y(£-  j)  -  ?]  . 


(6) 


Then  y>  >  0  only  if  |  >  0  ,  but  in  this  case  there  is  always  a  single, 
stable,  equilibrium.  The  result  is  that  the  applied  5  is  decreased  at 
equilibrium  by  a  certain  amount  j.0 .  Also  j(>  increases  as  S  increases, 
although  if  \p  has  a  finite  asymptotic  value,  j0  cannot  exceed  this, 
whatever  the  value  of  &  .  We  may  note,  however,  that  the  presence 
of  additional  circuits  of  this  kind,  with  higher  thresholds,  which  add 
their  effects  with  increasing  S  ,  would  provide  an  effective  damping 
mechanism  over  an  arbitrarily  large  range. 

Consider  next  a  two-neuron  circuit,  with  one  neuron  passing 
from  s2  to  s2 ,  and  the  other  from  s2  to  Sj  .  Suppose,  first,  that  these 
are  both  excitatory,  and,  for  simplicity,  that  they  are  identical  in 
character.  Let  |t  be  the  excess  of  Si  over  the  threshold  of  the  neuron 
originating  at  sx ,  let  £i  represent  the  excitation  produced  here  by  the 
other  neuron,  and  let  |2  and  e2  represent  the  corresponding  quantities 
at  s2 .  Then,  still  neglecting  the  conduction  time,  we  have 


del/dt  =  a  [<j}(£2  +  £2)  ~  £1]  » 
de2/dt  =  a  [<£(!i  +  £1)  —  £2]  . 


(7) 


If  it  happens  that  £1  =  £2  and  that  the  initial  values  of  the  £'s 
are  equal,  then  it  follows  from  symmetry  that  £1  =  £2  for  all  t ,  and 
the  pair  of  equations  (7)  can  be  replaced  by  a  single  equation  of  the 
form  (2).    In  the  general  case  for  any  £t  and  £2 ,  an  equilibrium  is 


Figure  3 


determined  by  an  intersection  of  the  curves  in  the  (ex ,  e2) -plane  de- 
fined by  the  two  equations  [Figure  3] 

£i  =  <M£2  +  £2>»    ei=<j>l£i  +  £1).  (8) 


26    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

For  any  point  to  the  right  of  the  first  curve  sx  is  decreasing,  and  for 
any  point  above  the  second  curve  e2  is  decreasing.  In  case  <£'(0)  ^  1  , 
the  two  curves  have  always  one  and  only  one  intersection ;  at  this 
both  e's  are  positive  only  if  at  least  one  £  >  0  and  the  other  is  not  too 
small,  and  they  define  a  stable  equilibrium.  In  case  ^'(0)  >  1  there 
will  be  one  or  three  intersections.  If  there  are  three,  one  of  these  is 
always  the  origin,  and  this  is  always  a  stable  equilibrium;  if  there 
is  only  one,  this  equilibrium  is  always  stable  and  it  may  be  the  origin. 
In  the  former  case,  the  intersection  farthest  from  the  origin  is  also 
stable.  In  particular,  continuous  activity  following  the  withdrawal 
of  the  outside  stimuli  can  occur  only  in  circuits  for  which  ^'(0)  >  1 , 
and  then  only  in  case  at  least  one  of  the  initial  e's  has  become  suffici- 
ently large.  More  detailed  discussions  of  this  type  of  circuit  in  which 
an  exponential  form  is  assumed  for  the  functions  <j> ,  but  these  are 
not  assumed  to  be  identical,  have  been  given  by  Rashevsky  (1938), 
and  Householder  (1938b).  Rashevsky  (1938)  has  introduced  these 
circuits  in  his  theory  of  conditioning  [cf.  chap.  xi]. 

Circuits  containing  both  excitatory  and  inhibitory  neurons  are 
somewhat  more  interesting,  because  of  the  possibility  of  periodical 
phenomena  (Landahl  and  Householder,  1939;  Sacher,  1942).  The 
scratch-reflex  is  one  of  numerous  examples  of  a  repetitive  or  fluctuat- 
ing response  to  a  stimulus.  Consider  the  case  of  a  single  inhibitory 
and  a  single  excitatory  neuron,  both  of  which  originate  and  termi- 
nate at  the  same  place,  s  .  A  self-stimulating  neuron  of  mixed  type 
may  be  regarded  formally  as  a  special  case.  For  a  mixed  neuron,  it 
is  common  to  assume  (Rashevsky,  1938)  that  the  functions  of  <j>  and  \p 
have  a  constant  ratio  for  all  values  of  their  common  argument,  and  we 
shall  make  this  assumption  for  simplicity.  Then  the  equations  may 
be  written 

de/dt  =  AE(£  +  e  -  j)  -as, 

dj/dt  =  BE(§  +  s-j)-bj,  (9) 

E  —  04/ A  =  by/B  . 

By  dividing  out  a  suitable  factor  from  E  and  incorporating  it  into 
A  and  B  ,  we  may  suppose  without  making  any  restrictions  that 

£7(0)  =1.  (10) 

Now  if  it  should  happen  that  a  =  b  ,  we  could  subtract  the  sec- 
ond of  these  equations  (9)  from  the  first,  replace  e  —  j  everywhere 
by  <r ,  and  have  a  single  equation  of  the  same  form  as  equation  (1). 
Hence  we  suppose  a  ^  b  .  The  pair  of  equations  (9)  in  e  and  j  can 
be  reduced  to  a  single  second-order  equation  in  a  as  follows.  Differen- 
tiate  equations  (9)  once  each,  and  the  equation 
:i-  ■   j  ■  —  j  ■—  a        -  (11) 


THE  DYNAMICS  OF  SIMPLE  CIRCUITS  27 

twice.    There  result  then  seven  equations,  from  which  the  six  quan- 
tities e  and  j  and  their  derivatives  can  be  eliminated : 


(12) 


a"  +  [a  +  b  -  (A  -  B)E(£  +  a)]a' 
-  (bA  -  aB)E(§  +  a)  +  aba  =  0. 
Equilibrium  occurs  for  e0 ,  /0  satisfying 

E(£  +  e  -  j)  =  ae/A  =  bj/B  ,  (13) 

and  hence  for  v0  =  e0  —  j0  satisfying 

(A/a-B/b)E($  +  ,r)  =a.  (14) 

The  action  of  the  circuit  is  somewhat  different  for  the  two  pos- 
sible signs  of  the  coefficient  of  E  .  If  this  is  positive,  this  equation 
has  the  same  form  as  equation  (5).   If 

A/a  -B/b^l, 

there  is  always  a  single  non-negative  root  <r0  of  equation  (14).  If  the 
relation  fails  there  is  at  least  one,  and  there  may  be  three.  Suppose 
cr0  is  a  positive  root,  the  largest  if  there  are  more  than  one.   Let 

(15) 

E  (£  +  a)  =  en  +  eiX  +  e2x2  •  •  • 

where 

e0  =  a0/(A/a  -B/b).  (16) 

Then  equation  (12)  can  be  written 

x"  +  [>+&-  (A  -  B)e1]  x 

(17) 
+  lab  -  (bA  -  aB)el]  x  +  ~-  =  0 

with  terms  of  second  and  higher  degree  in  x  omitted. 

Now  at  the  value  o-0  considered,  the  slope  of  the  left  member  of 
equation  (14)  must  be  less  than  one,  and  this,  in  view  of  the  expan- 
sion (15),  means  that  the  coefficient  of  x  in  equation  (17)  is  positive. 
Hence  the  characteristic  roots  of  the  linearized  equation  (17)  are 
either  complex  or  else  real  and  of  the  same  sign ;  if,  further, 

a  +  b>  (A  -B)elf  (18) 

the  real  parts  are  both  negative ;  and  if,  finally, 

(\M  -  \rB)2e1  <a-b<  (V~3  +  yjB)2  ex ,  (19) 

the  roots  are  complex.  Hence  if  the  relation  (18)  is  satisfied,  the 
equilibrium  an  is  stable,  and  if  relation  (19)  also  holds,  the  approach 
to  the  equilibrium  value  fluctuates  with  a  frequency  v  satisfying 


28   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

-  16ji2v2  =  [a-  b  -  (A  +  B)eiy  -  4  AB  e2 . 

When  the  coefficient  of  E  in  equation  (14)  is  negative,  there  is- 
always  a  single  non-positive  root.  If  this  is  negative  let  <r0  denote  it. 
In  the  transformed  equation  (17)  the  coefficient  of  x  is  always  posi- 
tive, but  the  discussion  is  otherwise  the  same  as  before. 

We  consider  finally  a  circuit  consisting  of  an  inhibitory  neuron 
extending  from  sx  to  s2  and  an  excitatory  neuron  from  s2  to  sx.  The 
equations  are 

de/dt  =  a  [<j> (|2  -  j)  -  e]  , 

dj/dt=bbp(i-i  +  e)  -n.  (20) 

There  is  always  a  single  equilibrium  obtained  by  equating  to  zero 
the  right  members  of  these  equations  (Figure  4) .  Let  e,0 ,  j0  represent 


Figure  4 

the  values  at  equilibrium,  and  let  neither  of  them  vanish.    Then  if 
we  set 

X  =  £  -  £0  ,       y  =  j  -  jo 

and  expand,  equations  (20)  have  the  form 

x'  =  —  a(x  +  ay)  +•••, 
y'  =  b(fix-y)  +..., 


(21) 


where  —a  and  /?  are  the  derivatives  of  4>  and  of  \p  at  /<>  and  at  e0 ,  re- 
spectively. The  characteristic  equation  is 


A2  +  (a  +  b)X  +  ab  (1  +  a  0)  —  0  . 


(22) 


Since  all  parameters  are  positive,  the  real  parts  of  the  characteristic 
roots  are  always  negative  and  the  equilibrium  is  stable.    If,  further 


THE  DYNAMICS  OF  SIMPLE  CIRCUITS  29 

(a-  b)2  <  4  abaft  , 

the  roots  are  complex  and  the  approach  to  equilibrium  is  fluctuating 
with  a  frequency  v  satisfying 

-16ji2r2=  (a-  b)2  -  4  abaft  . 

In  this  circuit  it  is  plain  that  permanent  activity  is  only  possible 
when  |2  >  0 .  Thus  the  simplest  circuits  which  exhibit  fluctuation 
are  those  consisting  of  one  excitatory  and  one  inhibitory  neuron,  and 
a  circuit  so  constituted  can  maintain  permanent  activity  in  the  ab- 
sence of  external  stimulation  only  if  both  neurons  originate  and  termi- 
nate at  the  same  synapse  and  A/a  >  B/b  .  This  is,  of  course,  quite 
evident  intuitively. 


THE  GENERAL  NEURAL  NET 

If  the  response  of  the  organism  can  be  expressed  as  some  func- 
tion of  the  stimulus,  this  function  must  depend  upon  whatever  para- 
meters are  required  for  describing  the  structure  of  the  nervous  sys- 
tem. The  psychologists  can  tell  us  much  about  the  empirical  charac- 
ter of  this  function  but  nothing  about  the  parameters.  The  anat- 
omists and  physiologists  can  tell  us  many  things  about  the  para- 
meters. Our  hope  is  for  a  synthesis  of  the  results  of  both  lines  of 
endeavor. 

If  we  knew  all  about  the  structure,  we  might  hope  to  devise  meth- 
ods for  deducing  the  function.  Actually,  with  complex  structures, 
this  becomes  exceedingly  difficult,  though  we  have  done  this  for  struc- 
tures of  some  very  simple  types.  If  we  knew  all  about  the  function, 
empirically,  we  might  hope  to  deduce  some  of  the  characteristics  of 
the  structure.  However,  there  is  never  a  single,  unique  structure,  but 
many  possible  ones,  all  leading  to  a  function  of  the  same  empirical 
characteristics.  And,  of  course,  we  do  not  know  all  about  either  the 
structure  or  the  function,  but  only  some  things  about  each. 

Certainly  the  structure  of  the  complete  nervous  system  can  be 
no  less  complex  than  the  behavior  which  is  an  expression  of  it,  and 
any  Golgi  preparation  of  a  section  from  the  retina  or  the  cortex  abun- 
dantly exhibits  such  complexity.  We  have  already  mentioned  one  of 
two  general  principles  concerning  this  structure  first  stated  by  Lo- 
rente  de  No  in  a  paper  which  appeared  in  the  Archives  of  Neurology 
and  Psychiatry,  Vol.  30  (1933),  pp.  245  ff.  His  statement  of  these 
is  as  follows: 

"Law  of  Plurality  of  Connections. — If  the  cells  in  the  spinal  or 
cranial  ganglia  are  called  cells  of  the  'first  order'  and  the  following 
ones  in  the  transmission  system  cells  of  the  second,  third  to  •••  nth. 
order,  it  can  be  said  that  each  nucleus  in  the  nervous  system  always 
receives  fibers  of  at  least  n  and  n  +  1  order,  and  often  of  n  ,  n  +  1 
and  n  +  2  order." 

"Law  of  Reciprocity  of  Connections.— If  cell  complex  A  sends 
fibers  to  cell  or  cell  complex  B  ,  B  also  sends  fibers  to  A  ,  either  direct 
or  by  means  of  one  internuncial  neuron." 

In  chapter  ii  we  assumed  the  function  <f>  or  y>  for  each  neuron  to 
be  linear  with  S  between  the  threshold  and  a  certain  maximal  value 
characteristic  of  the  neuron,  and  elsewhere  constant.   The  coefficient 

30 


THE  GENERAL  NEURAL  NET  31 

of  the  linear  variation  we  called  the  activity-parameter.  We  found 
that  if  a  chain  of  n  neurons  leads  from  a  synapse  s,0  to  a  synapse  sn , 
and  if  fixed  stimuli  &,•••,£»  are  applied  at  each  synapse  sx ,  •••  ,  sn , 
then  the  total  excitation  yn  +  <r„  +  Sn  present  at  sn  is  expressible  as 
a  linear  function  of  the  total  y0  present  at  s0  provided  y0  lies  between 
certain  positive  fixed  limits,  and  is  otherwise  constant.  In  special 
cases  the  limits  are  equal  and  yn  is  independent  of  y0 ,  s„.  being  said  to 
be  inaccessible  to  s0 .  When  sn  is  accessible  to  s0 ,  then  the  relation 

y,i  —  Sn  4-  3  +  Ay'  when  y0  <  y' , 

Vn  =  Sn  +  3  +  Ay0  when  y'  ^y0^y" ,  (1) 

yn  =  Sn  +  3  +  43/"  when  */„  >  y"  , 

obtained  in  chapter  ii  for  a  chain,  differs  formally  from  that  for  a 
single  neuron  only  by  the  presence  of  the  term  3  .   But  if  we  set 

Z  =  Sn  +  3,  (2) 

we  have  more  simply 

yn  =  Z  +  Ay0  (3) 

when  y0  lies  between  the  stated  limits,  and  when  it  does  not  the  near- 
est limit  appears  in  this  equation  in  place  of  y,0  •  The  similarity  to 
the  behavior  of  a  single  neuron  is  now  complete,  the  term  Z  cor- 
responding to  the  stimulus  applied  at  the  terminus.  However,  this 
term,  as  well  as  the  limits  y  and  y" ,  depend  here  upon  the  particular 
stimuli  applied  at  the  various  synapses  of  the  chain. 

In  the  discussion  of  more  general  types  of  net,  only  the  occur- 
rence of  circuits  can  present  essential  complications  and  hence  we 
limit  ourselves  to  these,  considering  first  the  case  of  a  simple  circuit 
of  n  neurons.  Such  a  circuit  is  obtained  by  closing  a  chain  of  n  neu- 
rons, bringing  s„  and  s0  into  coincidence.  But  if  sn  is  inaccessible  to 
s0  before  the  closure  then  the  closure  makes  no  change  in  the  value 
of  yn  .   Hence  we  suppose  sn  accessible  to  sc  . 

Following  Pitts  (1942a)  in  substance,  we  find  it  convenient  to 
employ  semi-dynamical  considerations,  taking  into  account  the  con- 
duction-time. Let  us  introduce  as  the  time-unit  the  time  required  for 
a  nervous  impulse  to  traverse  the  circuit  completely.  Having  defined 
in  chapter  ii  the  5  and  4  employed  in  equation  (1),  we  shall  have 
no  further  occasion  to  refer  to  the  parameters  of  the  individual  fibers, 
or  to  the  y  at  any  point  except  s0  =  sn  ,  wherefore  it  is  legitimate  to 
drop  all  subscripts  as  designations  of  neurons  and  synapses.  Fur- 
ther, it  increases  somewhat  the  generality  without  adding  essential 
complications  to  allow  the  stimulus  at  this  point  during  the  interval 
0  =  t  <  1  to  be  different  from  the  constant  value  to  be  assumed  there- 


32    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

after.  Then,  if  y0  is  now  taken  to  represent  the  value  of  y  during  this 
initial  interval,  we  have 

y(t)=y0  (0^t<l), 

y(t  +  l)=Z  +  Ay(t)  (t^O),  (4) 

when  y(t)  lies  between  the  limits  y'  and  y",  and  when  this  is  not  the 
case  the  nearest  limit  replaces  y(t)  in  the  latter  equation. 

Now  it  is  clear  from  the  nature  of  the  mechanism  that  the  fol- 
lowing- possibilities  are  exclusive  and  exhaustive: 

i)  y(t)  approaches  asymptotically  a  value  yx  on  the  interval 
from  y  to  y" . 

ii)  y(t)  reaches  and  remains  constant  at  a  value  in  excess  of 
y"  or  else  below  y'. 

iii)     y(t)  ultimately  alternates  between  two  fixed  values. 

If  we  set 

t  =  v  +  r  (0^t<1),  (5) 

where  v  is  an  integer,  then  when  A  ¥=  1 ,  the  solution  of  the  difference- 
equation  (4)  has  the  form 

1-  Av 

y(v  +  T) Z  +  y0Av  =  y„  +  {y«-y„)Av  (6) 

1-  A 

where 

y„  =  Z/(l-A),  (7) 

until  y  falls  outside  the  interval  from  y  to  y".  Hence  case  (i)  occurs 
provided  |A|  <  1  and 

v''s§  v.  =£  y".  (8) 

If  A  >  1 ,  the  interval  between  y(t)  and  y:X  ,  wherever  the  latter  may 
be,  continues  to  increase  while  y(t)  lies  on  the  interval,  and  after 
having  passed  either  limit — which  will  occur  in  a  finite  time — it  re- 
mains constant.  If  A  <  —  1  ,  the  interval  y  —  y^  increases  numeri- 
cally but  with  alternating  sign  until,  after  a  finite  time,  one  limit  or 
the  other  is  passed.  Thereafter,  if  equation  (8)  is  satisfied  without 
the  equality,  (iii)  occurs,  while  if  equation  (8)  fails  or  if  an  equality 
holds,  then  (ii)  occurs.  When  A  =  —  1  ,  alternation  between  fixed 
values  starts  immediately  if  relation  (8)  is  satisfied  without  the  equal- 
ity, and  otherwise  (ii)  occurs.  Finally  if  A  =  1  ,  it  is  evident  from 
the  difference-equation  (4)  itself  that  the  y(y)  form  an  arithmetic 
progression  until  one  limit  is  passed,  the  later  terms  being  identical. 
Additional  essential  complications  are  involved  in  the  discussion 
of  nets  consisting  of  two  or  more  circuits.  However,  certain  simplifi- 
cations can  be  performed  at  once.    We  wish  to  determine  y  at  each 


THE  GENERAL  NEURAL  NET  33 

synapse.  But  if  any  synapse  is  the  origin  of  only  one  and  the  termi- 
nus of  only  one  neuron,  the  two  neurons  constitute  a  chain,  and  after 
the  E  is  determined  for  this  chain,  this  synapse  requires  no  further 
consideration.  Again,  let  a  neuron  N  form  a  synapse  with  two  or 
more  neurons  Nx ,  N2  ,  •  •  •  .  The  results  are  the  same  if  we  suppose  N 
replaced  by  two  or  more  neurons  N7,  N2',  •  ••  ,  with  identical  prop- 
erties all  originating  at  the  origin  of  N ,  but  N*  forming  a  synapse 
with  Nx  alone,  2VY  with  iV2  alone,  •••  (Pitts,  1942b).  Thus  the  only 
synapses  requiring  separate  consideration  are  those  at  which  two 
or  more  neurons  terminate.  Each  of  the  synapses  of  the  set  under 
consideration  is  the  terminus  of  two  or  more  chains  which  originate 
at  other  synapses  of  the  set,  and,  the  distribution  of  stimulation  be- 
ing fixed,  each  chain  is  characterized  by  the  values  of  its  set  of  four 
parameters.  In  case  the  terminus  of  any  of  these  chains  is  inacces- 
sible from  its  origin,  the  a  which  it  produces  is  calculable  indepen- 
dently of  the  value  of  y  at  its  origin,  and  we  may  delete  this  chain 
and  add  this  a  to  the  5  at  the  terminus.  If  there  happens  to  be  only 
one  other  chain  terminating  there,  this  can  be  combined  with  the 
chain  or  the  chains  originating  there  and  the  synapse  dropped  out 
of  the  set  being  considered.  We  therefore  suppose  that  each  synapse 
is  accessible  to  the  origin  of  every  chain  which  terminates  there. 
There  is  also  the  possibility  that  when  the  S  applied  at  any  synapse 
is  increased  by  the  maximum  a  that  can  be  produced  together  by  all 
the  chains  which  terminate  there,  this  is  still  below  the  y',  or  that 
the  S  increased  by  the  minimal  a  of  all  together  is  above  the  y"  of 
some  chain  which  originates  there.  If  so,  the  a  produced  by  this 
chain  can  be  calculated  at  once,  the  result  added  to  the  S  applied 
there,  and  the  chain  deleted.  It  is  clear,  of  course,  that  these  dele- 
tions, which  are  made  possible  by  the  inaccessibility  of  one  synapse 
to  another,  will  be  different  for  different  distributions  of  stimulation. 

Having  performed  these  simplifications,  we  suppose,  before  pass- 
ing on  to  the  most  general  case,  that  only  one  synapse  remains.  The 
resulting  net,  which  consists  of  a  number  of  circuits  all  joined  at  a 
single  common  synapse,  we  shall  call  a  rosette  (Pitts,  1943a),  and 
the  common  synapse,  we  shall  call  its  center.  Now  if  the  conduction- 
time  is  not  the  same  around  all  these  circuits,  we  may  nevertheless, 
with  sufficient  accuracy,  regard  these  times  as  commensurable,  and 
we  shall  use  their  common  measure  as  the  time-unit.  Let  n  be  the 
number  of  circuits,  and  let  //,  be  the  conduction-time  of  the  i-th  circuit. 

Now  consider  the  contributions  of  the  i-th  chain  to  the  stimulus 
y  at  s  at  any  time.  If  y(t)  is  the  total  stimulus  at  time  t ,  then  the 
contribution  at  time  t  +  m  is 

Si  +  Aiy(t)  when  y{  ^y(t)  ^yj', 


34   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

the  y(t)  in  that  expression  being  otherwise  replaced  by  the  nearest 
limit.   Let  us  introduce  quantities  at(t),  fti(t)  denned  as  follows: 

0Li(t)=l     when     jf(f)=Vo 
uj(O=0    when    y(t)<yi, 

(10) 
Pi(t)=l     when     y(0  =2/i", 

A(0  =0     when     y(t)  >yi". 

Then  the  contribution  of  the  i-th  chain  to  y(t  +  //;)  may  be  written 

Si  +  A^imWfiiit)  y(t)  +  tl-aiW]  y{  +  [1 -&(*)]  1/i" }. 

If  we  introduce  the  operator  E  defined  by 

Ey(t)  =  y(t  +  l), 

let  fi  represent  the  largest  of  the  jut ,  and  set 

P§  =  (i  —  fij , 
then  we  have  finally 

E*y{t)=Z  +  ^AiE»  {*dt)  pi(t)y(t)  +  [1  -  a,(t)]  yi 

(11) 

+  ii-fawiyn. 

The  functions  a  and  /?  are  constant  except  when  y  crosses  one  of  the 
boundaries  y'  or  y"  associated  with  the  corresponding  chain.  Hence 
the  difference-equation  (11)  can  be  solved  on  the  assumption  that 
the  a's  and  /3's  are  constant,  and  the  solution  is  valid  as  long  as  it  lies 
on  the  particular  interval  associated  with  the  assumed  values  of  the 
a's  and  /3's. 

If  the  numbers  y,'  and  y"  are  arranged  in  order,  they  limit  at 
most  2n  +  1  intervals  (two  of  them  infinite),  and  each  interval  is 
associated  uniquely  with  a  particular  set  of  values  a,  ,  fit .  No  other 
set  of  the  a;  ,  /?/  is  possible.  Associated  with  each  of  these  sets  a* ,  /J» , 
is  a  unique  value  of  y  satisfying 

[l-2A»ai0i]  y-^  +  24i  [(1-002//+  (l-fii)yn   (12) 

provided  the  coefficient  of  y  is  non-null.  This  defines  a  possible  equi- 
librium of  the  difference-equation.  However,  if  this  value  of  y  does 
not  lie  on  the  associated  interval,  then  no  equilibrium  for  the  gen- 
eral equation  (11)  exists  on  this  interval.  If,  for  the  set  a,  ,  /3;  ,  the 
solution  y  of  equation  (12)  does  lie  on  the  associated  interval,  the 
solution  y(v  +  t)  of  the  difference-equation  (11)  corresponding  to 
the  a;  ,  /?,  equal  to  these  constant  values  differs  from  this  constant 
value  y  by  a  sum  of  terms  of  the  form  p(v)Av,  where  p(v)  is  a  poly- 
nomial in  v  multiplied,  possibly,  by  a  sine  or  a  cosine,  and  a  is  a  real 


THE  GENERAL  NEURAL  NET  35 

root  or  the  modulus  of  a  complex  root  of  the  equation 

a*  ~  2  Ai  ai  fr  xp*  =  0  .  (13) 

As  before,  v  is  an  integer  for  which 

t  =  V    +    T  (O^T^l). 

Hence  the  equilibrium  is  unstable  unless  every  root  of  equation  (13) 
has  a  modulus  less  than  unity. 

In  case  for  any  of  the  intervals  the  coefficient  of  y  in  equation 

(12)  vanishes,  this  equation  has  no  solution  unless  the  right  member 
also  vanishes.  But  then  equation  (13)  has  a  root  unity  and  the  corre- 
sponding solution  of  (11)  involves  a  simple  polynomial  of  non-null 
degree,  so  that  no  stable  equilibrium  occurs.  Thus,  in  brief,  in  order 
for  any  interval  to  possess  a  stable  equilibrium,  it  is  necessary  and 
sufficient  that  the  solution  y  of  equation  (12)  obtained  from  the 
associated  set  a,  ,  pt ,  shall  lie  on  this  interval,  and  that  the  equation 

(13)  shall  have  every  root  of  modulus  less  than  unity.  Fluctuating 
equilibria  of  the  sort  met  with  in  the  simple  circuit  are  here  possible, 
and  also  another  sort  arising  from  possible  complex  roots  of  the  char- 
acteristic equation  (13)  and  leading  to  terms  involving  sines  and 
cosines. 

In  the  general  case,  let  the  synapses  Si  and  the  chains  CK  be  sep- 
arately enumerated,  and  let  us  define  two  sets  of  quantities  PjK  and 
QjK  as  follows : 

PjK  =  1  if  Sj  is  the  origin  of  CK  , 

Pjk  =  0  if  Sj  is  not  the  origin  of  CK , 

QiK  =  1  if  Si  is  the  terminus  of  CK , 

QiK  =  0  if  s»  is  not  the  terminus  of  CK . 

All  simplifications  as  described  above  having  been  made,  we  can  de- 
fine for  each  synapse  s<  a  quantity 

Zi  =  Si  +  ^QiKEKt 

and  for  each  chain  CK  the  sets  of  quantities  aK ,  pK  with  values  0  or  1 
according  to  the  value  of  y ,  at  the  origin  CK ,  relative  to  yK'  and  yK". 
Then  the  difference  equations  satisfied  by  the  yt  have  the  form  (Pitts, 
1943a) 

E*  yi  =  Zi  +  2  Q^  AKE»«{(1-  aK)yK' 

(14) 

+    (1  -Pk)v"  t- Ok  &  2^*2//}. 

If  there  are  n,  chains  originating  at  Si ,  the  a's  and  p's  associated  with 
these  chains  are  able  to  take  on  at  most  2/1*  —  1  different  sets  of  val- 
ues, and  there  are  therefore  at  most  U  (2nt  —  1)  sets  of  values  for 


36   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

them  altogether.  Each  set  of  values  is  associated  with  a  region  in 
y-spa.ce  which  may  contain  a  single  point  whose  coordinates  yi  rep- 
resent a  steady-state  of  the  net.  For  this  to  be  so  (the  equilibrium 
being  stable)  two  conditions  must  be  fulfilled:  The  constant  y-x  de- 
fined by  equations  (14)  when  the  aK  and  /5K  are  given  these  values  and 
the  operator  E  is  taken  to  be  the  unit  operator  must  define  a  point 
which  lies  in  this  region ;  and  a  certain  algebraic  equation  (the  char- 
acteristic equation  of  these  difference-equations)  must  have  only  roots 
whose  moduli  are  less  than  one.  In  principle,  therefore,  the  steady- 
state  activity  of  nets  of  any  degree  of  complexity  can  be  determined, 
though  admittedly  the  procedure  could  become  exceedingly  laborious. 
Thus  given  three  synapses,  joined  each  to  each  by  a  total  of  six  chains, 
27  regions  in  ?/-space  may  exist  and  require  separate  consideration 
as  possible  locations  of  equilibria.  Moreover,  persistent  fluctuations 
may  arise,  no  steady-state  being  approached  at  any  time. 

While  the  solution  of  the  direct  problem  of  describing  the  output 
of  any  given  net  is  complete,  at  least  in  principle,  the  general  inverse 
problem  is  still  open.  However,  in  the  special  case  where  the  output 
function  is  such  that  the  a's  and  /5's  remain  constant,  Pitts  (1943a) 
has  shown  how  to  construct  a  rosette  to  realize  this  function. 

This  concludes  our  purely  formal  discussion  of  neural  structures, 
and  we  turn  now  to  some  special  structures  and  their  possible  rela- 
tion to  concrete  types  of  response. 


PART  TWO 


VI 

THE  DYNAMICS  OF  THE  SINGLE  SYNAPSE: 
TWO  NEURONS 

Thus  far  we  have  been  concerned  with  the  formal  development 
of  methods  for  determining  the  activity  of  structures  composed  of 
neurons.  We  shall  now  attempt  to  make  application  of  the  theory 
and  method  to  experimental  problems.  Two  paths  are  open  to  us.  We 
could,  on  the  one  hand,  examine  specific  neural  structures,  seeking' 
to  determine  for  each  the  response  which  it  mediates  as  a  function  of 
the  stimulus,  or  we  might  start  with  this  function  and  attempt  to 
construct  a  suitable  mechanism.  In  this  and  in  succeeding  chapters 
we  follow  the  first  course.  In  the  final  chapters  of  this  Part  II  we 
follow  the  second  course.  The  immediate  problem  is  considered  solved 
if,  from  the  theoretical  structure,  quantitative  relations  are  derived 
which  agree  with  the  experimental  data  within  a  suitable  margin  of 
error  for  some  range  of  the  variables  in  question,  and  if  the  number 
of  parameters  is  not  too  large.  Now  many  of  the  parameters  may  be 
explicit  functions  of  certain  other  variables  which  have  been  kept 
constant  throughout  the  experiment.  Thus  in  many  cases,  the  struc- 
ture will  have  different  properties  when  the  constants  of  the  experi- 
mental situation  are  changed.  In  these  cases  we  may  say  that  the 
structure  studied  makes  predictions  regarding  activity  outside  the 
domain  of  activity  intended  to  be  covered.  Such  predictions  may  sug- 
gest an  experimental  approach  not  otherwise  evident.  If  the  predic- 
tions are  borne  out,  the  theory  is  immediately  extended  in  its  applica- 
tion. If  not,  the  structure  must  be  extended  or  revised  in  such  a  man- 
ner as  to  include  the  old  as  well  as  the  new  properties. 

Thus  on  whichever  course  we  set  out,  whether  working  from 
mechanism  to  behavior,  or  from  behavior  to  mechanism,  we  are  led 
finally  into  the  second  course  when  extensions  are  required.  In  this 
we  are  guided  by  a  consideration  of  the  elements  without  which  there 
could  be  no  correspondence  between  the  activity  of  the  structure  and 
the  activity  observed.  If  certain  of  the  observed  elements  interact, 
then  the  elements  of  the  structure  must  be  inter-connected.  If  the 
action  is  unilateral  in  the  experimental  situation,  a  unilateral  con- 
nection may  suffice.  If  the  observed  activity  depends  on  the  order  of 
the  events  of  the  past,  the  structure  must  contain  elements  which  ex- 
hibit this  property  of  hysteresis.  Thus  one  is  limited  to  a  consider- 
able extent  in  the  choice  of  mechanisms  to  be  studied. 

37 


38   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

For  first  application  we  choose  the  simplest  structures,  working 
gradually  to  those  of  increasing  complexity.  We  shall  find  in  the 
present  chapter  how  a  very  simple  mechanism  will  serve  for  the  inter- 
pretation of  such  superficially  different  sorts  of  data  as  those  con- 
cerning the  occurrence  and  duration  of  a  gross  response,  just-dis- 
criminable  intensity-differences,  adaptation-times,  and  fusion-frequen- 
cies in  vision  and  perhaps  other  modalities.  In  general,  even  where 
the  structure  is  relatively  simple,  it  is  not  possible  to  solve  in  closed 
form  the  equations  resulting  from  this  structure.  Thus  certain  re- 
strictions upon  the  parameters  may  have  to  be  introduced  in  order 
to  obtain  a  workable,  even  if  approximate,  solution.  As  the  choice  of 
the  restrictions  is  somewhat  arbitrary,  one  should  keep  in  mind  that 
other  equally  plausible  restrictions  could  lead  to  different  results  and 
increase  both  the  accuracy  and  the  scope  of  the  theory. 

The  simplest  structure  which  can  be  studied  is  a  single  neuron, 
and  the  simplest  assumption  that  can  be  made  about  its  activity  is 
that  it  is  of  the  simple  excitatory  type.  Its  activity  is  determined 
when  we  have  evaluated  e(t)  for  any  5  .  However,  one  does  not  ob- 
serve e  but  some  response  R  .  Thus  the  simplest  structure  in  which 
we  can  deal  with  observed  quantities  is  a  chain  of  two  neurons,  the 
first  being  acted  upon  by  some  stimulus  S  ,  and  the  second,  which  may 
be  a  muscular  element,  capable  of  producing  some  response  R  .  The 
response  R  is  produced  as  soon  as  e  reaches  the  threshold  of  the  sec- 
ond neuron.  Hence  if  we  set  h  =  s(t)  and  solve  for  I ,  then  since  the 
function  s(t)  depends  upon  S  through  the  function  $(S)  (chap,  i, 
equation  1),  we  obtain  the  reaction-time  t1(S)  as  a  function  of  the 
intensity  S ,  this  time  being  measured  from  the  application  of  the 
stimulus  until  e  reaches  the  threshold.  For  this  purpose  we  use  <j>  as 
given  by  equation  (6)  of  chapter  i  and  assume 

1                         h 
tl  =  --\0g  (1 ),  (1) 

a  ft  log  S/h, 

where  hx  is  the  threshold  of  the  afferent  neuron. 

This  relationship  should  apply  to  an  experiment  on  a  simple  re- 
flex in  which  a  stimulus  of  intensity  S  produces  a  response  R  after 
a  time  tr .  However,  the  total  time  tr  from  the  application  of  the 
stimulus  until  occurrence  of  the  response  as  registered  by  the  timing- 
device  involves,  in  addition  to  t1(S),  also  a  time  t0  which  measures 
the  time  for  conduction  plus  the  time  required  for  the  muscular  re- 
sponse to  effect  the  recording  instrument  plus  any  other  time  of  delay 
which  does  not  depend  appreciably  upon  S  .  We  may  then  expect  the 
equation 


SINGLE  SYNAPSE :  TWO  NEURONS 


39 


tr  =  to lOg   (1 

a 


h 


-) 


piogS/K 


(2) 


to  represent  the  experimentally  determined  relation  between  tr  and 
S  .  The  extent  to  which  this  does  so  in  some  cases  for  which  data  are 
available  may  be  seen  in  Figures  1,  2,  and  3  where  experimental  data 


BERGEN  ANO  CATTELL 
VISUAL  DATA 


O SUBJECT  B 
•  SUBJECT  C 


STIMULUS    INTENSITY   S- 


Figure  1. — Comparison  of  theory  with  experiment:  dependence  of  delay  of 
reaction  upon  intensity  of  the  stimulus  for  visual  stimuli.  Curves,  theoretical 
predictions  by  equation  (2)  ;  points,  experimental.  (Visual  data  from  Cattell, 
1886.)  Abscissa,  intensity  (on  logarithmic  scale)  of  stimulus;  ordinate,  interval 
between  presentation  of  stimulus  and  occurrence  of  response. 

(points)  and  theoretical  predictions  (curves)  are  shown  for  each  of 
a  number  of  rather  different  types  of  stimuli.  The  details  are  given 
in  the  legends. 

In  general,  we  cannot  expect  that  the  chain  from  receptor  to 
effector  involved  in  the  reflex  will  contain  so  small  a  number  of  ele- 
ments. But  certainly  this  demands  first  consideration  since  it  is  the 
simplest  possible  mechanism.  And  even  if  the  chain  were  known  to 
contain  a  larger  number  of  neurons,  the  slowest  synapse  in  the  series 
will  tend  to  govern  by  itself  the  temporal  form  of  the  response,  so 
that  if  the  remaining  synapses  are  relatively  fast,  the  equation  just 
deduced  will  still  provide  an  adequate  description  of  the  experimental 
situation. 

If  a  stimulus  S  is  presented  for  too  short  a  time  t ,  e  will  not 


40   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

have  reached  the  threshold  h  at  the  end  of  this  time.  But  for  a  given 
S  there  is  a  minimal  t  =  f  at  which  time  e  =  h .  This  i  (S)  is  the 
minimal  period  of  stimulation  with  the  intensity  S  which  just  suffices 
to  produce  the  response.   We  obtain  for  this  an  equation  similar  to 


STIMULUS    INTCNSITY  S  - 


Figure  2. — Comparison  of  theory  with  experiment:  dependence  of  delay  of 
reaction  upon  intensity  of  the  stimulus  for  auditory  stimuli.  Curves,  theoretical 
predictions  by  equation  (2)  ;  points,  experimental.  (Auditory  data  from  Pieron, 
1920).  Abscissa,  intensity  (on  logarithmic  scale)  of  stimulus;  ordinate,  interval 
between  presentation  of  stimulus  and  occurrence  of  response. 


STIMULUS   INTENSITY    S  - 


Figure  3. — Comparison  of  theory  with  experiment:  dependence  of  delay  of 
reaction  upon  intensity  of  the  stimulus  for  gustatory  stimuli.  Curves,  theoretical 
predictions  by  equation  (2) ;  points,  experimental.  Gustatory  data  from  Pieron, 
1920.)  Abscissa,  intensity  of  stimulus;  ordinate,  interval  between  presentation 
of  stimulus  and  occurrence  of  response. 


SINGLE  SYNAPSE :  TWO  NEURONS  41 

equation  (2) ,  but  with  t0  =  0  and  tr  replaced  by  t .  Thus  from  a  con- 
sideration of  a  chain  of  two  neurons  one  should  expect  that  if  all 
other  conditions  remain  unchanged  the  same  relationship  should  hold 
in  both  cases,  except  that  t0  would  be  absent  in  this  case.  Since  the 
two  cases  are  experimentally  distinct,  it  may  be  that  the  results  from 
the  two  types  of  experiments  are  widely  divergent.  If  so,  it  may  be 
necessary  to  assume  that  there  are  several  neurons  in  the  chain  or 
even  circuits  in  the  structure.  In  any  case,  the  kind  of  disagreement 
may  suggest  the  nature  of  the  change  to  be  made  in  the  neural  net. 

Let  us  consider  another  special  case  of  a  chain  of  two  neurons. 
Let  the  afferent  neuron  be  of  the  mixed  type  with  <£  =  y> ,  a  >  b  ,  and 
threshold  hx .  A  constant  S  >  hx  applied  to  such  a  neuron  results  in 
a  a  (=  e  —  j)  which  is  positive,  but  which  vanishes  asymptotically. 
Then,  the  stimulus  being  presented  at  t  =  0  when  e  =  j  =  0  ,  if  S  is 
large  enough,  and  h  not  too  large,  a  will  first  reach  the  value  h  at 
some  time  tf .  From  this  one  can  determine  a  relation  tr(S)  similar 
to  that  of  equation  (1).  If  S  is  maintained  at  a  constant  value  for  a 
sufficient  time,  <r  reaches  a  maximum  and  declines.  Let  tr  +  r  be  the 
time  at  which  a  returns  to  the  value  h .  Then  we  can  determine  the 
duration  T  of  the  reaction  as  a  function  of  the  intensity  S  when 
e  =  j  =  0  initially. 

At  t  =  t*  >  t  +  t  let  S  be  replaced  by  S  +  AS,  where  A  S  may 
take  on  any  positive  value.  Negative  values  of  A  S  would  be  of  in- 
terest only  if  tr  <  t*  <  tr  +  r .  Then  for  t  >  t*  >  t r  +  T  we  may  write 

+  4>(S  +  A  S)  [e-&<<-'*>  -  «?-«(«-«•>]  . 

If  A  S  is  large  enough,  a  will  again  reach  h  at  some  time  t  —  t*  +  tr' 
and  the  response  will  again  be  initiated.  By  setting  a  =  h  in  equa- 
tion (3),  we  obtain  tr'(AS,  S,  t*)  from  the  smaller  root,  t.  For 
t*  >  >  1/6  and  tr'  <  <  1/6  ,  we  may  obtain  an  equation  for  tr'  which 
shows  the  time  V  to  depend  only  upon  A  S/S ,  and  not  upon  hx . 

At  some  time  t  =  t*  +  tr'  +  t  the  response  will  again  cease.  Using 
the  larger  root  of  a  =  h  in  expression  (3)  we  may  determine  the 
duration  T'(AS  ,S  ,t*)  of  the  response.  For  t*  >  >  1/6  and  r  >  >  1/a, 
we  may  write 

T'  =  -log[/log(l  +  zlSyS)]   -tf',  (4) 

6 

/  being  a  constant.  For  fairly  large  values  of  A  S/S  we  may  neglect 
tr'  in  equation  (4).  The  equation  thus  makes  a  definite  testable  pre- 
diction as  to  the  nature  of  the  relation  between  the  duration  of  the 
response  and  the  relative  increase  of  the  stimulus. 


42   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

If  we  let  t*  be  large  and  if,  for  a  fixed  S  ,  we  restrict  A  S  to  the 
least  value  it  can  have  while  U  remains  finite  (i.e.,  the  two  roots  tr' 
and  tr  +  t  of  equation  (3)  coincide),  we  obtain,  if  we  use  equation 
(6)  of  chapter  i, 

AS/S  =  d  =  d0,  (5) 

da  being  a  constant.  That  is,  when  a  constant  stimulus  of  intensity 
S  >  hx  has  been  applied  for  a  long  time,  the  smallest  additional  stim- 
ulus A  S  necessary  to  produce  the  response  must  be  a  constant  frac- 
tion of  the  intensity  of  the  original  stimulus  S .  On  the  other  hand, 
US  <  hi  we  have 

6=  (S0  +  l)hJS-  1.  (6) 

Thus  d(S)  decreases  hyperbolically  from  oo  to  S0  as  S  varies  from 
zero  to  hx ;  thereafter  S  is  a  constant.  The  quantity  <3  of  equations  (5) 
and  (6)  is  essentially  a  Weber  ratio,  and  its  variation  with  5  as  de- 
scribed in  the  above  equations  has  the  chief  qualitative  properties  of 
the  experimental  relation  for  most  types  of  stimuli.  This  problem 
will  be  discussed  in  more  detail  subsequently  (chap.  ix). 

Suppose  that  instead  of  replacing  S  by  S  +  A  S  at  time  T,  we 
remove  S  for  a  time  t'  after  which  only  A  S  is  presented.  This  is  the 
experimental  technique  for  studying  the  processes  of  adaptation  and 
recovery.  Then  for  t  >  t*  >  t' , 

CT=z<£(S)  [e-°<*-'*>  -  r*<*-**}  +  e-bt  -  e-at] 

(7) 

At  the  time  t*,  we  shall  suppose  a  <  h  .  Hence  at  some  time  t  =  t*  + 
t'  +  t" r ,  if  A  S  is  large  enough,  o-  =  h  ,  and  the  response  occurs.  From 
this  relation,  together  with  equation  (7),  we  can  determine  the  reac- 
tion-time t"r(AS ,  S ,  t*,  V)  from  the  smaller  root,  and  the  duration 
t"  (AS ,  S ,  t\  t')  of  the  response  from  the  larger  root,  at  any  stage 
of  the  process  of  recovery  following  preadaptation  to  the  intensity  S  . 
We  can  further  determine  the  minimal  AS  required  for  stimulation 
as  well  as  the  minimal  interval  of  exposure  at  a  given  AS  . 

'    If  t*  >  >  1/b  ,  t'  >  >  1/a ,  t'  >  >  tr"  and  tr"  <  <  1/b  ,  we  have 

cl>(AS)  (1  -  e-"*r")  -  ^(S)  e-w  -  h  =  0.  (8) 

Thus 

tr"  =  -  -log(  1  -  *±-L——     ,  (9) 

so  that  the  reaction  time  tr"  increases  with  S  but  decreases  with  both 
A  S  and  t' . 


SINGLE  SYNAPSE :  TWO  NEURONS  43 

Equation  (9)  makes  a  definite  prediction  as  to  the  relationship 
between  the  reaction  time  and  the  variables  S  ,  A  S  ,  and  t'  and  is  thus 
subject  to  experimental  test. 

If,  for  fixed  S  and  t',  we  now  restrict  A  S  to  the  smallest  value 

for  which  the  response  can  still  take  place,  we  obtain  a  relation  of  the 

form 

AS  S 

log =  e-bt'log  —  ,  (10) 

h'  K 

where  log  K  =  log  hx  +  h/(i .  Thus,  except  for  minimal  if,  the  loga- 
rithm of  the  testing  stimulus  A  S  is  an  exponentially  decreasing  func- 
tion of  the  time  t'  of  recovery.  As  the  time  t'  becomes  infinite,  A  S 
approaches  h'.  The  intensity  S  determines  by  how  much  the  ordinate 
is  multiplied  in  the  graph  of  A  S  against  P.  The  type  of  relationship 
between  A  S  and  t'  of  equation  (10)  for  the  case  of  visual  stimuli  is 
found  in  the  work  of  various  investigators  (cf.  S.  Hecht,  1920). 

Suppose  next,  still  assuming  that  <j>  =  y> ,  that  any  constant  stim- 
ulus has  been  applied  for  a  long  time  and  that  at  t  =  0  the  stimulus 
is  increased  at  a  rate  such  that  cUf>/dt  =  X  .  After  a  time  t ,  we  find 
that 

X  X 

<r  =  —  (1-  e~bt) (1-  e~at).  (11) 

b  a 

If  X  <  abh/ (a  —  b) ,  where  h  is  the  threshold  of  the  second  neuron,  a 
will  never  exceed  h  and  there  can  be  no  response.  This  is  analogous 
to  the  failure  of  slowly  rising  currents  to  produce  excitation  in  peri- 
pheral nerves  and  corresponds  to  the  effect,  commonly  experienced, 
that  a  stimulus  which  rises  slowly  in  intensity  often  fails  to  evoke  a 
response.  If  X  is  larger,  and  response  occurs,  then  from  equation  (11) 
we  can  determine  the  reaction-time  tr'"  by  solving  for  t  with  a  =  h . 
It  is  clear  that  the  reaction-time  depends  very  much  upon  the  manner 
in  which  the  stimulus  is  presented. 

Let  us  consider  the  effect  of  a  different  mode  of  application  of  a 
stimulus  to  an  afferent  neuron  of  the  mixed  type  with  <f>  =  t/'  •  Let  a 
stimulus  S  be  given  for  a  period  of  time  rT  followed  by  no  stimulus 
for  a  period  of  time  (l—r)T.  Let  this  be  repeated  indefinitely.  For 
each  successive  interval  we  can  determine  e  and  j  from  the  differ- 
ential equation  together  with  the  requirement  that  £  and  j  both 
be  continuous.  The  value  of  a(t)  during  the  interval  0  <  t  <  rT 
of  stimulation  after  a  large  number  of  repetitions  may  be  obtained 
as  follows.  Let  e„-t  represent  the  value  of  e  at  the  beginning  of  the 
n-th  resting  period,  and  e'n-i  the  value  at  the  beginning  of  the  n-th 
period  of  stimulation,  where  e0  —  0 .   We  use  equation  (8)  of  chap- 


44    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


ter  i,  replacing  c0  by  c'M_i  and  t  by  rT  to  calculate  e»_i ,  and  we  use  the 
same  equation  to  calculate  e'„  by  setting  <j>  =  0  ,  replacing  s0  by  £„_i , 
and  £  by  (1— r)T  .  When  we  do  this  we  find  by  simple  induction  that 

e'B  =  0(l  -  e-^)  (1  -  e-naT)e-a^T/(l  -  e-"T) , 

and  as  n  becomes  large  the  exponential  containing  it  can  be  neglected. 
The  expression  for  j'n ,  similarly  defined,  is  the  same  with  b  replacing 
a.  To  obtain  j(t)  during  the  interval  in  question,  we  need  only  re- 
place j0  by  j'n  and  b  by  a  in  the  same  equation  (8) .  After  taking  the 
difference  e  —  j  and  performing  elementary  algebraical  simplifica- 
tions, we  obtain  finally  the  desired  expression: 


—  A  p-bt 


a  =  <},e 


/  1  -  e-hil-r)T  \  /  1  -  e-a^-r)r  \ 


Now  a  reaches  a  maximum  at  t  =  t*  given  by 


t*  = 


log 


a 


a(l-  c*<1-r>f)  (1  -  e-bTl 
6(1  -  e-^-'^Ml  -  <raT) 


(13) 


unless  t*  >  rT ,  in  which  case  the  maximum  value  of  a  is  a(rT).  Set 
a  =  h  in  equation  (12)  with  t  replaced  by  rT  or  t*  according  to  the 
case.  Then  for  given  r  and  T ,  that  value  of  5  which  satisfies  the 
equation  is  the  least  stimulus  that  will  produce  a  steady  response 
when  repeated  in  this  manner. 

For  this  r  and  S ,  let  T*  be  the  particular  value  of  T  employed. 
Then  f  =  1/T*  is  a  critical  frequency  separating  response  from  no 
response  for  the  value  of  5  in  question.  Then,  if  <j>  =  ft  log  S/hi  and 
H  =  h/0,  for  rT*  <  t*  or  f  >  r/t*, 


I    6-°' 
V   ~~1 


brT*  _  g-br* 


-bT* 


irT*   (>-aT*     \  S 

log-  =  H, 

1  -  e-aT*      I        K 


(14) 


and  for  rT*  >  t*  or  /*  <  r/t*, 

b 

1  --  e~aT* 
b 


-a(l-r)T* 


-.  -a- 


1  -  <rw* 


a 


9-b(l-r)r* 


H 


L   (a  -  b)  log  (S/h,)  J 


a-b 


(15) 


For  very  large  frequencies  /*,  we  have  from  expression   (14),  for  r 
not  too  near  zero  or  unity, 


a-b  ,      S 

/*  = r(l  —  r)  log—. 

2H  ha 


(16) 


This  states  that  the  frequency  above  which  response  fails  to  occur 
is  proportional  to  log  S/Jh  as  well  as  to  r(l-r),  the  function  of  r 


SINGLE  SYNAPSE:  TWO  NEURONS 


45 


being  maximal  at  r 


and  symmetric  about  r  =  \ .    In  terms  of 


visual  stimulation  /'*  is  the  frequency  above  which  the  typical  response 
to  intermittent  stimulation  ceases,  and  thus  /*  may  be  identified  with 
the  critical  flicker-frequency.  Equation  (16)  then  states  that  the  criti- 
cal flicker-frequency  increases  with  the  logarithm  of  the  intensity  for 
large  /*.  This  is  essentially  the  Ferry-Porter  law.  Furthermore,  with- 
in a  limited  range  and  for  a  fixed  stimulus-intensity,  this  frequency 
is  the  same  for  a  given  value  of  the  light-dark  ratio,  LDR  =  r/(l— r) , 
as  for  its  reciprocal. 

For  very  small  frequencies,  we  find  from  equation  (15)  that  in- 
dependently of  r  and  /*,  when  S  >  h',  there  results  a  response  to  flick- 
ering (intermittent)  illumination,  whereas  when  S  <  h'  there  results 
no  response,  h'  being  a  constant  which  is  the  effective  threshold.  Thus 
a  plot  of  /*(log  S/hx)  begins  at  (log  h'/hlf  0),  rises  vertically  at 
first,  then  flattens  off  while  approaching  a  final  slope  which  depends 
upon  r . 

The  relationship  between  f*  and  r  is  generally  determined  for  a 
constant  apparent  brightness  given  by  S'  =  Sr .  Using  the  approxi- 
mate expression  (16)  with  S'  =  constant,  we  find  that  f*(r)  rises 
rapidly  from  zero  to  a  maximum  for  r  <  ^(LDR  <  1)  and  then  falls 
to  zero  for  r  =  1   (Figure  4).    The  position  and  the  height  of  the 


.4 


Figure  4 


.8 


1.0 


maximum  depends  upon  S".  However,  when  r  is  near  zero  or  unity, 
the  approximations  break  down.  Furthermore,  equation  (16)  holds 
only  for  large  enough  /*.  Hence  the  exact  relation  f*{r)  may  be  of 
considerable  complexity.  For  the  various  experimental  relationships, 
one  may  consult  Bartley  (1941).  Most  of  the  results  quoted  agree 
with  the  above  prediction  that  for  constant  S',  f*(r)  is  decreasing  in 


46   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

r  when  r  is  just  less  than  one-half  (LDR  just  less  than  unity) . 

More  generally,  if  instead  of  alternating  a  stimulus  5  with  no 
stimulus,  we  had  alternated  S  +  AS  with  S ,  we  should  have  obtained 
the  same  results  with  log  S/lh.  replaced  by  log(l  +  A  S/S)  unless 
S  <  hx .  From  this  it  is  clear  that  for  a  constant  S  +  A  S  ,  an  increase 
in  S  >  h^  decreases  the  critical  frequency  f*.  Similarly,  an  equal  in- 
crease in  both  S  and  S  +  A  S  decreases  f*.  That  is,  an  illumination 
added  to  both  phases,  as  from  stray  light,  decreases  the  critical  flick- 
er-frequency. 

Although  we  have  referred  to  visual  phenomena  only,  one  may 
well  expect  that  analogous  properties  of  some  other  modalities  also 
could  be  accounted  for  roughly  by  just  such  a  simple  mechanism  as 
the  one  considered  here. 

We  have  assumed  throughout  that  a  >  b  and  <p  =  xp .  For  a  con- 
stant stimulus  this  gives  a  a  which  rises  rapidly  to  a  maximum  and 
then  subsides  more  slowly  to  zero.  This  resembles  the  "on"  activity 
of  the  "on-off"  fibers  of  the  retina.  Had  we  chosen  a  <  b  ,  we  should 
find  a  <  0  upon  application  of  a  constant  stimulus,  but  on  cessation 
of  the  stimulus  a  would  increase  to  a  maximum  and  subside  to  zero. 
This  resembles  the  activity  of  the  "off"  fibers.  We  should  have  ob- 
tained results  entirely  similar  to  these  above  but  with  (1—r)  re- 
placed by  r  ,  If,  however,  we  suppose  elements  of  both  types  to  be 
present  at  two  different  positions  with  different  parameters  and  with 
some  simple  interaction,  the  complexity  of  our  results  increases  great- 
ly. Again,  if  we  remove  the  restriction  $  =  y> ,  and  let  R$  =  xp  ,  with 
R  a  fraction  having  a  value  between  zero  and  one,  we  find,  that  for 
constant  S  ,  o-  increases  to  a  maximum  and  decreases  to  a  constant 
value  (1—R)$.  This  corresponds  to  behavior  of  the  continuously 
acting  elements  of  the  retina.  We  proceed  to  consider  some  properties 
exhibited  by  a  neural  element  of  this  latter  type. 

We  have  now  a  chain  of  two  neurons,  the  afferent  member  of 
which  is  of  the  mixed  type  with  <f>  —  xp/R  ,  0  <  R  <  1  .  Let  the  stimu- 
lation again  be  intermittent,  of  frequency  f  =  1/T  and  fractional 
stimulation  r  .  Let  the  intermittent  stimulation  be  continued  indef- 
initely. The  value  of  a  at  the  end  of  each  period  of  stimulation,  that 
is,  <r(t)  for  t  ->  oo  and  t  =  rT(mod  T) ,  can  be  determined  in  the 
manner  described  above.  If  6  is  that  value  of  a  divided  by  a  ( co )  for 
r  =  1 ,  6  is  essentially  the  ratio  of  a  at  the  end  of  a  period  of  stimula- 
tion for  a  particular  /  and  r ,  to  the  value  which  a  would  have  if  a 
stimulus  of  the  same  intensity  were  applied  continuously.  0  would 
be  better  defined  as  the  ratio  of  (a  -  ^)/(tr00  -  h),  but  we  shall  neg- 
lect the  threshold  as  compared  to  a  .   We  may  then  write 


SINGLE  SYNAPSE:  TWO  NEURONS 


47 


6  = 


1-  R 


1  -  e*rT 


-brT 


-uT 


R 


o-bT 


(17) 


Equation  (17)  gives  a  relation  between  the  relative  net  excitation  8 , 
the  fraction  r  of  time  of  stimulation,  and  the  frequency  /  =  1/T  .  For 
/  =  o  ,  6  =  1  f or  all  r .  But  for  /  =  oo ,  6  =  r .  Furthermore,  6  in- 
creases with  /  for  small  /  and  the  height  is  greater  for  small  r .  The 
type  of  relationship  between  0  and  /  for  various  values  of  r  is  shown 


m 


e 


48    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

in  Figure  5.  Notice  that  the  maximum  moves  to  the  right  with  in- 
creasing r  .  However,  if  instead  of  the  8  used,  we  had  taken  the  aver- 
age value  over  the  interval  rT ,  we  should  have  obtained  essentially 
the  same  results,  but  with  the  maximum  moving  to  the  left.  The 
equation  corresponding  to  expression  (17)  is,  however,  much  more 
complicated.  The  quantity  6  suggests  immediately,  in  the  terms  of  the 
visual  field,  the  relative  brightness  during  flicker.  Experiments  by 
Bartley  (1941)  show  essentially  the  same  type  of  variation  as  ex- 
pression (17)  but  there  is  no  significant  change  in  the  position  of  the 
maximum  with  r  on  a  range  from  \  nearly  to  1.  From  what  has 
been  stated  above,  one  could  probably  find  a  simple  average  which 
would  give  this  result. 

If  cj)  is  proportional  to  S  over  the  range  to  be  considered,  then 
multiplying  S  by  1/6  would  make  the  responses  equal.  For  the  fre- 
quency /  very  large,  6  =  r ,  that  is,  S  must  be  increased  to  S/r  to  ap- 
pear the  same  as  a  continuously  applied  S .  This  is  just  a  statement 
of  the  Talbot  law. 

From  this  brief  consideration  of  the  dynamics  of  two-neuron 
chains  we  have  been  able  to  derive  equations  predicting  quantitative 
relations  among  various  experimental  variables.  These  include  the 
relation  between  reaction-time  and  intensity  of  stimulation,  for  given 
change  in  stimulation  and  period  of  accommodation.  Similarly,  the 
duration  of  response  is  determined  in  terms  of  these  same  variables. 
A  Weber  ratio  is  determined  as  a  function  of  accommodation-time. 
Furthermore,  a  relation  is  determined  connecting  flicker-frequency, 
light-dark-ratio  and  intensity.  And  finally,  relations  are  determined 
between  relative  brightness  and  the  light-dark-ratio  and  frequency. 
In  some  cases,  quantitative  agreement  with  experiment  is  exhibited. 
In  others,  general  qualitative  agreement  is  obtained.  It  is  well  worth 
noting  at  this  point  that  while,  on  the  one  hand,  to  the  extent  that  the 
formulae  are  verified  these  manifold  relations  are  all  brought  within 
the  scope  of  a  single  unifying  principle,  on  the  other  hand  the  dis- 
cussion explicitly  introduces  a  great  many  problems  which  the  experi- 
ments only  vaguely  suggest. 


VII 

THE  DYNAMICS  OF  THE  SINGLE  SYNAPSE: 
SEVERAL  NEURONS 

When  a  pair  of  afTerents,  instead  of  the  single  one  assumed  in 
the  preceding  chapter,  form  a  common  synapse  with  a  single  efferent, 
the  resultant  a  at  this  synapse  is  capable  of  varying  with  time  in  a 
much  more  complicated  manner.  We  shall  consider  briefly  two  pos- 
sible applications  of  such  a  mechanism,  one  in  which  both  afferents 
are  supposed  to  be  affected  by  the  same  stimulus,  one  in  which  the 
stimuli  are  assumed  to  be  different. 

Consider  first  the  very  special  case  in  which  both  afferents  are 
stimulated  by  the  same  constant  stimulus.  Let  one  of  the  afferents 
be  of  the  simple  inhibitory  type  with  the  associated  yu  and  6T  .  Let 
the  other  be  of  the  mixed  type  with  </>2 ,  xp2  ,  a2  and  b2 .  Let  a2  >  >  bx 
or  b2  and  let  </>2  —  y\  —  y^  =  h  ,  the  threshold  of  the  efferent.  These 
assumptions  are  made  to  reduce  the  number  of  parameters.  We  em- 
ploy equation  (8)  of  chapter  i,  with  its  analogue  for  j  .  Then,  a2  be- 
ing large,  the  term  e-a*[  quickly  dies  out  so  that  except  for  very 
small  t , 

a  -  h  =  ?/>!  e~bit  +  y<2  e^1 .  (1) 

But  the  frequency  v  of  response  (chap,  i)  in  the  efferent  neuron  is 
proportional  to  a  —  h  when  this  is  not  too  large.  Since  y,  and  y>2  are 
arbitrary,  w e  may  replace  a  —  h  by  v  .  We  can  then  attempt  to  inter- 
pret equation  (1)  as  giving  the  variation  with  time  of  the  frequency 
of  the  response  of  an  efferent  neuron  when  a  constant  sustained  stim- 
ulus is  applied  to  its  afferents.  Equation  (1)  is  shown  in  Figure  1 
(curve)  for  particular  values  of  the  constants  while  in  the  same  figure 
are  shown  the  results  of  experiments  by  Matthews  (1931)  (points). 
In  these  experiments  the  stretch  receptors  in  muscle  were  stimulated 
by  means  of  attached  weights  and  the  variation  with  time  of  the  fre- 
quency of  discharge  was  determined.  Since  experimentally  the 
weights  tend  to  sink  with  time,  one  might  separate  the  stimulus  into 
two  parts,  a  constant  part  acting  on  the  second  neuron  and  a  variable 
part  acting  on  the  first,  making  this  one  excitatory  with  large  a.  The 
variable  part  would  presumably  be  roughly  an  exponentially  decreas- 
ing function  whose  decay-constant  must  be  equal  to  6,  of  equation 
(1).  This,  also,  would  lead  to  equation  (1).  Whether  or  not  the  ac- 
tual rate  of  decay  corresponds  sufficiently  with  the  experimentally 
determined  decay-constant  is  then  a  question  of  fact.    Formally,  we 

49 


50    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


u 

LU 


T 
O 

x 

>- 
u 

z 

<JJ 

3 

a 


time  r 


1.0 

IN  SECONDS 


1.5 


Figure  1. — Comparison  of  theory  with  experiment:  adaptation  to  muscular 
stretch.  Curve  theoretical,  predicted  by  equation  (1)  ;  points,  experimental 
(Matthews,  1931).   Abscissa,  duration  of  stretch;  ordinate,  frequency  of  response. 

obtain  the  same  result  regardless  of  the  point  of  view  adopted. 

The  structure  consisting  of  two  afferents  and  one  efferent  where 
the  afferent  Nx  is  now  simple  excitatory  leads  to  other  interesting 
results  when  the  stimulating  conditions  are  altered.  Let  Si  >  h^.  and 
S2  >  h2  be  constant  stimuli  applied  at  times  t±  =  0  and  t2  to  the  neu- 
rons Nx  and  N2  respectively.  Let  h  be,  as  before,  the  threshold  of  the 
efferent  neuron.  Also  let  ex ,  e2  and  j2  be  zero  initially.  Then  the  effer- 
ent neuron  is  excited  at  the  time  t'  when  a  =  ex  +  e2  —  j2  =  h  ,  or 


[1 


-&2< 


t'-h)]=hm       (2) 


^(1  -  e-^')  +  &  [1  -  e^»(*'-*.)]  -  xp2 

The  value  of  t'  depends  on  Sx ,  S2  and  £>  • 

To  simplify  the  problem  further,  let  <£2  =  y*2  and  let  o-2  be  always 
less  than  h  .  This  can  be  done  readily  by  restricting  S2  or  requiring 
that  cj>2  <  h  .  Then  no  response  can  occur  prior  to  the  moment  t  =  0  , 
even  if  t2  <  0  .  If  we  set  tw  =  t'  —  t2  we  obtain  by  solving  equation  (2) 


SINGLE  SYNAPSE:  SEVERAL  NEURONS 


51 


t'  =  --log 


a, 


h 


,     fa(S2) 

^ (e-w, 


-a2 1 


fa(Sx)      0i  (50 


■) 


(3) 


Finally  if  we  set  £r  =  t0  +  £',  where  t0  is  a  constant  as  in  equation  (2) 
of  chapter  vi,  we  can  determine  the  total  reaction  time  tr  as  a  function 
of  S2  through  fa ,  Si  through  fa ,  and  of  tw ,  which  differs  by  t0  from 
the  time  by  which  S2  precedes  the  initiation  of  the  response.  As  S2  is 
a  stimulus  which  precedes  Sx  and  affects  the  response  time  to  Si ,  but 
is  itself  incapable  of  producing  the  response,  it  may  be  considered  a 
warning  stimulus.  Hence  we  may  take  equation  (3)  to  predict  the 
kind  of  results  to  be  obtained  in  an  experiment  in  which  a  particular 
stimulus  of  intensity  Si  has  been  preceded  by  a  warning  stimulus  S2 
and  produces  a  response  in  a  time  tr(S1}  S2 ,  tw)  depending  on  the 
strength  of  the  warning  stimulus  as  well  as  upon  the  manner  in  which 
Sx  and  S2  are  spaced  in  time. 

For  the  particular  case  in  which  a  fixed  Si  and  S2  are  used  and 
for  tl0  >  >  tr ,  we  may  write  equation  (3)  as 


tr  =  t0' log  [1  +  D(e-6*<-  -  e-°»'»)] 

a, 


(4) 


in  which 


U'  =  t0 log  (1-h/fa)  >  D  =  fa/ (fa  -  h) 


12  16  20  2* 

PREPARATORY  INTERVAL    IN    SECONDS  tw  — * 


Figure  2. — Comparison  of  theory  with  experiment:  effect  of  time  of  occur- 
rence of  warning  stimulus  upon  the  reaction-time.  Curve,  theoretical  predictions 
by  equation  (4);  points,  experimental  (Woodrow,  1914).  Abscissa,  interval  be- 
tween presentation  of  warning  and  effective  stimuli;  ordinate,  interval  between 
presentation  of  effective  stimulus  and  occurrence  of  response. 


52   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

In  Figure  2  is  shown  a  comparison  of  equation  (4)  (curves)  with 
experimental  data  (points)  by  H.  Woodrow  (1914)  in  which  the  effect 
of  the  interval  between  warning  stimulus  and  stimulus  proper  on  the 
reaction-time  was  measured.  The  upper  curve  was  obtained  under 
the  condition  that  the  successive  values  of  tw  in  the  experiment  were 
mixed  randomly;  the  lower  curve  was  obtained  under  the  condition 
that  the  value  of  tw  was  kept  the  same  for  a  number  of  trials  before 
being  changed  to  another  value.  As  the  conditions  are  different  in  the 
two  cases,  one  might  expect  that  the  parameters  would  also  be  differ- 
ent. Further  details  are  given  in  the  legend  of  Figure  2.  For  a  dis- 
cussion of  a  mechanism  which  can  differentiate  between  these  two 
conditions,  the  reader  is  referred  to  Landahl  (1939a). 


VIII 
FLUCTUATIONS  OF  THE  THRESHOLD 

It  has  been  assumed  thus  far  that  the  threshold  of  a  neuron  is  a 
constant  which  does  not  depend  on  time.  Actually  it  varies  from 
moment  to  moment  and  when  we  speak  of  the  threshold  as  a  constant 
we  must  understand  by  this  some  mean  value  of  a  group  of  measure- 
ments of  the  threshold.  A  more  complete  description  would  give  also 
a  measure  of  the  variability.  The  threshold  may  vary  with  many 
changes  in  the  organism.  These  variations  would  generally  be  rather 
slow.  But  within  the  neuron  and  in  its  immediate  surroundings  there 
occur  rapid  minute  fluctuations  in  the  concentrations  of  the  various 
metabolites.  The  work  of  C.  Pecher  (1939)  indicates  very  strongly 
that  it  is  these  fluctuations  in  concentration  that  are  responsible  for 
the  variations  in  the  thresholds  of  the  peripheral  fibers  with  which 
he  experimented.  His  calculations  showed  that  as  few  as  some  thous- 
and ions  was  sufficient  to  produce  excitation.  From  the  kinetic  theory, 
one  should  then  expect  that  the  per  cent  variation  in  the  threshold 
should  be  one  hundred  divided  by  the  square  root  of  the  number  of 
ions  necessary  for  excitation  (Gyemant,  1925).  This  value  in  terms 
of  the  coefficient  of  variation  is  of  the  order  of  a  few  per  cent  and 
is  comparable  with  the  values  obtained  experimentally. 

We  may  make  the  calculation  of  the  variation  as  follows.  In 
order  for  excitation  to  occur,  it  is  necessary  to  stimulate  a  minimal 
region  of  a  neuron.  Suppose  this  to  be  a  node  of  Ranvier.  Let  the 
width  of  the  node  be  d  oo  104  cm,  the  radius  of  the  fiber  r  oo  10-4  cm. 
The  effect  of  an  ion  is  small  at  distances  of  a  few  diameters.  Thus 
ions  a  few  diameters  removed  from  the  cell  surface  will  have  little 
influence  on  the  surface.  Let  this  distance  of  influence  be  d  oo  107  cm. 
Then  the  volume  within  which  the  ions  affect  the  excitability  of  the 
neuron  is  2nrdd  .  If  C  oo  10-5  is  the  molar  concentration  of  the 
ions,  and  if  N  is  Avogadro's  number,  the  total  number  of  ions  influenc- 
ing the  excitability  is  2nrddCN  and  thus  the  per  cent  fluctuation 
is  given  by  100 /\/2jirddCN  oo  2%.  Had  we  used  the  area  of  an  end- 
foot  (c\3l0-7),  the  same  sort  of  result  would  have  been  obtained.  But 
because  of  the  variations  in  these  quantities  one  cannot  exclude  the 
possibility  that  rather  large  variations  may  occur.  The  calculations 
only  indicate  that  the  fluctuations  about  the  threshold  may  be  appre- 
ciable. As  long  as  the  range  in  variation  is  not  comparable  with  the 
threshold  itself,  the  kinetic  theory  requires  that  the  fluctuations  be 
distributed  normally  to  a  high  degree  of  approximation.    That  this 

53 


54   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

is  the  case  for  single  nerve  fibers  is  illustrated  in  the  data  by  C. 
Pecher  (Landahl,  1941c). 

If  for  a  particular  neuron  the  mean  value  of  the  threshold  is  h , 
the  coefficient  of  variability  is  v  ,  and  if  p(C)  represents  a  normal 
curve  of  unit  area,  then  the  probability  P  of  a  response  in  the  absence 
of  a  stimulus  is  given  by  the  integral  of  p(C)  from  1/v  to  oo.  If  t  is 
the  least  time  for  a  fluctuation  to  have  effect,  then  after  a  time  t/P 
one  could  reasonably  expect  a  chance  response.  The  mean  frequency 
of  such  responses  would  then  be  given  by  P/r  per  second.  These  re- 
sponses would  not  be  periodic.  If  t  is  taken  to  be  of  the  order  of  mag- 
nitude of  10~3  seconds,  then  for  v  =  30%  the  mean  time  between  re- 
sponses would  be  a  few  seconds,  while  for  v  =  20%,  the  mean  time  be- 
tween responses  would  be  a  number  of  hours.  From  this  we  see  that 
the  probability  of  a  chance  response,  even  over  a  considerable  period 
of  time,  becomes  negligible  rapidly  as  v  becomes  much  smaller  than 
one-fifth.  But  one  should  consider  also  the  slower  changes  in  the  en- 
vironment of  the  neuron  which  not  only  changes  the  threshold  but 
also  its  degree  of  variation. 

Variations  in  the  threshold  would  cause  the  response-times  and 
other  measurable  variables  to  be  distributed  in  some  manner  about 
the  value  corresponding  to  the  mean  value  of  the  threshold.  As  an 
illustration,  let  us  estimate  the  dependence  of  a  measure  of  the  varia- 
tion in  response-times  on  the  intensity  of  the  stimulus.  We  consider 
the  case  of  a  simple  excitatory  afferent  stimulated  by  a  constant  stim- 
ulus S  and  acting  on  an  efferent  of  threshold  h  .  Suppose  that  tx  is 
the  value  of  t  for  which  e  =  h .  Then  if  h  is  decreased  by  an  amount 
vh ,  e  =  h  is  satisfied  by  t  =  U.  —  a.   Then 

1        /  vh     \ 

a  =  -log[    1+ (D 

a        V  <£  —  h   J 

is  an  average  variation  in  the  reaction-time  due  to  the  variation  in 
threshold.  In  general,  since  one  must  consider  more  than  one  chain, 
one  may  suppose  that  variations,  essentially  independent  of  <f> ,  are 
introduced  at  other  synapses  and  at  the  end-organ.  Let  a0  be  a  meas- 
ure of  the  total  effect  of  this  variation.  Then  the  measured  variation 
in  the  response-time  will  be  given  by  the  square  root  of  the  sum  of 
the  squares  of  these  two  variations.  As  v  is  generally  quite  small 
a  =  vh/((p  —  h)a  and  thus 


at  =  Veto2  +  vya*{4>/h  -  l)2,  (2) 

and  we  have  a  relation  between  a  measure  of  the  variation  in  re- 
sponse-times and  the  stimulus-intensity  in  terms  of  ^  .   We  may  com- 


FLUCTUATIONS  OF  THE  THRESHOLD 


55 


pare  this  result  with  the  results  of  experiments  by  Berger  and  Cattell 
(1886)  cited  in  chapter  vi,  in  which  the  mean  variations  of  the  re- 
sponse-times were  measured.  We  may  use  the  same  parameters  ex- 
cept for  (70  which  is  arbitrary  for  this  curve.  Thus  we  have  one  para- 
meter to  determine  this  relation.  The  data  are  incomplete  in  the  re- 
gion of  the  threshold,  but  the  comparison  in  Figure  1  is  made  for 


.040 


.030  - 


O.020 


.010 


10  100 

STIMULUS  INTENSITY     S— * 


1000 


Figure  1. — Comparison  of  theory  with  experiment :  the  variation  in  reaction- 
times  as  a  function  of  stimulus-intensity.  Curve,  theoretical  predictions  by  equa- 
tion (1);  points,  experimental  (Cattell,  1886).  Abscissa,  intensity  of  stimulus; 
ordinate,  mean  variation  in  reaction  times. 

illustrative  purposes  primarily.  Nothing  has  been  said  of  the  type 
of  distribution  one  would  expect  for  a  particular  stimulus-intensity. 
This  would  require  a  more  detailed  analysis.  We  wish  only  to  indi- 
cate the  kind  of  effects  due  to  the  variations  in  the  threshold.  In  the 
next  chapter  we  shall  show  how  they  provide  a  possible  basis  for  the 
distribution  of  judgments  in  situations  that  require  some  form  of 
discrimination. 


IX 


INTERCONNECTED  CHAINS:  PSYCHOPHYSICAL 

DISCRIMINATION 


In  chapter  iii  we  dealt  with  the  general  problem  of  the  interac- 
tions among  interconnected  parallel  chains  of  neurons,  and  more  es- 
pecially with  the  mutual  reduction  of  the  a's  developed  by  the  simul- 
taneously stimulated  chains  when  the  interconnections  are  inhibitory. 
We  also  exhibited  a  mechanism  capable  of  transmitting  excitation 
when  the  intensity  of  the  stimulus  lies  on  a  limited  range  only.  In 
this  chapter  we  shall  apply  these  considerations  to  the  interpreta- 
tion of  sequences  of  a  stimulus  and  a  response  of  a  type  often  stud- 
ied by  psychologists,  which  we  shall  speak  of  as  discriminal  se- 
quences. By  a  discriminal  sequence  we  shall  mean  any  sequence  in 
which  the  response  is  one  of  a  limited  set  of  qualitatively  different 
possible  responses,  while  that  feature  of  the  stimulus  that  deter- 
mines which  response  of  the  set  is  to  occur  is  either  its  absolute  in- 


m 


m 


ti  \J4    X    S.   Jz  \ti 


ft    I 


t 


% 


1  ^      Ml 


V 


r> 


5 


£; 


$ 


Figure  1 
56 


PSYCHOPHYSICAL  DISCRIMINATION  57 

tensity  or  its  intensity  relative  to  that  of  some  other  specified  stimulus. 

We  consider,  then,  parallel  neurons  or  chains  interconnected  by 
inhibitory  neurons  (Figure  1),  and  we  impose  here  the  further  re- 
strictions that  their  inhibitory  effect  is  numerically  the  same  as  the 
excitatory  effect  due  to  the  neuron  with  the  same  origin.  Then  a  stim- 
ulus Sx  may  produce  a  response  Ri ,  and  S2  a  response  R2  ,  if  the  stim- 
uli are  presented  separately.  But  if  the  stimuli  are  presented  to- 
gether, then  the  response  Rx  will  be  produced  if  Sx  exceeds  S2  by  an 
amount  which  depends  upon  the  thresholds  of  the  efferent  neurons. 
If  the  difference  between  Sx  and  S2  is  too  small,  neither  response  oc- 
curs. If  the  thresholds  are  negligibly  small,  the  response  Rx  occurs 
alone  if  51  >  S,  and  R2  if  S2  >  S1 . 

Because  of  fluctuations  of  the  type  discussed  in  the  preceding 
chapter,  the  values  of  s-l  and  e2  produced  by  the  afferent  neurons  will 
not  generally  be  exactly  equal  even  when  Sx  =  S2 .  If  St  is  slightly 
greater  than  S2 ,  there  is  a  certain  finite  probability,  less  than  one- 
half,  that  the  response  R2  will  be  given  instead  of  Ri ,  the  probability 
decreasing  as  the  difference  between  Sx  and  S2  is  increased. 

Suppose  that  fluctuations  occur  only  in  the  thresholds  of  the 
afferent  neurons.  The  fluctuation  of  the  threshold  of  any  neuron 
causes  fluctuation  of  the  o-  produced  by  this  neuron,  and  we  shall  pos- 
tulate the  distribution  of  o-  rather  than  that  of  h .  Furthermore,  be- 
cause of  the  interconnections  between  the  neurons,  an  increase  in  o- 
at  the  terminus  of  one  afferent  has  the  same  effect  as  a  decrease  in  a 
at  the  other,  so  that  formally  we  may  regard  the  fluctuation  as  oc- 
curring at  only  one  synapse  (Landahl,  1943).  Thus  we  shall  assume 
that  the  thresholds  are  constant  but  that  at  synapse  sx  ,  <r  —  e  —  j  +  C 
where  s  and  j  result  from  the  activity  of  the  afferents  but  C  is  nor- 
mally distributed  about  zero. 

We  shall  assume  that  the  net  is  completely  symmetrical.  Let  h' 
be  the  threshold  of  either  efferent.  As  we  are  assuming  that  y  and  b 
of  the  interconnecting  inhibitory  neurons  are  equal  to  <£  and  a  of  the 
parallel  excitatory  neurons,  a1  =  e3  —  j\  =  —  a2  =  —  (e4  —  y3).  Thus 
o-i  and  (72  are  equal  and  opposite.  Using  equation  (6)  of  chapter  i,  we 
may  write  the  stationary  values  as 

Si 

ax  —  -  a2  =  p\0g— .  (1) 

Now  a!  exceeds  h'  if  the  stimulus  S-l  exceeds  S2  sufficiently.  But  then 
e,  will  exceed  e2  by  some  value  h  .  We  may  summarize  as  follows: 

If  er  —  e2  —  C  >  h ,  response  Ri  is  given ; 


58   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

if  h  i?  £i  —  £.2  —  £  =  —  h  ,  there  is  no  response ; 
if  —  h  >  £i  —  £2  —  £  ,  response  R2  is  given. 

If  p(£)d£  is  the  probability  that  £  has  a  value  in  the  range  £  to 
£  +  d£ ,  then  the  probability  that  response  Rx  occurs  is  obtained  by 
integrating  p(£)  with  respect  to  £  from  minus  infinity  to  (e^  —  e2  —  h). 
This  becomes  evident  when  we  see  that  if  £  is  any  value  less  than 
£1  —  £2  —  h ,  response  Rx  is  produced.  If  we  let  f\  be  the  probability 
of  response  Rx ,  P2  the  probability  of  response  R2 ,  and  P0  the  prob- 
ability of  neither  response,  and  if  we  define 


then  we  may  write 


P(x)  =  f*  p(£)d£,  (2) 

J  -"50 

P^  =  P(e1-B2-h),  (3) 

P2  =  P(-e1  +  e2-h),  (4) 

Po  =  l-P*-Pz.  (5) 

If  Si  =  S2 ,  response  Ri  may  be  considered  the  correct  response 
and  R2  the  wrong  response.  In  this  case  Pr  =  Pc ,  the  probability  of 
a  correct  response,  and  P2  =  Pw ,  the  probability  of  a  wrong  re- 
sponse. Any  failure  to  respond,  or  any  response  other  than  cor- 
rect or  wrong  such  as  "equal,"  "doubtful,"  could  be  included  in  the 
proportion  to  be  identified  with  P0  •  It  is  commonly  the  case  that  when 
a  categorical  judgment  is  required,  so  that  either  Rx  or  R2  is  made 
at  each  trial,  the  subject  must  lower  his  criteria  for  judgment.  We 
may  interpret  this  with  reference  to  the  structure  studied  by  assum- 
ing a  lowered  threshold.  For  this  case  we  set  h  =  0  ,  whence  P0  =  0 
and  Pi  +  P2  =  1.  Thus  from  a  knowledge  of  only  the  standard  error 
of  the  probability  distribution,  one  is  able  to  calculate  the  probabil- 
ities of  the  various  responses  to  any  given  pair  of  stimuli  when  the 
judgments  are  categorical;  the  additional  parameter  h  enters  when 
"doubtful"  judgments  are  allowed. 

Since  complete  symmetry  has  been  assumed,  it  follows  that 
Px  =  P2  for  Si  =  S2 .  In  general,  this  is  not  true  for  the  observed  pro- 
portions. The  amount  by  which  the  observed  proportions  differ  from 
equality  is  a  measure  of  the  bias  of  the  subject.  The  simplest  inter- 
pretation of  the  bias  is  that  the  afferent  thresholds  are  not  exactly 
equal  so  that  the  mean  values  £x  and  £2  are  not  equal  for  Sx  =  S2 ,  but 
£x(£)  =  s2(S)  +  x0.  Although  x.0  will  depend  on  Si,  we  shall  not 
consider  this  any  further,  preferring  rather  to  incorporate  #0  into  £ , 
so  that  p(£)  has  a  mean  value  of  x0  instead  of  zero.  Thus  modified, 
the  mechanism  may  be  applied  to  the  experimental  data  by  F.  M. 
Urban  shown  in  Table  1  (Urban,  1908).   These  are  the  average  re- 


PSYCHOPHYSICAL  DISCRIMINATION 


59 


suits  from  observations  made  on  seven  subjects.  The  first  entry  in  the 
table,  .012,  gives  the  proportion  Pg  of  judgments  that  a  weight  of  84 
gms  is  heavier  than  the  standard,  which  is  100  gms  in  every  case.  This 
is  for  the  case  in  which  judgment  of  "equality"  is  permitted,  the  pro- 
portion being  Pe  =  .027.  On  the  same  line  in  the  last  column  is  given 
the  proportion  Pi  =  1  —  Pg  —  Pe  of  times  the  weight  of  84  gms  is 


.50 

Experiment 


a 


/.CO 


Figure  2. — Comparison  of  theory  with  experiment:  distribution  of  judgments 
of  relative  weights.  In  2a  the  abscissa  of  each  point  is  the  experimental  value 
(Urban,  1908)  of  the  proportion  of  judgments  of  the  indicated  type  and  the  ordi- 
nate is  the  proportion  theoretically  predicted  by  equations  (l)-(5).  Thus  perfect 
agreement  would  be  indicated  if  all  points  lay  on  the  straight  line.  In  2b  the 
curve  is  theoretical,  the  points,  experimental.  The  abscissa  of  each  point  is  the 
proportion  of  "greater  than"  or  of  "less  than"  judgments  as  the  case  may  be,  the 
ordinate,  the  proportion  of  "doubtful"  judgments  when  comparing  the  same  "vari- 
able" stimulus  with  the  standard". 


60    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


n 

ID 

./> 

ro 

o 

>£> 

CO 

ro 


00 


lO 

f) 

n 

iT) 

i" 

CD 

r 

S330NVD  fc)3l3W  Nl  30N3fc)3JJI0  AHSN31NI   1HOH 

Figure  3. — Comparison  of  theory  with  experiment:  distribution  of  judgments 
of  relative  brightness.  Curve,  theoretical  from  equations  (l)-(5) ;  points,  experi- 
mental (Kellogg,  1930).  The  ordinate  is  the  difference  between  the  intensities 
of  the  "variable"  and  the  "standard"  stimuli,  in  meter-candles.  The  abscissa  is 
in  every  case  the  proportion  of  judgments  of  the  type  in  question:  for  the  solid 
circle,  "greater  than"  categorical  judgment;  for  the  open  circle,  "greater  than" 
with  "doubtful"  judgments  permitted;  for  the  crosses,  the  proportion  of  "doubt- 


PSYCHOPHYSICAL  DISCRIMINATION  61 

judged  to  be  lighter  than  the  standard.  Directly  below  the  first  entry- 
is  the  number  .020  which  is  the  proportion  of  times  the  weight  of  84 
gms  is  judged  heavier  than  the  standard  when  the  judgment  of  "equal- 
ity" is  ruled  out,  and  so  on  for  the  other  entries. 

In  Table  1  (p.  72)  are  given  the  corresponding  probabilities  com- 
puted by  equations  (1)  through  (5),  using  for  the  standard  deviation 
of  the  distribution  5.7  gms,  for  the  threshold  h  =  2.1  gms,  and  for  the 
bias  or  constant  error  x  =  2.75  gms.  These  values  were  determined  from 
the  three  values  in  parentheses  on  the  left,  whence  the  corresponding 
values  on  the  right  are  the  same.  It  is  to  be  noted  that  the  parameters 
are  all  measured  in  grams.  This  is  done  for  convenience  only,  as  ac- 
tually there  is  an  unknown  constant  which  must  multiply  each  para- 
meter to  give  the  values  in  terms  of  e  and  j  .  Furthermore,  a  linearity 
between  S  and  e  is  implied,  which  is  nearly  the  case  in  the  small  range 
considered.  In  both  the  Table  and  Figure,  the  proportions  Pg ,  and 
Pi  are  used.  These  are  respectively  the  proportions  of  judgments  of 
"greater,"  and  "lesser."  Their  relation  to  Pc  and  Pw  is  evident. 

The  agreement  between  the  theory  and  experiment  is  illustrated 
in  Figure  2.  Complete  agreement  would  be  indicated  if  all  points  fell 
on  the  line  of  slope  one.  The  relation  between  the  proportions  of 
judgments  of  "equality"  and  the  proportions  of  the  correct  and  wrong 
responses  is  also  shown  in  Figure  2.  If  these  results  are  plotted  on 
probability  paper,  the  predicted  results  will  be  simply  three  parallel 
lines.  The  experimental  data  confirm  this  rather  well  (Landahl, 
1939b) .  The  results  from  each  of  the  seven  subjects  showed  the  same 
trend. 

When  one  considers  the  visual  data  by  F.  M.  Urban,  one  finds 
that  the  averaged  data  for  several  subjects,  as  well  as  those  for  individ- 
uals, cannot  be  so  simply  interpreted.  With  judgments  of  "equality" 
allowed,  neither  the  proportions  of  correct  responses  nor  those  of 
wrong  responses  follow  the  integral  of  the  normal  distribution  with- 
in the  limits  of  experimental  error;  with  judgments  of  "equality"  ex- 
cluded the  proportions  are  not  peculiar.  This  suggests  that  the  thresh- 
old h  is  not  constant  but  is  affected  by  the  value  of  <jx  .  In  order  to 
preserve  symmetry,  the  effect  on  the  threshold  of  Sx  >  S2  must  be 
supposed  the  same  as  that  of  S2  >  &i  if  St  and  S2  are  simply  inter- 
changed. We  may  suppose  that  the  lowering  of  the  threshold  h ,  due 
to  the  change  of  experimental  situation  when  judgments  of  "equality" 
are  ruled  out,  is  the  result  of  the  activity  of  some  outside  group  of 

ful"  judgments  has  been  added  to  the  proportion  of  "greater  than"  judgments. 
In  the  inset  the  curve  represents  the  values  of  h  predicted  by  equation  (6)  plotted 
against  the  difference  of  the  intensities  of  the  two  stimuli  and  the  points  repre- 
sent the  values  computed  directly  from  the  data. 


62    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

neurons.  If  a  neuron  of  negligible  threshold  having  afferent  synapses; 
at  sx  and  s2  tends  to  excite  this  outside  group  of  neurons,  then  the 
threshold  h  will  decrease  with  the  absolute  value  of  ux  or  <r2,  these 
being  proportional  to  the  absolute  values  of  a3  and  <x4 .  Thus  h(\<j\) 
should  decrease  linearly  for  small  |<r|,  though  h  cannot  become  nega- 
tive.  If  we  set 

h  =  h0e-°w,  (6) 

we  have  a  suitable  form,  with  but  one  new  parameter  introduced. 

In  Figure  3  is  shown  a  comparison  between  theory  and  experi- 
ment for  the  visual  data  by  W.  N.  Kellogg.  The  curves  are  computed 
from  the  equations  by  setting  the  standard  deviation  equal  to  0.58 
meter-candles,  —  x0  =  —0.10  meter-candles,  ho  =  0.49  meter-candles 
and  9  =  1.14  meter-candles1.  The  intensity  of  the  standard  was  21.68 
meter-candles.  In  the  inset  of  the  Figure  is  shown  a  comparison  be- 
tween equation  (6)  and  the  values  of  h  determined  from  the  data. 
These  are  symmetric  about  —  xc .  This  type  of  relationship  between 
h  and  the  difference  between  the  stimuli  was  found  for  each  of  the 
individuals  upon  whom  the  experiment  was  carried  out.  The  deriva- 
tive of  h(S)  is  discontinuous  at  —x(l.  One  would  not  expect  to  ob- 
serve such  a  discontinuity  even  if  it  were  present.  If  the  threshold  of 
the  neuron  which  produces  the  change  of  h  with  stimulus  difference,, 
had  not  been  neglected,  the  value  of  h  would  have  been  a  constant  in 
the  neighborhood  of  —  x,0 .  For  these  reasons,  a  dotted  curve  is  intro- 
duced in  the  Figure  to  indicate  that  the  discontinuities  are  not  ex- 
pected to  appear  in  the  data.  For  further  details  the  reader  is  re- 
ferred to  the  paper  by  H.  D.  Landahl  (1939b). 

In  the  case  of  the  auditory  data   (Figure  4)  by  W.  N.  Kellogg 
(1930)  one  finds  that  a  further  asymmetry  is  present.  A  rather  accu- 
rate representation  of  the  data  results  if  one  assumes  that  the  effect 
on  the  threshold  h  due  to  stimuli  for  which  Sr  >  S2  is  not  the  same  as 
that  for  which  S2  >  <S\  .    Since  for  this  modality  S2  —  Si  may  be  a. 
rather  large  fraction,  the  first  term  of  the  expansion  of  equation  (1) 
leads  to  noticeable  error.  Thus  the  parameters  are  measured  in  terms. 
of  the  logarithm  of  the  ratio  of  the  stimuli.    If  we  let  h.0  =  0.43, 
x<,  =   .02  ,   the  standard   deviation   0.18   or  approximately   9  (milli- 
volts)2, and  9  =  2.8  for  o^  >  0  ,  we  obtain  the  curves  shown  in  Figure 
4.   The  points  are  the  experimental  values  obtained  by  averaging  the 
results  from  a  number  of  subjects.   The  inset  shows  the  relationship 
between  h  and  log  Si/S2  which  is  decidedly  asymmetric.    If  this  be 
considered  significant,  one  might  attempt  to  correlate  the  asymmetry 
with  the  mode  of  presentation  or  perhaps  with  the  modality.    The. 


PSYCHOPHYSICAL  DISCRIMINATION 


63 


o 
o 

_j 
o 

l" 


O 
O 
O 

-I 
_J 
U 


>■ 
o 

a 
< 


.SinOAIllllAI  Nl   A1ISN31NI   QNOOS 


^t 

* 

■<* 

rO 

'-D 

1 

i 

i 

1 

o 
rvj 

1 
o 

iO 


1 

fi) 

1 

oS 
1 

1 
o 

'"  * 

1 
i" 

0» 
O) 

00 


o 

0> 


o 

00 


o 

r- 

o 

I* 

o   z 

S°- 

o 


o 


_  o 


rf> 


-    AJ 


%  001 

Figure  4. — Comparison  of  theory  with  experiment:  distribution  of  judgments 
of  relative  loudness  (data  from  Kellogg,  1930).  The  representation  is  the  same 
as  that  in  Figure  3  except  for  the  use  of  the  logarithmic  scale  of  intensities, 
necessitated  by  the  large  relative  range  in  the  stimuli. 


64   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

asymmetry  appeared  fairly  clearly  in  the  results  from  each  of  the 
individuals  taking-  part  in  the  experiment. 

On  the  basis  of  the  mechanism  considered,  it  is  essential  that  the 
stimuli  be  presented  simultaneously.  However,  a  complication  of  the 
mechanism  has  been  considered  for  which  simultaneous  presentation 
is  not  necessary.  Essentially  the  same  results  may  be  obtained  if  the 
stimuli  are  presented  in  succession  (Landahl,  1940a). 

Since  the  integral  of  the  normal  distribution  cannot  be  given  in 
closed  form,  it  is  convenient  to  introduce  an  approximation  by  which 
closed  solutions  can  be  obtained.  This  is  especially  desirable  when 
one  wishes  to  use  the  results  in  other  situations.  With  the  distribu- 
tion (Landahl,  1938a) 

p(C)=-e-fci<i  (7) 


we  obtain,  if  sx>  &.., 


log2Pw  +  k(ei-e2)  =0  (8) 


to  determine  the  probability  of  a  wrong  response  when  a  categorical 
judgment  is  required.  The  applicability  of  this  approximation  has 
been  tested  by  a  comparison  of  theory  and  experiment  made  else- 
where (Landahl,  1938a). 

While  our  definition  of  a  discriminal  sequence  rules  out  the  con- 
currence of  the  alternative  responses  and  hence  requires  that  the  in- 
hibitory neurons  connecting  the  parallel  excitatory  chains  shall  have 
activity-parameters  at  least  as  great  as  those  of  the  excitatory  chains 
themselves,  it  is  natural  to  consider  also  the  case  where  this  restric- 
tion is  removed.  We  noted  in  chapter  iii,  in  the  case  of  two  parallel 
chains  with  inhibitory  interconnections,  what  qualitatively  different 
effects  might  follow  the  simultaneous  stimulation  of  both  chains  as 
different  relations  are  imposed  upon  the  parameters  of  the  constitu- 
ent neurons.  We  select  for  further  consideration  here  only  the  sym- 
metric structure  consisting  of  two  parallel  chains  with  crossing  in- 
hibition where  a  >  /5  (cf.  equations  (4)  chapter  iii),  in  which  case 
concurrent  transmission  along  the  two  paths  will  occur  when  the  two 
stimuli  are  sufficiently  strong  and  not  too  greatly  different. 

These  chains  may  lead  from  neighboring  cutaneous  receptors, 
from  neighboring  retinal  elements,  or  from  organs  of  two  disparate 
sensations.  The  responses  which  they  occasion  may  be  overt  bodily 
movements  or  they  may  be  merely  awareness  of  the  sensations.  The 
mechanism  has  a  possible  application  wherever  there  is  interference 
by  a  stimulus  of  one  type  with  evocation  of  a  response  by  another, 
and  reciprocally.   It  is  at  once  apparent  that  the  interference  by  the 


PSYCHOPHYSICAL  DISCRIMINATION  65 

one  stimulus  with  the  other's  response  may  occur  even  though  the 
first  stimulus  would  be  inadequate,  if  presented  alone,  to  produce  its 
own  response.  Thus  if  the  application  were  made  to  the  interaction 
of  auditory  with  visual  perception,  the  mechanism  provides  that  even 
a  subliminal  auditory  stimulus  would  raise  the  absolute  threshold  for 
visual  stimulation.  With  appropriate  modifications — crossing  excita- 
tion instead  of  crossing  inhibition — the  possibility  of  a  mutual  lower- 
ing of  threshold  could  be  similarly  treated. 

To  link  the  mechanism  with  crossing  inhibition  more  substan- 
tially with  possible  experimental  results,  we  consider  the  effect  of 
threshold-fluctuations,  or,  what  is  more  convenient  and  mathemati- 
cally equivalent,  random  variations  in  the  o-'s  at  sx  and  s2 .  Since 
a^j?we  cannot,  as  before,  represent  the  combined  effect  by  varia- 
tions at  only  one  synapse,  but  any  variations  occurring  also  at  Si 
and  s2  could  be  formally  accounted  for  by  suitably  modifying  the 
distribution-functions  at  the  first  two  synapses  alone  and  we  suppose, 
for  simplicity,  that  with  this  modification  the  resulting  distributions 
are  identical  at  the  two  synapses.  Denote  these  functions  by  p(C), 
C  being  the  random  addition  to  either  synapse  and  having  zero  as  its 
mean. 

The  mutual  influence  of  the  stimuli  upon  absolute  thresholds  can 
be  determined  from  an  investigation  of  near-threshold  stimulation 
where  the  functions  <f>  and  xp  can  be  represented  linearly: 

<j>(S)  =aS-a',y>(S)  =ps-p,     a  >  p  .  (9) 

Corresponding  to  equations  (1),  chapter  iii,  we  have 

a1  =  a(S1  +  Cx)  -a'-/?(S2  +  C2)  +  /J', 

(10) 

<r2  =  a(S2  +  k)  -  a'  -  fiiSi  +  d)  +  fi' . 

Then  the  respective  conditions  for  the  responses  R1  and  R2  are, 

aS1-pS2  +  at1-pz2-h>0,  (11) 

aS2-  pS1  +  a^~  PCi~h>0,  (12) 

where 

h  =  a'-  p  +  h' ,  (13) 

and  each  efferent  from  sx'  and  s2  has  the  threshold  h' .  Let  (RR), 
(RO) ,  (OR)  and  (00)  denote  the  occurrence  of  both  responses,  the 
first  only,  the  second  only,  and  no  response,  respectively.  The  prob- 
abilities of  these  events,  where  Si  and  S2  have  given  values,  are 

P(RR)=     f     p(Ci)  J       f»(f.)<Zf»<Ztif  (14) 


66    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 
P(RO) = 

(15) 

-Si+7»/(<H3)  °°  oo  oo 

f  Piti)    j  P(C2)  dC2dCi  +  J     p(Ci)     J  P(Ca)  #.dk, 

-oo  ^(fi+Si)/a+/i/a-S2  -Si+7i/(<H3)  a(£i+Si)//3-ft/0-S2 

P(Otf)  = 

(16) 

-81+h/(a-p)         a(£i-:S,)/0-fi/|3-S2  oo  oo 

j  p(Ci)     Jp(C2)^C2^C1+J    2>(Ci)       JV(C2)<#3<#i, 

-oo  -oo  -Si+fe/(o-|3)  /3(fi+S!)/a+Va-S2 

-S1+7i/(a-(3)  (3(fi+S1)/a+ft/a-S2 

P(00)=J     2?(d)        J"p(C*)#»#i.  (17) 

-oo  a(£i+Si)/0-/!/0-S2 

Other  expressions  for  P(RR)  and  for  P{00)  can  be  obtained  by 
interchanging-  subscripts.  These  four  P's  are  functions  of  Sx  and  S2 
whose  values  are  experimentally  determinable;  their  sum  is  unity  so 
that  only  three  are  independent.  If  p(C)  is  given  they  depend  upon 
the  parameters  a ,  /5  and  h;  if  the  distribution  29(C)  is  assumed  to  be 
normal  there  is  an  additional  parameter,  the  standard  deviation.  Pear- 
son's tables  of  tetrachorics  can  be  utilized  for  determining  these  para- 
meters from  the  empirical  frequencies.  The  quantities  St  and  S2  are 
not  the  intensities  of  the  external  stimuli  but  some  monotonic  func- 
tions of  these;  however,  at  the  near-threshold  level  it  is  permissible 
to  regard  these  functions  as  linear. 

In  the  preceding  paragraphs  we  have  considered  the  case  of  two 
stimuli  simultaneously  presented  to  the  organism,  with  two  alterna- 
tive responses  permitted.  In  chapter  vi  we  considered  the  case  where 
a  response  may  follow  the  sudden  increase  in  the  intensity  of  a  stim- 
ulus previously  maintained  at  a  constant  level.  We  turn  now  to  a 
process  of  more  complex  form  in  which  each  of  a  group  of  stimuli 
differing  only  in  intensity  elicits  a  distinct  response.  This  involves 
absolute  discrimination  though  a  similar  mechanism  may  be  at  work 
when  relative  discrimination  occurs.  The  important  point  to  notice  is 
that  an  increase  in  the  intensity  of  the  stimulus  does  not  merely 
change  the  strength  of  the  response  or  bring  into  activity  additional 
elements,  but  may  so  alter  the  response  that  none  of  the  elements 
involved  before  the  change  is  included  among  those  active  after  the 
change. 

Consider  the  net  of  Figure  5  which  is  a  parallel  interconnected 
structure  containing  also  circuits  (Householder,  1939b).  Let  an  affer- 


PSYCHOPHYSICAL  DISCRIMINATION  67 


ne 

Figure  5 

ent  neuron  form  synapses  with  a  number  of  neurons  ne  whose  thresh- 
olds differ  but  which  are  otherwise  equivalent.  Let  all  those  which 
have  the  same  threshold  be  brought  together  to  act  on  a  single  neuron. 
Only  two  of  an  indefinite  number  of  final  neurons  are  shown  in  the 
Figure.  Thus  all  the  neurons  ne  having  the  threshold  hk  are  brought 
together  to  act  on  a  single  neuron  ne3 .  Let  f(h)dhhe  the  number  of 
these  neurons  having  thresholds  between  h  and  h  +  dh .  Consider- 
ing only  the  stationary-state  activity,  and  using  the  linear  relation- 
ship of  equation  (5),  chapter  i,  with  ft  =  1 ,  we  may  write 

s(S,h)  =  (S-h)  f(h)  (18) 

as  the  value  of  the  excitation  at  st ,  the  terminus  for  the  neurons  ne 
with  the  threshold  h  . 

Let  equivalent  inhibitory  neurons  of  negligible  thresholds  origi- 
nate at  each  synapse  s»  and  terminate  at  every  other  synapse  s,  with- 
out duplication.  If  a(S,h)  is  the  value  of  a  at  the  synapse  Si  at  which 
the  neurons  of  threshold  h  terminate,  then  the  value  of  j(S)  at  this 
synapse,  due  to  the  activity  of  the  neurons  ni  terminating  there,  is 

j(S)=Xfo(S,h)  f(h)dh,  (19) 

X  being  a  constant  measuring  the  activity  of  the  inhibitory  neurons. 
Thus 

<j(S,h)=s(S,h)  -  X  f  a(S,h)  f(h)dh.  (20) 

This  is  a  special  form  of  equation  (15),  chapter  iii,  where  the 
/J  and  x'  of  the  former  equation  are  here  replaced  by  X  and  h ,  respec- 
tively, N(x',  x)  becomes  f(h),  and  the  former  h  is  negligible. 

If  f(h)  is  continuous  and  /(0)  =  0,  e(S,h)  vanishes  at  h  =  0 
and  h  =  S .  Hence  e(S,h)  has  at  least  one  maximum  in  this  range. 
Assuming  it  to  have  only  one,  we  see  that  e(S,h)  will  equal  j(S)  at 
only  two  values  of  h ,  hx  and  h2 ,  and  in  this  range  e(S,h)  >  j(S)  so 
that  <r  (S,h)  >  0  .  That  is,  the  neurons  ne3  corresponding  to  thresh- 
olds in  the  range  hx  to  h2  are  excited.  Thus  we  may  write 


68  MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


j(S)  =  X  fh*<r(S,h)  dh 

Jht 

so  that,  as  in  chapter  iii, 

X  f^e(Sfh)  dh 

1+X(K-K) 


(21) 


(22) 


Thus  from  e(S,M  =  e(S,h2)  =  j(S),  we  may  determine  h^S)  and 

MS). 

Define  a  Weber  ratio  S(S)  by  the  equation 


MS)  =MS  +  S<$). 


(23) 


The  definition  implies  that  when  the  intensity  S  is  changed  to  S  +  Sd  , 
a  completely  different  set  of  neurons  is  activated,  so  that  we  may  ex- 
pect a  response  to  S  +  Sd  which  is  distinct  from  the  response  to  S  . 

In  order  to  proceed  we  must  prescribe  the  function  f(h).  The 
simplest  assumption  is  that  f(h)  is  proportional  to  h  over  a  sufficient- 
ly wide  range,  which  amounts  to  setting  f(h)  =  In  if  any  constant 


X    White 


Brodhun 


Konlg 


O    670     \X  (Jl 


A    White 

a    670    a  U 


5 


10 


log 


10 


Figure  6. — Comparison  of  theory  with  experiment:  intensity-discrimination 
at  varying  intensity-levels  for  visual,  auditory,  and  tactile  sensations.  Solid 
curves,  theoretical  prediction  by  equations  (25) -(28)  ;  points  and  dotted  curve, 
experimental.  Visual  data  from  Konig  and  Brodhun,  1888  and  1889.  Abscissa, 
intensity  (on  logarithmic  scale)  of  stimulus;  ordinate,  ratio  of  just-discriminable 
difference  to  total  intensity. 


PSYCHOPHYSICAL  DISCRIMINATION 


69 


multiplier  is  absorbed  in  the  constant  X  .  The  graph  of  e(S,h)  is  an 
inverted  parabola  with  a  maximum  at  5/2 ,  and  thus  th  and  h*  are 
equidistant  from  S/2  .  Define  the  "relative  interval,"  x  ,  by 


Then 


Sx  =  S-2h1  =  2ha-S. 
e(S,h1)=e{S,h2)  =SH1  -  x2)/4  =  j(S) 


(24) 
(25) 


Introducing  s  =  h(S—h)  in  equation  (22)  and  the  result  into  equa- 
tion (25),  we  obtain 


where  u  is  defined  by 


ux3  +  x2  —  1  =  0 


u  =  2XS/Z 


(26) 


(27) 


and  is  thus  proportional  to  the  intensity.  The  value  of  x  for  S  +  Sd 
is  given  by 

x(S  +  Sd)  =  [6-  x(S)]/(d  +  1),  (28) 


7  J 


Auditory  data;  Rlesz 

1000  cycles  per  sec. 

70   "    "   " 


h 
5 


2  • 


1  - 


Figure  7. — Comparison  of  theory  with  experiment:  intensity-discrimination 
at  varying  intensity-levels  for  visual,  auditory,  and  tactile  sensations.  Solid 
curves,  theoretical  predictions  by  equations  (25) -(28) ;  points  and  dotted  curves 
experimental.  Auditory  data  from  Reisz,  1938.  Abscissa,  intensity  (  on  logarith- 
mic scale)  of  stimulus;  ordinate,  ratio  of  just-discriminable  difference  to  total 
intensity. 


70    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


•7 


.6  - 


.1 


Tactile  data:  Macdonald  and  Robertson 


o   <33< 


1.5 


2.0 


2.5 


log! 


Figure  8. — Comparison  of  theory  with  experiment:  intensity-discrimination 
at  varying  intensity-levels  for  visual,  auditory,  and  tactile  sensations.  Solid 
curves,  theoretical  predictions  by  equations  (25) -(28)  ;  points  and  dotted  curves 
experimental.  Tactile  data  from  Macdonald  and  Robertson,  1930.  Abscissa,  in- 
tensity (on  logarithmic  scale)  of  stimulus;  ordinate,  ratio  of  just-discriminable 
difference  of  total  intensity. 

and  evidently  u(S  +  S6)  =w(l  +  6).   Writing  equation  (26)  with 
S  replaced  by  S  +  S6  ,  and  introducing  (28)  we  obtain 


ILX" 


(36u  +  1)  x?  +  (S62u  +  26)  x-  (u63  -  26  -  1)  =  0 .  (29) 


By  eliminating  x  between  equations  (29)  and  (26),  we  obtain  6(u), 
the  desired  relation  between  the  Weber  ratio  6  and  the  intensity  of 
the  stimulus.  The  result  can  be  expressed  in  the  form  (Householder, 
1942c) 

S  =  pu-a^  (30) 

where  a  is  very  nearly  constant  and  lies  between  1/2  and  1/3  . 

In  Figures  6  to  8  are  shown  a  comparison  between  theory  and  ex- 
periment for  visual,  auditory,  and  tactile  data  (Householder,  1939). 
The  abscissa  log  u  is  used  for  convenience;  u  is  proportional  to  the 
stimulus  S .  It  should  be  emphasized  that  there  is  but  the  one  para- 
meter involved  in  each  curve. 

In  Figure  9  is  shown  the  theoretical  and  experimental  results  for 
the  case  of  visual  discrimination  of  lengths  (Householder,  1940).  In 
order  to  make  the  comparison  it  is  necessary  to  assume  only  a  propor- 
tionality between  the  length  of  the  line  and  the  value  of  e  resulting 


PSYCHOPHYSICAL  DISCRIMINATION 


71 


from  the  movement  of  the  eye  from  one  end  of  the  line  to  the  other. 
The  effect  of  binocular  cues  also  has  been  considered  (Householder, 
1940) . 

In  the  case  of  discrimination  of  weights  (Householder,  1940),  it 
is  found  that  it  is  unsatisfactory  to  assume  a  simple  proportionality 
between  the  weight  W  and  the  value  of  e  at  the  end  of  the  afferent 
neuron.  If  a  logarithmic  relation  is  assumed  together  with  a  particu- 
lar distribution  of  the  thresholds,  it  is  then  possible  to  determine  a  re- 
lationship between  d  and  W  in  terms  of  two  parameters.  Experimental 
results  from  discrimination  of  weights  are  shown  as  points  in  Figure 
10,  the  weights  being  placed  in  the  subject's  hand.  The  theoretical 
predictions  are  shown  as  curves.  The  weight  of  the  hand  is  a  third 
parameter  in  this  case  and  had  to  be  estimated  indirectly  from  the 
data  in  each  case  since  the  direct  measurement  was  not  included.  The 
estimated  values  were  400  gms  for  each  hand  of  the  male  subjects  and 
350  gms  for  each  hand  of  the  female  subjects— values  which  are  quite 
plausible. 


ae 


.010 

00  =  .005 
log  K  =  8.18042 

• 

008 

OC =.34277 

.006 

•  / 

004 

>/• 

002 

• 

.000 

1     i    i 

-   .  i 

■ 

, 

.005 


.01 


.02 


.05 

e 


.10 


.20 


.50 


Figure  9. — Comparison  of  theory  with  experiment:  discrimination  of  lengths 
of  line-segments,  visually  perceived.  Curve,  theoretical,  based  on  equations  (25)- 
(28);  points,  experimental  (Chodin,  1877).  Abscissa,  visual  angle  of  shorter 
segment;  ordinate,  just-discriminable  angular  difference. 


72   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

Our  discussion  has  dealt  mainly  with  two  mechanisms,  and  it  is 
important  to  distinguish  the  experimental  situations  to  which  they 
were  applied.  The  first  mechanism  was  applied  to  relative  discrimina- 
tion between  stimuli  simultaneously  presented,  and  the  distribution  of 
"correct"  and  "wrong"  judgments  was  predicted.  The  second  was  ap- 
plied to  absolute  discrimination,  each  stimulus  being  presented  alone, 
and  a  Weber  ratio  was  deduced.  A  third  mechanism  was  first  intro- 
duced in  chapter  vi,  and  could  be  applied  to  sensitivity  following  adap- 
tation. Still  a  fourth  was  suggested  in  chapter  iii,  but  only  as  illus- 
trating the  possibility  of  a  mechanism  failing  to  transmit  under  in- 
tense stimulation.  No  application  to  concrete  quantitative  evidence 
was  made  of  this.  The  last  mechanism  discussed  here  will  be  extended 
in  chapter  xii  to  provide  a  mechanism  for  the  discrimination  of  colors. 


TABLE  I 
LIFTED  WEIGHTS 


EXPERIMENTAL 

THEORETICAL 

s 

r, 

Pe 

Pi 

PB 

Pe 

Pi 

Grams 

P  r 

o  p  o  r  t  i 

o  n  s 

84 

0.012 

0.027 

0.961 

0.004 

0.021 

0.975 

.020 

.980 

.010 

.990 

88 

.021 

.082 

.897 

.025 

.077 

.898 

(  .053) 

.947 

.053 

.947 

92 

.096 

(.181) 

.723 

.103 

(.181) 

.716 

.185 

.815 

.179 

.821 

96 

.275 

.266 

.459 

.284 

.265 

.451 

.420 

.580 

.409 

.591 

LOO 

.502 

.267 

.231 

.551 

.250 

.199 

(  .683) 

.317 

(  .683) 

.317 

L04 

.842 

.103 

.055 

.796 

.140 

.064 

.920 

.080 

.880 

.120 

L08 

.915 

0.065 

.020 

.932 

0.054 

.014 

0.963 

0.037 

0.966 

0.034 

PSYCHOPHYSICAL  DISCRIMINATION 


73 


WEIGHT  DIFFERENCE  AW  IN  GRAMS 


Figure  10. — Comparison  of  theory  with  experiment:  discrimination  of  lifted 
weights  for  three  observers  Qi,%,  and  03 ,  left  hand  L,  right  hand  R,  and  aver- 
age for  both  hands  M.  Curve,  theoretical,  based  on  equations  (25) -(28) ;  points 
experimental  (Holway,  Smith,  and  Zigler,  1937).  Abscissa,  lesser  weight  in 
grams,  ordinate,  just-discriminable  difference  in  grams. 


X 

INTERCONNECTED  CHAINS:  MULTIDIMENSIONAL 
PSYCHOPHYSICAL  ANALYSIS 

In  some  cases,  for  example  in  the  making  of  aesthetic  judgments, 
the  stimulus-objects  are  complex  and  may  provide  stimulation  in  any 
number  of  distinct  modes.  Then  if  a  statement  of  preference  is  called 
for — one  of  two  incompatible  responses — the  sequence  of  stimulus  and 
response  may  be  regarded  as  a  discriminal  sequence  as  defined  in  the 
preceding  chapter  provided  we  regard  each  of  the  two  stimuli  as  a 
resultant  of  the  components  in  the  various  modes.  The  composition, 
we  may  suppose,  is  effected  in  some  way  within  the  organism  through 
the  concurrence  at  some  point  of  the  afferent  chains  leading  from  the 
several  receptors. 

The  simplest  scheme  for  representing  the  neural  processes  which 
mediate  a  discriminal  sequence  of  this  type  is  the  following.  Suppose 
that  each  complex  stimulus-object,  CP(p  =  1 ,  2),  provides  stimuli  of 
intensities  CP;(i  =  1  ,  ■■•  ,  n)  in  the  n  modalities,  and  that  these  stim- 
uli send  impulses  independently  along  discrete  afferent  chains  to  the 
synapse  Si  where  they  occasion  the  production  of  o-  =  SPi ,  the  SPi  for 
each  p  combining  additively  to  yield  the  SP  hitherto  employed: 


sP  =  zs 


Pi 


Each  Spi  is  then  some  function  of  Cpi  alone;  still  regarding  only  the 
near-threshold  range  we  may  take  these  functions  to  be  all  linear, 

SP  =  ^LiCpi-M.  (1) 

We  may  now  use  either  of  the  procedures  introduced  in  the  previous 
chapter,  with  a  =  /5 ,  according  to  the  choice  of  location  for  the  ran- 
dom element.  In  either  case  there  are  but  three,  or,  with  more  special 
assumptions,  only  two,  functions  P .  The  functions  remain,  however, 
functions  of  the  SP ;  for  any  pair  of  stimuli  d  and  C2 ,  which  provide 
some  (unknown,  since  the  L,  are  unknown)  <SX  and  S2 ,  it  is  possible 
to  determine  experimentally  the  relative  frequencies  P ,  and  those 
stimulus-pairs  which  yield  the  same  values  of  the  P's  will  be  those 
which  yield  a  fixed  difference 

Si-s;=2£i(Cn-&i).  (2) 

The  same  result  follows  if  we  assume  crossing  inhibition  connecting 
also  the  pair  of  afferents  affected  by  the  two  stimuli  for  each  modal- 
ity.   This  would  justify  extending  the  assumption  of  linearity  to  a 

74 


MULTIDIMENSIONAL  PSYCHOPHYSICAL  ANALYSIS  75 

much  greater  range,  since  only  the  positive  difference  \C1i  —  C2i[  af- 
fects the  subsequent  members  of  any  chain. 

It  is  natural  to  identify  this  scale  of  S-values  with  Thurstone's 
"psychological  scale"  (Thurstone,  1927;  cf.  Guilford,  1936).  The  psy- 
chological scale  is  introduced  quite  abstractly  in  psychophysics,  the 
assumptions  being  that  each  stimulus-object  produces  a  "discriminal 
response"  which  can  be  measured  on  this  scale,  and  that  repeated 
presentation  of  the  same  object  leads  to  varying  responses  as  so 
measured,  the  distribution  being  normal  on  this  scale.  By  following 
a  well-defined  procedure  the  empirical  frequencies  of  the  judgments 
can  be  utilized  for  determining  the  modal  S  associated  with  each  stim- 
ulus-object, the  determination  being  unique  up  to  a  linear  transfor- 
mation. 

In  our  terms,  for  any  stimulus-object  Cp  ,  the  modal  response 
(on  the  part  of  the  afferent  chain,  at  the  synapse  sp)  is  Sp  ,  while 
Sp  +  CP  is  the  particular  response  on  a  given  presentation;  if  Cp  is 
taken  to  be  normally  distributed  the  identification  of  our  S-scale  with 
that  of  psychophysics  is  immediate;  otherwise  some  scale-transfor- 
mation is  required.  In  either  case  the  methods  of  psychophysics  can 
be  utilized  to  determine  the  SP  ,  for  each  Cp  ,  at  least  up  to  a  linear 
transformation.  Then  if,  further,  the  CPi  are  directly  measurable,  the 
quantities  Lt  can  be  determined  up  to  a  common  multiplicative  factor. 
These  quantities  furnish  measures  of  the  relative  contributions  of  the 
separate  modes  of  stimulation  to  the  judgment  as  a  whole.  We  note, 
incidentally,  the  possible  application  of  factor  analysis  with  a  large 
population  of  subjects  (Thurstone,  1937). 

Physical  measurement  of  the  CPt  is  possible  but  rarely  and  in  the 
least  interesting  of  the  cases,  and  psychophysicists  have  endeavored 
to  obtain  from  the  empirical  frequencies  an  insight  into  the  number 
of  distinct  modes  of  stimulation  of  the  organism  by  the  complex 
stimulus-objects  of  a  given  class.  It  is  clear  that  from  judgments  of 
preference  alone  no  such  information  is  to  be  had,  for  by  the  condi- 
tions of  the  experiment  the  subject  is  required  to  make  a  one-dimen- 
sional ordering  of  the  objects.  But  by  a  slightly  revised  experimental 
procedure,  interpreted  in  terms  of  a  suitable  neural  mechanism,  it 
is  possible  to  obtain  the  data  necessary  for  such  a  multidimensional 
analysis  (cf.  Householder  and  Young,  1940;  Young  and  Householder, 
1941). 

In  this  procedure  each  stimulus-object  is  replaced  by  a  pair,  and 
the  subject  is  now  asked  which  of  two  given  pairs  is  most  unlike.  The 
formal  psychophysical  analysis  required  for  determining  the  S  corre- 
sponding to  each  pair  is  identical  with  that  required  for  determining 
the  S  for  each  object  in  the  previous  experiment;  only  the  interpre- 


76   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

tation  is  different,  for  the  S  now  measures,  on  the  psychological  scale, 
the  extent  to  which  the  two  members  of  the  pair  are  different.  If  the 
two  members  of  any  pair  are  identical  the  associated  S  must  be  zero, 
so  that  the  additive  term  is  determinate  and  the  S's  can  in  this  case 
be  determined  uniquely  up  to  a  common  multiplicative  factor. 

The  ordering  of  the  pairs  along  the  S-scale  is,  however,  only  an 
intermediate  step  since  we  wish  to  associate  the  individual  objects 
with  points  of  a  metric  space  of  sufficiently  many  dimensions  in  such 
a  way  that  the  S  for  any  pair  is  the  distance  between  the  points 
which  represent  the  members  of  the  pair.  But  the  solution  of  this 
problem  depends  upon  the  character  of  the  space's  metric,  and  the 
metric  in  turn  is  a  property  of  the  mediating  neural  mechanism. 

Suppose,  then,  that  the  objects  Ap  and  BP  constitute  the  pair  CP 
and  that  the  afferents  for  the  i-th  modality  affected  by  AP  and  BP 
are  connected  by  crossing  inhibition.  Then,  if  the  thresholds  are  low, 
beyond  the  locus  of  the  crossing  the  corresponding  <r  is  proportional  to 

Cpi=\Api-Bpi\  (P  =  l,2)  (3) 

along  the  afferent  from  Api  or  Bpi ,  whichever  is  the  greater,  while  it 
is  negative  along  the  other.  If  these  two  paths  join  at  some  subse- 
quent synapse,  then  succeeding  neurons  of  this  chain  are  stimulated 
in  amounts  proportional  to  CPi  as  given  by  equation  (3).  From  here 
the  mechanism  is  just  like  the  one  previously  discussed.  Hence  for 
any  pair  (A  ,  B)  in  place  of  equation  (1)  we  have 

S  =  2Li\Ai-Bi\  . 

If  the  quantities  At  and  Bl  are  physically  measurable,  or  if  they  are 
measurable  by  psychological  methods  independent  of  the  method  now 
being  outlined,  e.g.,  in  terms  of  J.  N.  D/s,  the  quantities  Li  have  sig- 
nifiance  and  it  is  an  empirical  problem  to  determine  whether  or  not  a 
set  Li  exists  satisfying  all  equations  of  this  type.  If  the  Ai  and  Bi 
are  not  so  measurable  they  are  precisely  the  quantities  to  be  deter- 
mined from  this  procedure,  and  we  may  introduce  units  so  chosen  that 
every  L%  =  1 .  In  this  case,  which  we  assume  hereafter, 

S  =  2\Ai-Bi\.  (4) 

If  it  happens  that  for  any  three  stimulus-objects,  the  S  for  one 
pair  is  equal  to  the  sum  of  those  for  the  other  two,  then  this  experi- 
ment provides  no  data  for  a  multidimensional  analysis.  This  does 
not  imply,  however,  that  only  a  single  modality  is  involved.  If,  for 
any  three  objects,  the  S  of  any  pair  exceeds  the  sum  of  those  for  the 
other  two,  the  equations  (4)  are  inconsistent,  and  further  analysis  is 
impossible  on  the  basis  of  the  mechanism  here  proposed.  Passing  over 


MULTIDIMENSIONAL  PSYCHOPHYSICAL  ANALYSIS  77 

these  cases  to  which  the  present  method  does  not  apply,  we  suppose, 
therefore,  that  for  some  set  of  these  objects,  say  A(0),  A(x),  and  A(2), 
the  S  of  each  pair  is  exceeded  by  the  sum  of  the  other  two,  and  we 
attempt  a  two-dimensional  representation.  It  is  evident  that  the  val- 
ues of  the  various  S's  alone  can  determine  the  various  A»  at  most  up 
to  an  additive  constant — the  zero  for  each  modality  is  arbitrary  un- 
less prescribed  by  considerations  irrelevant  to  the  formal  experi- 
mental procedure.  Such  prescription,  if  available,  may  be  observed 
later  by  an  appropriate  adjustment;  for  the  present  choose  A<0)  as 
one  reference-point  and  assume  that  each  Ai(0)  =  0 .  If  S{pq)  repre- 
sents the  S  corresponding-  to  the  pair  (A(p),  A(9))»  we  may  suppose, 
further,  that 

£(02)    >   gCL2> 

relabeling  the  objects  if  necessary.  For  determining  the  four  quan- 
tities A(i)j,  only  three  equations  are  available.  Hence  we  may  make 
the  assumption 

r±l  -f±2  ) 

subject  to  possible  later  revision.  Finally,  we  may  suppose  that 
Ax{2)  <  A2(2),  since  we  can  only  separate  but  not  identify  the  two 
modes,  and  we  have 

Ax<2>  =  [£<°2>  -  £<12>]/2  ,  A2<2>  =  [S<02)  +  S<12>]/2  , 
A^  =  A^  =  S(01> /2  . 

Now  consider  any  object  A(3).  If  either 

£(03)   =£(01)    _|_   £(13)   —£(02)    _|_   £(23) 

or  else 

£(03)   =  £(01)    _   £(13)   —  £(02)    _   £(23) 

then 

£(03)—  ^(3)    +   42<3) 

and  the  quantities  on  the  right  are  indeterminate.  If  neither,  there 
are  at  least  two  independent  equations  involving  Aa(3)  and  A2(3).  If 
there  are  three,  a  two-dimensional  representation  is  impossible,  but 
if  for  every  fourth  object  A(3)  there  are  at  most  two  new  equations, 
two  dimensions  are  sufficient. 

Thus,  apart  from  the  arbitrariness  indicated,  with  a  sufficiently 
large  number  of  stimulus-objects  A  all  the  Ai  can  be  determined.  It 
is  perhaps  clear  enough  from  the  above  how  one  must  proceed  when 
more  than  two  dimensions  are  required.  The  quantities  S  may  be 
regarded  as  distances  in  the  representative  space,  and  the  space  is 
metric  but  not  Euclidean.  The  assumption  of  linearity  imposed  upon 
the  mechanism  is  not  highly  restrictive,  in  principle,  since,  if  the 


78   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

stimuli  are  all  low  in  intensity,  the  linear  expressions  may  be  re- 
garded as  the  first  terms  of  Taylor  expansions.  Additional  cross- 
connections  could  be  introduced  into  the  mechanism,  either  inhibitory 
or  excitatory  in  character,  but  the  essential  features  would  not  be 
changed  nor  the  metric  of  the  space  greatly  modified.  In  particular, 
it  is  important  to  note,  the  arbitrariness  is  inherent  in  the  nature  of 
the  experiment  and  can  only  be  removed  by  further  experiment  or 
observation  of  a  different  kind. 

In  chapters  vi  and  vii  we  considered  the  process  at  a  single  syn- 
apse terminating  one  or  more  afferent  chains,  and  related  this  to 
various  sequences  of  stimulus  and  response.  In  all  these  sequences 
there  was  but  a  single  response,  however  complex  in  form,  dependent 
upon  the  variation  of  the  single  a  ,  evoked  by  a  single,  simple  or  com- 
plex stimulus,  and  perhaps  modified,  in  degree  or  in  the  time  of  its 
occurrence,  by  other  stimuli.  In  this  and  the  preceding  chapters  we 
have  considered  two  or  more  synapses  with  as  many  afferents,  ter- 
minating afferent  chains  from  various  receptors.  The  structures  were 
associated  with  classes  of  sequences  of  stimulus  and  response  of  the 
following  sort.  Each  of  a  group  of  stimuli,  simple  or  complex,  when 
presented  in  isolation  can  evoke  a  certain  characteristic  response, 
whereas  the  concurrence  of  these  stimuli  modifies  the  separate  re- 
sponses by  enhancing,  reducing,  or  even  preventing  them.  We  have 
by  no  means  exhausted  the  possible  applications  of  the  various  struc- 
tures ;  by  varying  the  assumed  relations  among  the  parameters  an  al- 
most countless  variety  of  sequences  is  suggested.  For  example  we 
have  been  considering  only  the  case  of  crossing  inhibition  and  have 
mentioned  only  in  passing  the  possibility  of  having  crossing  excita- 
tion, a  mechanism  that  would  mediate  sequences  of  a  quite  different 
sort.  Numerous  possible  complications  are  easily  suggested.  The  re- 
sponse to  a  given  stimulus  might  be  modified,  not  directly  by  another 
external  stimulus  but  by  the  response  to  that  stimulus,  this  calling 
for  a  connection  running  from  the  second  effector  back  to  some  point 
in  the  afferent  chain  leading  to  the  first  effector.  Circuits  of  the  type 
discussed  in  chapter  iv  might  be  introduced  at  various  points  and  their 
effects  studied  and  related  to  observable  sequences. 

The  procedure  of  starting  with  the  simpler  structures  and  seek- 
ing applications  thereof  has  this  decided  advantage,  that  we  can  feel 
assured  that  the  postulated  mechanism  is  not  more  complicated  than 
necessary  for  mediating  the  adduced  sequence.  Thus  if  one  stimulus 
can  in  any  way  modify  the  response  evoked  by  the  isolated  occurrence 
of  another,  then  some  connection  must  lead  from  the  first  receptor 
to  the  effector  for  that  response,  whether  the  connection  is  direct, 
through  the  spinal  cord  only,  or  indirect,  through  the  thalamus,  or 


MULTIDIMENSIONAL  PSYCHOPHYSICAL  ANALYSIS  79 

elsewhere.  It  is  highly  unlikely  that  the  actual  mechanism  mediating 
any  of  the  stimulus-response  sequences  here  outlined  is  as  simple  as 
the  one  postulated  for  it,  but  by  comparing  the  deductions  from  the 
simpler  postulates  with  laboratory  data  we  can  take  note  of  the  devia- 
tions and  be  guided  thereby  in  our  endeavor  to  improve  the  picture. 
While  this  procedure,  from  mechanism  to  suggested  application, 
could  be  pursued  indefinitely  through  increasing  degrees  of  complex- 
ity, we  turn  instead,  in  the  following  chapters,  to  the  reverse  proce- 
dure, considering  certain  forms  of  activity  and  attempting  to  con- 
struct mechanisms  capable  of  mediating  these. 


XI 


CONDITIONING 

A  most  important  property  of  neural  circuits  is  that  their  ac- 
tivity may  continue  indefinitely  after  cessation  of  the  stimulus.  The 
possible  application  of  this  property  to  memory  is  evident,  but  to 
conditioning  it  is  much  less  so.  We  now  suggest  a  mechanism  for  ex- 
plaining conditioning  and  learning. 

Consider  first  a  few  properties  of  the  structure  of  Figure  1 


i  x 


Figure  1 

(Rashevsky,  1938) .  For  the  present  we  shall  ignore  the  presence  of  the 
dotted  neurons.  This  structure  consists  of  two  neuron-chains,  leading 
through  a  final  common  path  to  a  response  R ,  together  with  a  uni- 
lateral interconnection  and  a  simple  circuit  C  .  The  chief  character- 
istic of  conditioning  is  that  a  particular  response  R ,  normally  pro- 
duced by  the  "unconditioned"  stimulus  Su  but  not  by  the  stimulus  Sc , 
may  after  the  repeated  concurrence  of  the  stimuli  Sc  and  Su  become 
capable  of  being  evoked  by  Sc  alone.  This  may  require  one  or  more 
concurrences  of  Sc  and  Su ,  and  while  Sc  and  Su  need  not  be  presented 
at  exactly  the  same  time,  the  time  between  them  cannot  be  too  long. 
Suppose,  only  for  the  sake  of  simplicity,  that  all  the  neurons  are  of 
the  simple  excitatory  type,  and  let  <£o  represent  the  maximum  value 
of  <f>  for  any  Sc .  Then  for  the  net  of  Figure  1,  let 

80 


CONDITIONING  81 

<£<,  <  h'  <  </>0  +  e0  ,  (1) 

</>,(,  <  h"  <  </>.0  +  </>.0tt  >  (2) 

</>o«  <  ^  ,  (3) 

where  £0  is  the  value  of  the  excitation  at  sc  due  to  the  circuit  when  in 
steady-state  activity,  and  where  the  </>'s  refer  to  the  afferent  neurons. 

If  the  unconditioned  stimulus  Su  exceeds  the  threshold  of  /„  suf- 
ficiently, the  response  R  may  be  elicited.  But,  as  we  assume  that  the 
circuit  C  is  not  active  initially,  the  stimulus  Sc  cannot  produce  the 
response  R  because  4>0  <  h' .  Furthermore,  neither  Su  nor  Sc  can 
bring  C  into  activity  when  there  is  too  long  a  time  between  their  oc- 
currence. Thus  Su  alone  can  produce  R  but  Sc  cannot.  Now  suppose 
that  Su  and  Sc  are  applied  together  for  a  sufficiently  long  time.  Though 
simultaneous  presentation  is  not  a  necessary  condition,  it  will  be  con- 
sidered here  to  simplify  matters.  Because  of  condition  (2) ,  the  thresh- 
old of  the  circuit  will  be  exceeded  and  the  circuit  will  pass  over  into 
a  state  of  steady  activity.  If,  now,  a  large  enough  Sc  is  applied  alone 
for  a  long  enough  time,  the  threshold  h'  will  be  exceeded  because  of 
condition  (1).  Thus  the  response  R  may  now  be  produced  by  the 
hitherto  inadequate  stimulus  Sc  alone,  and  the  structure  exhibits  one 
of  the  principal  features  of  conditioning. 

If  we  add  now  the  inhibitory  neurons  III'  and  III,  the  resulting 
structure  will  exhibit  another  feature  important  in  the  phenomenon 
of  conditioning.  Whenever  Su  is  applied,  the  effect  of  neuron  III  is 
blocked  by  III'.  But  if  Su  is  not  applied,  then  the  continuous  or  re- 
peated application  of  Sc  may  cause  III  to  produce  enough  inhibition 
at  s'  to  block  the  action  of  Sc  if  conditioning  has  previously  taken 
place.  This  corresponds  to  the  loss  in  effectiveness  of  the  conditioned 
stimulus  which  occurs  when  it  is  applied  repeatedly  without  rein- 
forcement by  the  unconditioned  stimulus. 

If,  instead  of  a  single  circuit  C  ,  we  assume  that  there  are  a  num- 
ber of  them  having  different  thresholds,  we  should  be  able  to  show 
that  a  more  intense  Su  and  Sc  would  tend  to  produce  a  more  intense 
response.  Furthermore,  by  considering  repeated  applications  of  Su 
and  S ,  it  is  possible  to  determine  the  effect  of  the  number  of  repeti- 
tions on  the  conditioning.  By  combining  these  extensions  of  the  struc- 
ture, N.  Rashevsky  (1938,  chap,  xxv,  equation  44)  obtains  an  expres- 
sion 

eB  =  A(il-e°»)  (4) 

for  eR  ,  the  excitation  tending  to  produce  the  response  R  when  S  is  ap- 
plied as  a  function  of  the  number,  n ,  of  repetitions.  The  constant  A  in- 
creases with  the  intensity  of  the  conditioned  stimulus,  while  the  con- 


82    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

stant  a  increases  with  both  Su  and  S  and  depends  on  the  time  between 
repetitions  and  the  time  of  stimulation  at  each  repetition.  This,  too,  is 
in  qualitative  agreement  with  results  of  experiments  on  conditioning. 

This  conditioning  mechanism  requires  the  continued  activity  of 
some  circuits  and  the  objection  could  be  raised  that  the  great  sta- 
bility of  well-established  memory-patterns,  which  resist  disruption 
by  either  such  major  disturbances  as  shock  and  narcosis  or  by  the 
cumulative  effect  of  countless  minor  vicissitudes  over  a  period  of  time, 
is  inconceivable  in  terms  of  so  vulnerable  a  structure.  To  this  ob- 
jection at  least  two  replies  are  possible.  One  is  that  a  quantitative 
theory  is  useful  in  proportion  to  the  extent  of  the  phenomena  for 
which  it  can  account,  and  it  is  not  less  useful  for  failing  to  account 
for  others,  However,  the  objection  can  be  met  in  a  more  positive  way 
by  supposing  that  the  relatively  rapid  changes  permitted  by  the  above 
mechanism  lead  in  some  fashion  to  more  permanent  structural 
changes.  It  is  quite  possible  that  the  usual  intermittent  rise  and  fall 
of  e  and  j  at  a  given  synapse  would  have  no  physiological  effect  be- 
yond the  exciting  and  inhibiting  effects  which  we  have  postulated, 
whereas  the  maintenance  at  some  sufficiently  high  value  of  either  or 
both  would  cause  permanent  changes  involving,  among  other  things, 
a  modification  of  threshold  (cf.  Douglas,  1932).  The  theory  of  the 
process  of  conditioning  would  then  hold  as  outlined,  but  would  re- 
quire a  supplement  to  account  in  detail  for  the  observed  stability. 

But  regardless  of  the  mechanism,  the  facts  of  conditioning  re- 
quire some  change  in  e  with  successive  trials,  leading  to  a  change  in 
response.  The  simplest  assumption  possible  is  that  s  is  proportional 
to  the  number  of  trials,  at  least  when  the  number  of  trials  is  small, 
and  this  is  the  essential  content  of  equation  (4),  with  Aa  the  constant 


Figure  2 


CONDITIONING  83 

of  proportionality.  The  theory  could  be  developed  formally  regard- 
ing- Aa  as  a  purely  empirical  constant  with  no  definite  physiological 
significance.  But  the  model  enables  us  to  relate  this  constant  to  such 
variables  as  the  strengths  of  the  conditioned  and  the  unconditioned 
stimuli,  temporal  factors,  and  the  like,  and  so  provides  the  possibility 
of  relating  a  larger  group  of  variables  in  a  single  formulation. 

For  interpreting  some  experimental  results  in  these  terms,  we 
consider  the  net  shown  in  Figure  2  (Landahl,  1941).  Let  a  stimulus 
Sc  normally  produce  a  response  Rc ,  and  let  a  pleasant  response  R1 
always  follow  Rc  in  the  experimental  situation.  Let  Sw  normally  pro- 
duce Rw ,  which,  in  the  experimental  situation,  leads  to  an  unpleas- 
ant stimulus,  less  pleasant  stimulus  or  to  an  equally  pleasant  stimulus 
but  after  a  longer  time.  Let  the  circuits  M  and  C  each  represent  a 
large  group  of  circuits  of  different  thresholds.  Let  the  part  of  the 
structure  composed  of  neurons  III,  III',  IV,  IV  be  equivalent  to  the 
corresponding  part  of  Figure  1  of  chapter  ix.  We  shall  consider  only 
simultaneous  presentation  of  SK  and  Sc .  On  the  first  trial,  neither 
C  nor  C  can  become  active,  and  thus  we  have  acting  at  sc  a  quantity 
e,oc;  and  similarly  at  slc  a  quantity  e,ow  .  Then,  if  one  of  the  two  re- 
sponses must  be  made,  the  probability  Pc  of  the  response  Sc  may  be 
given  by  the  approximate  equation  (8)  of  chapter  ix,  with  £!  —  e2 
replaced  by  £oc  —  £ow  • 

After  Sc  and  Sw  have  been  presented  together  n  times,  the  re- 
sponse Rc  will  have  been  made,  say,  c  times  and  the  response  Rw ,  w 
times.  We  shall  refer  to  n  as  the  number  of  trials,  c  as  the  number 
of  correct  responses,  and  w  as  the  number  of  wrong  responses.  Then 
Pc ,  the  probabilty  of  a  correct  response,  may  be  identified  with  the 
proportion  of  correct  responses,  so  that  approximately 

Pc  =  dc/dn;  (5) 

and  similarly 

Pw  =  dw/dn ,  (6) 

Thus 

Pc  +  Pw  =  l,     c  +  w  =  n.  (7) 

When  a  stimulus  Sc  is  presented,  a  certain  group  of  the  circuits 
M  are  brought  into  activity.  Then,  each  time  there  is  a  response  Rt , 
conditioning  may  take  place  in  some  circuits  of  C  and  the  amount, 
as  measured  by  the  increase  in  the  excitatory  factor  Aec  at  sc ,  will  not 
be  dependent  upon  the  time  tc  between  the  presentation  of  Sc  and 
response  Ri .  But,  if  the  circuits  M  are  acted  upon  by  inhibitory  neu- 
rons from  various  external  sources,  or  if  the  circuits  are  replaced  by 
single  neurons,  the  activity  will  decay  with  the  time,  tc ,  roughly  ex- 
ponentially.    Thus,  we  may  obtain  an  expression  for  Aec  similar  to 


84   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

equation  (4).  To  a  first  approximation,  equation  (4)  becomes  in  this 
case  Aec  =  ncb  (Sc ,  tc)  where  b  =  Aa  depends  on  Sc  and  tc ,  and  where 
c  is  the  number  of  repetitions  of  Sc  and  Ri  together  and  thus  replaces 
n  .  If  Sc  and  tc  are  constant,  the  total  e  at  sc  is  given  by 

ec  =  s0c  +  be  .  (8) 

We  can  obtain  a  similar  expression  for  ew  at  sw  .  This  is  the  case  when 
the  final  response  is  pleasant,  but  if  the  final  response  is  unpleasant 
we  should  expect  the  effect  of  the  conditioning  to  be  the  opposite. 
That  is,  the  centers  C  are  such  that  they  have  inhibitory  fibers  lead- 
ing to  sw  .  Then  the  coefficient  corresponding  to  b  will  be  some  nega- 
tive quantity  —  /S .   Thus 

ew  =  eow  —  pw,  (9) 

as  w  is  the  number  of  repetitions  of  wrong  response  leading  to  R2 . 

Let  us  apply  these  results  to  the  particular  experimental  situation 
which  arises  when  Lashley's  jumping-apparatus  is  used.  Here  an  ani- 
mal is  forced  to  jump  toward  either  of  two  stimuli.  Choice  of  one 
leads  to  reward,  choice  of  the  other  may  lead  to  punishment.  For 
simplicity,  we  assume  that  the  times  tc  and  tw ,  respectively,  from 
presentation  of  the  stimuli  to  the  reward  and  punishment,  are  con- 
stants. If  we  then  introduce  equations  (8)  and  (9)  into  equation  (8) 
of  chapter  ix,  and  eliminate  Pw  and  c  by  means  of  equations  (6)  and 
(7),  we  obtain  a  differential  equation  in  w  and  n .  From  this,  with 
the  initial  condition  w  =  0  for  n  =  0  ,  we  obtain 

1  2bekle°c~e°w) 

w  — fog QO) 

k(b  -  ft)     &  2&e*(e"-£°«>  -  (b  -  /S)  (1  -  e-kbn) 

for  ft^jS.  For  6  =  >5 ,  the  result  is  a  rising  curve  which  approaches  a 
limit  exponentially.  In  terms  of  the  mechanism,  we  may  consider  the 
experiment  as  requiring  a  discrimination  between  two  stimuli  whose 
values,  in  effect,  change  in  successive  trials.  The  correct  stimulus  be- 
comes effectively  larger  due  to  the  conditioning  while  the  wrong  de- 
creases. Thus,  the  probability  of  a  wrong  response  diminishes. 

In  Figure  3  is  shown  a  comparison  between  the  theory  and  the 
experimental  data  by  H.  Gulliksen  (1934).  The  lower  and  upper 
curves  were  obtained  respectively  by  setting  eoc  —  eow  —  0  ,  kb  =  .0121 , 
ft  =  0  and  k(eoc  —  cow)  =  -.46  ,  kb  =  .0229,  §  =  0  .  Besides  giving  a 
quantitative  relation  between  w  and  n,  our  considerations  actually 
give  a  great  deal  more.  From  the  results  of  the  preceding  paragraphs, 
we  can  obtain  a  function  b  (Sc ,  Ri ,  tc),  that  is,  b  is  a  function  of  the 
intensity  of  the  stimulus,  the  strength  of  the  reward,  and  the  time  tc . 


CONDITIONING 


85 


DATA   BY  HGULLIKSEN 


■  00  200  300 

NUMBER  OF   TRIALS  fl  — 

Figure  3. — Comparison  of  theory  with  experiment:  simple  learning.  Curves, 
theoretical  from  equation  (10);  points  experimental  (Gulliksen,  1934).  Abscissa, 
number  of  trials;  ordinate,  number  of  errors. 


Similarly,  one  can  obtain  fl(Sw ,  R? ,  tw).  Thus  we  should  be  able  to 
make  predictions  for  data  as  in  Figure  3,  but  for  various  strengths 
of  reward  or  punishment  and  for  other  variables.  In  this  way  a  con- 
siderable amount  of  data  could  be  brought  into  a  single  formulation 
and  the  prediction  tested  by  experiment. 

By  considering  various  modifications  of  the  experimental  situa- 
tion consistent  with  the  restrictions  imposed,  we  can  derive  other  re- 
lations which  could  be  checked  by  experiment  (Landahl,  1941).  If 
the  responses  R%  and  R2  are  made  identical  but  tc  ¥^  tw ,  we  have  an 
analogy  to  a  situation  in  which  there  are  two  paths  to  a  goal-response 
requiring  different  times,  tc  and  tw ,  to  traverse.  We  would  refer  to 
the  shorter  path  as  correct,  and  thus  tc  <  tw .  For  constant  Sc  and 
Sw ,  the  coefficient  b  will  be  a  function  of  tc  or  tu,  only.  If  Sc  and  Sw 
are  not  too  different,  ec  will  equal  eoc  +  b(tc)c  and  sw  will  equal 
Sow  +  b  (tw)w  since  the  final  response  is  the  same.  But,  as  b  decreases 
with  t,  b(tc)  >  b(tw).  Thus,  at  least  when  Sc  =  Sw  initially,  the 
probability  of  the  wrong  response  will  decrease  towards  zero  since, 
for  small  c  and  w  ,  c  =  w  ,  sc  —  e„  =  6  (tc)c  —  b(tw)w  >  zero.  Thus, 
we  could  determine  the  number  of  errors  as  a  function  of  the  number 
of  trials  for  various  t c  and  tw  .  If  eoc  <  s„w  ,  the  correct  response  may 
never  be  learned. 

Elimination  of  a  blind  alley  can  be  accounted  for  since,  the  cor- 
rect path  being  entered  last,  the  time  between  Sc  and  the  reward  is 
less  than  the  time  between  Sw  and  the  reward.  Hence  on  later  trials 
there  is  a  tendency  to  turn  away  from  the  wrong  stimulus.   An  equa- 


86   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

tion  for  the  number  of  errors  as  a  function  of  the  number  of  trials 
has  been  obtained  on  this  basis  and  studied  in  relation  to  such  para- 
meters as  strengths  of  reward  and  punishment,  length  of  the  alley, 
and  distance  (time)  from  blind  to  goal  (Landahl,  1941).  One  finds 
that  generally  fewer  errors  will  be  required  to  eliminate  a  blind  alley 
if  it  is  close  to  the  goal.  The  dependence  on  the  length  of  the  alley  of 
the  number  of  errors  required  to  eliminate  the  blind  is  found  to  be 
fairly  complex.  According  to  the  strength  of  the  reward,  we  find  that 
a  blind  far  from  the  goal  will  be  eliminated  with  more  difficulty  the 
longer  it  is,  while  if  it  is  near  the  goal  it  will  be  eliminated  more 
readily  if  it  is  long.  What  we  wish  particularly  to  emphasize  is  that 
from  relatively  simple  structure  can  be  deduced  fairly  complex  ac- 
tivity. 

It  is  possible  to  generalize  the  mechanism  to  include  a  choice  from 
among  any  number  N  of  stimuli  by  constructing  a  net  similar  to  that 
of  Figure  2,  but  with  N  afferents  and  N(N  —  1)  crossing  inhibitory 
neurons  (cf.  chap.  iii).  Suppose  that  out  of  the  N  stimuli  there  is 
but  one  correct  stimulus  Sc .  Hence,  instead  of  considering  the  indi- 
vidual wrong  responses,  we  may  consider  their  average  effect.   Thus 

if  sc  is  the  net  value  at  sc  due  to  the  correct  response  and  if  sw  is  the 
average  value  of  all  the  e«/s  we  may  write  [Landahl,  1941,  equation 
(9)] 

NP 

\og—-^-  +  k(ec-8w)=0  (11) 

N  —  1 


'C 


in  place  of  equation   (8)   of  chapter  ix.    We  note  that  for  ec 

Pw  =  (N  —  1)/N  as  would  be  expected  by  chance,  while  for  large 

ec  —  sw ,  Pw  tends  toward  zero. 

In  the  experimental  situation,  let  a  stimulus  S'»  (i  s=  1 ,  2  ,  •  •  • ,  M) 
accompany  a  group  of  stimuli  Sj  (j '=  1 ,  2  ,  •  •  • ,  N)  of  equal  intensity, 
one  and  only  one  of  which  will  elicit  its  response.  Among  the  stimuli 
S)  is  a  stimulus  Sic ,  the  "correct"  stimulus  corresponding  to  S'i , 
which  when  chosen  results  in  a  reward ;  response  to  any  other  stimu- 
lus Sj  when  accompanied  by  S\  results  in  punishment,  or  at  least  no  re- 
ward. The  number  N  may  be  referred  to  as  the  number  of  possible 
choices,  while  the  number  M  is  the  number  of  associations  to  be 
learned  in  the  experiment.  After  a  wrong  response  is  made,  the  ex- 
perimenter may  choose  to  assist  (prompt)  the  subject  in  making  the 
correct  response  or  he  may  not.  He  may  do  so  each  time,  not  at  all 
or,  in  general,  some  fraction,  1  —  /  ,  of  the  times.  Thus  /  is  a  variable 
under  the  control  of  the  experimenter  just  as  are  M  and  N  .  We  shall 
assume  that  throughout  any  particular  experiment  M ,  N ,  and  /  are 


CONDITIONING  87 

not  changed.   Then 

£c  =  £oc  +  bc  +  b(l-f)w,  (12) 

since  conditioning  improves  with  each  correct  response  as  well  as  with 
a  fraction  (1  —  /)  of  the  wrong  responses.  The  prompted  correct 
responses  are  not  counted  in  c  so  that  we  do  not  change  the  relation 
n  =  c  +  w .  At  each  wrong  choice,  a  quantity  p  is  subtracted  from 
£ow  .  This  contributes  only  p/  (N  —  1)  to  the  average.  Thus 

**  =  8a»-fi/{N-l).  (13) 

The  parameter  b  gives  a  measure  of  the  amount  of  conditioning 
per  trial.  If  a  response  to  one  stimulus  has  no  effect  on  the  condition- 
ing at  centers  corresponding  to  other  stimuli,  the  n  is  independent  of 
M  .  But,  if  the  response  to  one  stimulus  results  in  the  stimulation  of 
inhibitory  neurons  terminating  at  the  various  other  conditioning  cen- 
ters, then  b  will  be  less  when  there  are  more  items  M  to  be  learned. 
We  may  account  for  this  by  introducing  as  a  rough  approximation, 
the  relation 

k 

where  r\  and  'Q  are  two  parameters  replacing  bk. 

Assuming  eow  =  eoc ,  substituting  equations  (5),  (12)  and  (13) 
into  (11),  and  eliminating  c  by  equation  (7),  we  obtain  a  differential 
equation  in  zv  and  n .  For  the  initial  condition  w  =  0  for  n  =  0,  and 
with  b  eliminated  by  relation  (14),  the  solution  of  the  differential 
equation  is 

(N-l)et»  N 

w  = log .      (15) 

Nf-f-0        (Nf-  f-0) e-vm  +  N  -  Nf  +  f 

This  equation  gives  the  number  w  of  errors  as  a  function  of  the  num- 
ber n  of  trials  for  any  number  M  of  items,  for  any  number  of  N  pos- 
sible choices,  and  for  any  fraction  (1  —  /)  of  prompting  by  the  ex- 
perimenter. All  this  involves  only  two  parameters  C  and  r\ .  As  we 
have  considered  a  highly  over-simplified  mechanism  and  introduced 
a  number  of  approximations,  it  is  not  to  be  expected  that  the  predic- 
tions of  equation  (15)  should  hold  over  too  wide  a  range  of  values 
of  M  and  N  . 

In  Figure  4  are  data  obtained  from  a  single  experiment  for  the 
purpose  of  illustrating  a  rather  special  case  of  the  experimental  pro- 
cedure outlined  above.    The  experiment  corresponds  to  the  case  in 


88    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


n-»- 

Figure  4. — Comparison  of  theory  with  experiment:  simple  learning.  Curve, 
theoretical  from  equation  (15);  points,  experimental  (Landahl,  1941).  Abscissa, 
number  of  trials ;  ordinate,  number  of  errors. 


which  M  =  N  and  /  =  1 .  Then  setting  y\  =  1.15  and  C  =  .098  we  ob- 
tain the  three  curves  for  which  N  =  4,  N  =  8,  and  N  =  12  respectively. 
The  values  of  v\  and  C  may  be  determined  from  one  point  of  each  of 
any  two  curves.  The  third  curve  is  then  without  unknown  parameters. 
But  another  family  of  such  curves  is  also  determined  by  equation 
(15),  without  any  additional  parameters,  for  the  case  in  which 
prompting-  follows  each  wrong'  response,  i.e.,  f  =  0  .  In  fact,  /  can 
be  given  any  value  in  the  range  0  =  /  =  1 ,  and  M  and  N  need  not  be 
equal. 

From  the  previous  discussion,  we  should  expect  b  to  depend  upon 
the  strength  of  reward.  To  a  first  approximation,  we  may  set 
6  =  alP/k ,  where  p  is  a  measure  of  the  strength  of  reward.  Similar- 
ly, we  can  set  ft  ==■  a2/pk  where  p  is  a  measure  of  the  strength  of  pun- 
ishment. Equation  (15)  then  determines  a  hyper-surface  in  the  seven 
variables,  w ,  n ,  N ,  M ,  f ,  p ,  and  p  in  terms  of  three  parameters 
a-! ,  a2 ,  and  C  • 


CONDITIONING  89 

Furthermore,  we  may  wish,  to  determine  what  the  performance 
in  terms  of  the  probabiliy  of  a  correct  response  would  be  if,  after  n' 
trials,  the  experiment  is  discontinued  for  some  interval  of  time.  The 
addition  of  a  single  parameter  enables  one  to  determine  the  perform- 
ance in  terms  of  the  various  experimentally  controlled  variables.  One 
can  then  also  determine  the  time  required  for  the  performance  to  drop 
to  some  preassigned  level.  It  is  found,  for  example,  that  while  in- 
creasing the  number  of  trials  beyond  some  value  has  little  or  no  effect 
on  the  performance  at  the  time,  it  may  have  a  considerable  beneficial 
effect  on  the  performance  at  some  subsequent  time.  This  is  essen- 
tially what  occurs  in  overlearning. 

But  these  results  apply  only  to  the  case  of  recognition-learning 
as  we  have  assumed  that  the  stimulus  to  which  response  is  to  be  made 
is  present  at  the  time  of  choice.  A  generalization  of  the  results  can 
be  made  so  as  to  include  the  case  of  recall-learning  (Landahl,  1943). 
A  number  of  parameters  must  be  introduced  in  this  case,  but  also  two 
new  experimental  variables  enter.  One  variable  determines  whether 
the  experiment  is  that  of  recognition  or  recall.  The  other  variable  is 
the  number  of  correct  responses,  as  in  this  case  c  cannot  be  deter- 
mined from  c  =  n  +  w  due  to  the  "equality"  response,  which  in  this 
case  is  a  lack  of  response.  Thus  with  a  small  number  of  parameters 
specified  experimentally,  c  and  w  can  be  determined  for  various  values 
of  the  seven  other  variables  and  the  result  may  then  be  compared  with 
experiment. 


XII 

A  THEORY  OF  COLOR-VISION 

According  to  the  Young-Helmholtz  theory,  any  color  can  be 
matched,  at  least  after  sufficient  desaturation  by  admixture  with 
white  light,  by  combining  in  suitable  proportions  lights  of  three  given 
colors  (cf.  Peddie,  1922) .  These  three  colors  may  be  chosen  arbitrari- 
ly except  that  no  one  is  to  be  matchable  by  combining  the  other  two. 
The  fact  that  three  primary  colors  are  sufficient  for  a  match  of  all 
others  strongly  suggests,  as  Helmholtz  brought  out,  that  retinal  ele- 
ments of  three  distinct  types  are  involved  in  the  perception  of  color. 
This  interpretation  has  not  gained  universal  acceptance  by  investi- 
gators, two  of  the  objections  being  that  anatomical  studies  fail  to 
differentiate  three  types  of  receptor,  and  that  the  degree  of  acuity 
with  monochromatic  illumination  is  too  high  to  be  accounted  for  by 
only  one-third  the  total  density  of  elements.  Hence  theories  have 
been  proposed  to  yield  quantitative  predictions  of  the  discriminal  pre- 
cision in  judging  color-differences  in  particular  without  postulating 
three  distinct  types  of  receptor  (e.g.  Shaxby,  1943).  However,  the 
case  here  is  analogous  to  the  case  of  discrimination  of  intensity-dif- 
ferences in  general:  different  intensities,  and  also  different  colors, 
can  occasion  qualitatively  different  responses  so  that  at  some  place 
in  the  neural  pathway  from  receptor  to  effector  the  locus  of  the  a 
must  be  capable  of  varying  with  variation  of  the  stimulus.  This  state- 
ment is  practically  self-evident  since  obviously  different  final  path- 
ways are  involved  in  reaching  the  different  effectors,  so  that  the  only 
question  is  where  the  variation  occurs — centrally,  along  the  afferent 
pathway,  or  along  the  efferent  pathway.  Wherever  this  may  take 
place,  the  method  demands  explanation.  But  the  statement  is  also  a 
consequence  of  the  well  known  Muller's  law  of  specificity,  as  is  clearly 
brought  out  by  Carlson  and  Johnson  (1941). 

Hence  the  neural  mechanism  which  mediates  any  discriminal  pro- 
cess must  provide  for  a  segregation  of  the  neural  pathways  affected 
by  different  stimuli.  In  the  case  of  a  simple  stimulus  characterized 
by  a  single  parameter,  S  ,  there  needs  to  be  only  a  one-dimensional 
array  of  synapses  reached  by  neurons  from  the  receptor  in  such  a 
way  that  when  S  has  one  value  the  resultant  a  is  positive  at  one  set 
of  the  synapses,  and  when  the  value  of  &  is  changed  sufficiently  the 
resultant  a  is  positive  at  a  different  set  of  the  synapses.  A  mechanism 
capable  of  bringing  this  about  was  described  in  chapter  ix.  When  the 
stimulus  is  complex  and  requires  three  parameters,  Sv(i  =  1 ,  2 ,  3), 

90 


A  THEORY  OF  COLOR- VISION  91 

to  specify  it,  then  a  three-dimensional  array  of  synapses  is  required. 

In  speaking-  of  one-dimensional  and  three-dimensional  arrays  we 
are  not  referring1  to  their  actual  spatial  distribution  in  the  cortex, 
since  manifestly  all  synapses  are  distributed  in  three  spatial  dimen- 
sions. We  refer  only  to  an  abstract  mode  of  classifying  all  the  syn- 
apses in  question.  In  the  case  of  the  discriminating  mechanism  of 
chapter  ix,  each  synapse  of  the  discriminating  center  is  character- 
ized by  a  certain  h  in  such  a  way  that  given  any  interval  bounded  by 
the  numbers  h'  and  h" ,  it  is  possible  to  say  unambiguously  of  any 
given  synapse  whether  its  associated  h  does  or  does  not  lie  within  this 
interval.  In  the  mechanism  now  to  be  discussed,  of  which  we  say  that 
the  synapses  form  a  three-dimensional  array,  each  synapse  is  char- 
acterized by  the  set  of  three  parameters  Sx ,  S2 ,  and  S3  and  given  any 
set  of  these  intervals  Si  to  Si",  we  can  say  unambiguously  of  any 
synapse  that  each  Si  associated  with  it  does  or  does  not  lie  upon  the 
interval  from  Sh'  to  Si". 

We  shall  now  describe  a  mechanism  which  generalizes  that  of 
chapter  ix  and  provides  the  segregation  of  pathways  required  for 
the  discrimination  of  colors.  While  it  may  not  be  the  simplest  one 
possible,  it  does  possess  the  necessary  qualitative  properties  and  no 
other  mechanism  has  been  proposed  which  does.  We  follow  Helm- 
holtz  and  assume  three  types  of  retinal  receptors,  each  connected  to 
all  the  synapses  of  the  three-dimensional  array  constituting  the 
"color-center."  We  utilize  the  three-receptor  hypothesis  because  it  is 
convenient,  not  because  we  are  necessarily  convinced  that  it  is  "true." 
We  consider  a  small  region  of  the  retina  only,  containing  one  of  each 
of  the  three  primary  receptors,  and  we  disregard  the  problem  of 
spatial  localization  or  other  attributes  of  the  sensations  which  accom- 
pany the  stimulation  of  these  receptors,  confining  ourselves  exclu- 
sively to  the  sensation  of  color.  Admittedly  the  other  attributes  de- 
mand explanation,  but  we  regard  the  problems  as  distinct. 

The  spatial  arrangement  required  of  the  synapses  at  the  color- 
center  is  highly  arbitrary,  and  while  we  imagine  a  specific  spatial 
localization  of  the  various  synapses  this  is  for  convenience  of  de- 
scription only.  With  this  understood,  we  suppose:  ( 1 ),  each  primary 
receptor  is  associated  with  a  particular  one-dimensional  array,  or 
axis,  of  synapses,  the  three  axes  being  mutually  orthogonal  and  all 
concurrent  at  a  point  O;  (2),  the  stimulation  of  the  i-th  primary  re- 
ceptor in  the  amount  Si  occasions  the  production  of  a  throughout  the 
color- center,  the  density  being  greatest  all  along  its  associated  axis 
and  being  everywhere  a  function  a,  (Si  ,  P)  of  Si  and  of  the  assumed 
position  P;  (3),  as  a  function  of  position  <n  depends  upon  two  para- 
meters only,  the  distance  r  =  OP ,  from  0,  and  the  angle  dt  between 


92    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

OP  and  the  i-th  axis.  The  assumption  that  a,  depends  upon  6,  alone 
is  made  for  the  sake  of  simplicity,  and  a  natural  generalization  would 
be  to  allow  all  three  angles  to  enter,  these  angles  being  related  by  the 
identity 

2cos20i  =  l.  (1) 

For  further  simplification  we  suppose  that  the  dependence  upon  6i  is 
through  the  cosine  so  that  the  connections  from  each  primary  recep- 
tor are  symmetric  about  the  associated  axis,  and  we  suppose  in  addi- 
tion that  when  the  units  for  the  Si  are  properly  chosen  the  three  func- 
tions a,  (S ,  r ,  cos  6)  are  identical.  Neither  of  these  restrictions  is 
essential. 

The  immediate  result  at  the  color-center  of  the  stimulation  of  the 
three  primary  receptors  is  then  the  production  at  every  point  P  of 

^(S  ,P)  =2l*i(Si,r,cosOl).  (2) 

If,  finally,  the  functions  ai  are  so  chosen  that  for  any  5 ,  a  has  aways 
a  maximum  at  a  single  point  P  ,  then  the  introduction  of  sufficient  mu- 
tual inhibition  between  synapses  (cf.  chap,  iii)  will  make  the  resul- 
tant a  negative  everywhere  but  in  the  neighborhood  of  P ,  and  the 
desired  segregation  of  the  pathways  is  secured. 

If  the  functions  are  properly  chosen  the  analytical  result  will  be 
essentially  a  representation  of  the  familiar  color-pyramid,  as,  indeed, 
we  wish  it  to  be,  since  this  is  found  empirically  to  provide  an  accurate 
representation  of  the  phenomena.  What  we  provide,  and  what  is  not 
contained  in  the  theory  of  the  color-pyramid,  is  a  neural  mechanism 
capable  of  mediating  perceptions  organized  in  this  way.  We  have 
supposed  that  the  basis  for  a  difference  between  perceptions  lies  in 
the  discreteness  of  loci  of  excitation  and  the  mechanism  here  described 
is  capable  of  separating  the  loci.  By  the  same  rule  an  observed  sim- 
ilarity must  have  its  basis  in  some  community  of  the  loci.  Hence  if 
we  suppose  that  synapses  of  the  color-center  which  are  collinear  with 
the  origin  are  all  connected  to  some  further  center  common  to  that  set 
we  have  a  possible  basis  for  the  identity  of  colors  of  varying  intens- 
ity; if  the  centers  of  this  group  corresponding  to  complanar  rays 
themselves  lead  to  a  common  tertiary  center  we  have  a  basis  for  iden- 
tity of  colors  of  varying  intensity  and  saturation,  and  so  on. 

The  form  of  the  predictions  from  this  mechanism  must  coincide 
with  those  of  the  simpler  one  (chap,  ix)  when  only  intensity,  but  not 
color  or  saturation,  is  varied.  Hence  each  function  <n(S ,  r ,  1)  must 
be  proportional  to  r(S—r)  if  we  suppose  the  /^-centers  of  the  previous 
mechanism  to  be  uniformly  spaced.  Further,  each  a\  as  a  function  of 
Si  should  have  a  maximum  at  0; .  =  0  .  The  simplest  possible  supposi- 


A  THEORY  OF  COLOR-VISION  93 

tion  is  then  that 

en  =  r  (Si  —  r)  cos  9i .  (3) 

With  this  assumption  the  maximum  of  a  for  any  stimulation  Si  occurs 
for 

cos  $! :  cos  82:  cos  83  =  St  —  r:  S2  ~  ri  S3  —  r , 

with  r  the  smaller  root  of 

6r2  -  3r  2  Si  +  2  S>2  =  0  . 

A  more  exact  specification  of  the  form  of  the  functions  en  can  be  made 
only  by  a  detailed  comparison  with  experimental  facts. 


XIII 

SOME  ASPECTS  OF  STEREOPSIS 

In  this  chapter  we  present  some  purely  formal  considerations  of 
the  visual  perception  of  space  without  providing  any  specific  neural 
mechanism.  Such  a  formal  analysis  is,  of  course,  a  necessary  pre- 
liminary since  evidently  when  we  do  not  know  in  advance  the  struc- 
ture of  the  neural  mechanism  involved  we  must  know  what  is  re- 
quired of  it  if  we  are  to  deduce  the  structure. 

The  structure  of  subjective  space  is  developed  gradually  during 
the  life  of  the  individual  and  is  the  resultant  of  diverse  sensory  cues — 
visual,  auditory,  kinaesthetic,  and  perhaps  others.  The  recognition 
of  two  pin-pricks  simultaneously  applied  to  different  parts  of  the 
body,  as  distinct,  involves  discrimination  of  a  certain  primitive  type 
and  requires  a  certain  minimal  separation  of  the  points;  to  recog- 
nize that  one  pin-prick  is  located  at  such  a  distance  and  in  such  a  di- 
rection from  the  other  involves  a  judgment  much  more  advanced  in 
form  and  requires  a  neural  mechanism  of  much  greater  complication. 
To  account  for  the  first  judgment  no  assumption  is  required  beyond 
the  distinctness  of  the  neural  pathways,  and  of  the  cortical  centers 
ultimately  affected  by  the  two  pricks.  But  the  second  judgment,  while 
keeping  the  pricks  distinct,  assimilates  them  into  a  certain  continuum 
and  hence  relates  them  in  a  definite  way.  Consequently  the  cortical 
centers  must  be  connected  with  each  other  and  with  the  motor  cen- 
ters in  some  definite  way,  perhaps  in  such  a  way  as  to  make  possible 
the  continuous  movement  of,  say,  an  index  finger  from  the  location 
of  one  prick  to  that  of  the  other. 

Similar  remarks  may  be  made  of  vision.  Let  us  confine  ourselves 
for  the  present  to  monocular  vision.  The  judgment  that  two  objects, 
seen  simultaneously,  are  distinct,  and  the  judgment  as  to  their  rela- 
tive positions,  are  quite  different  judgments  and  the  first  does  not  by 
any  means  imply  the  second.  The  second  judgment  may  be  somehow 
associated  with  the  ocular  rotations  that  would  be  necessary  for  the 
fixating  of  first  the  one  point  and  then  the  other  (cf.  Douglas,  1932), 
in  which  case  if  the  visual  space  has  been  integrated  into  a  unified 
space  of  perception  as  a  whole,  there  may  be  associations  of  some  sort 
with  movements  of  the  body  or  of  a  member  from  the  one  point  to 
the  other.  In  either  case,  however,  that  of  the  pin-pricks  or  that  of 
the  visual  objects,  we  say  nothing  about  whether  the  motor  and  kin- 
aesthetic  connections  are  a  part  of  the  native  endowment  of  the  or- 

94 


SOME  ASPECTS  OF  STEREOPSIS  95 

ganism  and  formed,  possibly  before,  possibly  after  birth,  indepen- 
dently of  any  experience,  or  whether  they  are  somehow  due  to  con- 
ditioning. 

Whatever  may  be  the  nature  of  a  given  perception,  therefore,  the 
perception  of  a  disjunction  implies  a  neural  disjunction  of  some  sort, 
whereas  the  localization  of  the  disjoined  elements  within  a  field  of 
relations  implies  the  presence  of  interconnections  of  some  sort  be- 
tween nervous  elements  involved  in  the  perception  of  these  elements, 
whether  these  interconnections  were  previously  functional,  or  only 
became  functional  as  a  result  of  the  perception  itself.  And  while 
there  may  be  no  unique  structure  of  nervous  interconnections  capable 
of  yielding  the  system  of  relations  as  perceived — the  solution  of  the  in- 
verse problem  of  neural  nets  is  not  necessarily  unique — nevertheless 
there  are  limitations,  and  empirical  anatomical  data  may  serve,  in 
time,  to  complete  the  characterization. 

The  problem  we  wish  to  consider  is  the  following.  By  whatever 
means  it  is  acquired,  the  normal  human  adult  does  possess  a  percep- 
tual space  within  which  he  locates  the  objects  of  his  perception.  With 
objects  some  distance  away  from  him,  visual  cues  are  the  chief,  and 
frequently  they  are  the  only,  means  available  to  him  for  the  localiza- 
tion of  these  objects  within  this  space.  In  these  cases,  where  the 
visual  cues  are  the  only  ones,  how  is  the  localization  effected?  That 
is,  what  kinds  of  cues  can  be  provided  by  the  eyes  alone,  which,  to- 
gether with  past  associations  (but  not  the  memory  of  a  previous  lo- 
calization), may  serve  the  subject  in  localizing  the  object  within  his 
perceptual  space  If  the  eyes  themselves  provide  several,  more  or  less 
independent,  cues  for  localizing  the  same  object— and  they  certainly 
do— then  the  final  localization  must  come  as  a  kind  of  resultant  of 
all  of  these.  Under  normal  conditions  these  cues  would  doubtless  act 
in  harmony  and  reinforce  one  another,  thus  providing  a  fairly  ac- 
curate judgment.  But  under  abnormal  conditions  due  to  pathology 
or  instrumentation,  the  normal  harmony  would  be  disrupted,  making 
the  perceived  relations  a  bizarre  distortion  of  the  true  ones.  In  fact, 
it  is  by  considering  the  nature  and  occurrence  of  these  distortions 
when  the  cues  are  in  disharmony  that  we  might  hope  to  get  our  best 
information  as  to  their  separate  modes  of  operation. 

In  the  system  of  spatial  relations  among  disjoint  elements,  per- 
haps the  simplest  and  most  primitive,  beyond,  of  course,  the  mere 
fact  of  separation,  is  the  amount  of  separation.  This  is  easily  under- 
stood in  terms  of  kinaesthesis.  Greater  separation  requires  more 
movement  for  crossing  it,  a  more  intense  kinaesthetic  sensation,  and 
in  these  terms  our  discussion  of  the  discrimination  of  intensities  may 
find  an  application  here.  With  reference  to  visually  perceived  extent 


96   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

a  suggestion  has  been  made  already  in  this  direction  (Householder, 
1940;  cf.  chap  ix). 

Perhaps  the  most  obvious  cue  for  judging  distance  is  the  "ap- 
parent size"  of  the  object  when  the  actual  size  is  known  or  inferred 
from  previous  experience.  The  "apparent  size"  is  by  definition  the 
solid  angle  subtended,  but  is  not  necessarily  the  size  the  object  ap- 
pears to  have.  Thus  a  distant  man  appears  small,  but  when  he  is 
fairly  close  the  size  he  appears  to  have  stays  fairly  constant  while 
his  distance,  and  hence  his  "apparent  size,"  varies  over  quite  a  wide 
range.  The  "apparent  size"  is  proportional  to  the  size  of  the  retinal 
image  and  provides  a  distance  cue. 

An  object  seen  through  a  spyglass  appears  flattened  and  a  pos- 
sible explanation  can  be  found  from  considerations  of  apparent  size. 
Thus  consider  a  cube  with  one  edge  nearly,  but  not  quite,  sagittal  in 
direction,  and  suppose  it  is  viewed  through  a  spyglass  magnifying  in 
the  ratio  M  =  1  +  u .  That  is,  let  the  retinal  image  of  the  cube  as 
seen  through  the  spyglass  be  M  times  that  formed  by  the  cube  seen 
at  the  same  distance  by  the  naked  eye.  If  the  actual  distance  of  the 
front  face  is  d ,  and  if  the  edge  of  the  cube  has  length  s ,  then  the 
back  face  has  the  actual  distance  d  +  s  ,  but  due  to  the  magnification, 
front  and  back  faces  appear  to  be  only  1/M  times  as  far.  They  ap- 
pear, therefore,  to  have  the  distances  d/M  and  (d  +  s)/M  .  But  this 
leaves  for  the  apparent  depth  of  the  cube  only  the  distance  s/M  . 

While  it  is  well  known  that  qualitatively  the  effect  is  present  as 
described,  no  quantitative  data  are  at  hand,  and  the  theory  here  sug- 
gested might  fail  to  meet  the  more  exacting  requirements  of  a  quan- 
titative test.  In  the  first  place,  it  is  assumed  that,  in  the  absence  of 
other  cues,  the  perceived  distance  would  be  exactly  1/M  times  the 
actual  distance.  This  might  not  be  the  case.  The  percieved  distance 
might  be,  say  1/li  times  the  true  distance,  where  /u  <  M  .  But  if  so, 
we  should  expect  the  judged  depth  to  be  1///  times  the  actual  depth 
and  the  judged  size  to  be  M/li  times  the  actual  size.  Moreover  one 
could  perform  an  experiment  in  which  a  truncated  pyramid  is  pre- 
sented instead  of  a  cube,  the  dimensions  being  such  that  the  subject 
would  be  expected  to  perceive  it  as  a  cube.  For  this,  when  the  dis- 
tance is  judged  to  be  1///  times  the  actual  distance  d  and  the  magnifi- 
cation is  M  times,  the  retinal  image  is  the  same  as  would  be  produced 
by  a  cube  seen  with  the  naked  eye  at  a  distance  of  dj  ii .  Now  if  a 
cube  whose  edges  are  Ms/tu  is  placed  so  that  the  nearest  face  is  at  a 
distance  of  d/  n ,  then  the  visual  angles  subtended  by  an  edge  of  the 
front  face  and  an  edge  of  the  back  face  are  Ms/d-  and  Ms/  (d  +  Ms). 
These  must  be  the  visual  angles  of  the  faces  of  the  truncated  pyra- 
mid placed  at  a  distance  d  and  magnified  M  times.    Hence,  if  the 


SOME  ASPECTS  OF  STEREOPS1S  97 

depth  of  the  frustum,  as  well  as  each  of  the  edges  nearest  the  obser- 
ver, is  s ,  the  edges  of  the  other  face  must  be 

s(s  +  d)/(Ms  +  d)  =s[l  -us/d]  , 

if  Ms  is  small  by  comparison  with  d .    Note  that  this  result  is  inde- 
pendent of  // . 

While  no  mechanism  is  suggested  here  for  this  dependence  of 
perceived  distance  upon  apparent  size,  we  note  that  a  converse  mech- 
anism has  been  suggested  by  Landahl  (1939a)  to  account  for  the 
constancy  of  perceived  size  with  the  concomitant  (mutually  inverse) 
variation  of  apparent  size  and  actual  distance. 

The  factor  of  accommodation  (Stanton,  1942)  is  certainly  not 
sensitive  as  a  distance-cue,  but  there  is  evidence  that  it  does  operate 
(Grant,  1942).  Distant  objects  are  seen  with  the  relaxed  eye  (if  em- 
metropic), whereas  to  focus  clearly  on  nearby  objects  requires  an 
effort  of  accommodation.  Convergence,  a  binocular  cue,  is  more  certain 
(Householder,  1940c).  Convergence  upon  near  objects  also  require 
an  effort,  although  it  is  known  that  the  visual  axes  of  relaxed  eyes  are 
not  parallel,  but  diverge,  so  that  some  effort  is  required  also  for  bi- 
nocular vision  at  a  distance.  The  cues  of  both  accommodation  and 
convergence  are  muscular,  and  neither  is  sensitive  enough  to  provide 
the  fine  discrimination  known  to  be  possible  in  binocular  stereopsis. 
In  fact,  binocular  stereopsis  can  be  achieved  by  means  of  a  stereo- 
scope when  the  visual  axes  are  parallel  and  accommodation  is  relaxed. 
Thus  other  cues  must  be  sought. 

In  normal  binocular  vision  whereas  there  are  two  retinal  images, 
one  in  each  eye,  of  the  object  fixated,  there  is  but  one  cortical  image 
— the  individual  sees  but  one  object.  On  the  other  hand  if  the  atten- 
tion, but  not  the  fixation,  is  shifted  to  an  object  enough  nearer  or  far- 
ther than  the  point  of  fixation,  then  two  images  of  this  single  object 
are  seen.  Objects  somewhere  in  between  the  point  of  fixation  and  the 
object  seen  doubly  may  be  seen  singly  but,  if  so,  they  can  be  recog- 
nized as  nearer  to  or  farther  from  the  observer,  as  the  case  may  be, 
than  the  point  of  fixation. 

Let  CL  and  CR  represent  the  centers  of  rotation  of  the  left  and  right 
eyes,  respectively.  The  separation  between  nodal  point  and  center 
of  rotation  is  so  slight  that  for  present  purposes  we  may  regard  CL 
and  CR  as  being  also  the  nodal  points.  Let  P  represent  the  fixation- 
point.  Then  PCL  and  PCR  are  the  two  visual  axes;  let  these  cut  the 
retinas  at  PL  and  PR  .  The  two  retinal  images  of  P  are  therefore  lo- 
cated at  PL  and  PR  ,  and  when  so  located  these  images  fuse  and  the 
point  P  is  seen  singly.  Let  P  be  any  point  on  the  line  CLP  be- 
tween CL  and  P .  Then  P  is  also  imaged  at  PL  on  the  left  retina,  but 


98   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

on  the  right  retina  the  image  is  at  some  point  PR  in  the  temporal 
direction  from  PR  .  If  an  object  is  located  at  P',  and  if  P'  is  not  too 
far  removed  from  P ,  there  may  be  fusion  even  of  the  images  on  PL 
and  P'R  ,  though  because  of  the  more  temporal  location  of  P'R  with 
respect  to  PR  ,  P'  is  judged  nearer  the  observer  than  P  .  The  situation 
when  P'  lies  beyond  P  is  similar  except  that  P'R  is  then  medial  to  PR  . 
But  if  P'  is  moved  much  farther  or  much  closer,  fusion  is  no  longer 
possible,  and  two  images  result.  We  shall  say  that  PL  and  PR  are 
corresponding  points  on  the  two  retinas,  whereas  Ph  and  any  other 
point  P'R  different  from  PR  are  disparate. 

While  holding  the  fixation  at  P ,  any  other  point  Q  within  the 
binocular  visual  field  will  form  an  image  on  a  point  QL  of  the  left 
retina,  QR  on  the  right,  and  the  images  may  fuse,  but  only  if  Q  is  not 
too  near  or  too  far  away.  We  suppose  that  associated  with  each  point 
QL  of  the  left  retina  there  is  a  unique  point  QR  of  the  right  retina 
which  corresponds  to  it,  while  QL  and  any  other  point  Q'R  are  dis- 
parate. If  Q  has  as  its  images  the  point  QL  and  a  disparate  point  Q'R 
not  too  far  removed  from  QR  fusion  may  still  occur,  but  fusion  is  in 
some  sense  optimal  when  the  images  fall  on  corresponding  points. 
The  extent  of  the  disparity,  or  the  absence  of  it,  then  provides  a  cue 
for  the  localization  of  the  point  Q  in  the  subject's  visual  space.  It  is 
suggested  that  the  degree  of  convergence,  and  possibly  also  the  degree 
of  accommodation  serve  as  cues  for  the  localization  of  the  general 
region  about  P  with  respect  to  the  observer,  while  the  binocular  dis- 
parity gives  depth  to  this  region  and  makes  possible  the  localization 
with  respect  to  P  of  the  objects  in  its  immediate  neighborhood.  Thus 
pictures  seen  in  a  stereoscope  are  seen  as  vaguely  somewhere  far  off, 
although  the  positions  of  the  various  details  with  respect  to  one  an- 
other are  very  definite. 

The  simplest  neurological  picture  to  correspond  to  this  seems  to 
be  the  following.  Suppose  there  is  a  binocular  representation  in  the 
visual  cortex  which  is  three-dimensional  in  character.  Simultaneous 
stimulation  of  corresponding  points  QL  and  QR  on  the  two  retinas 
leads  to  maximal  excitation  a  at  an  associated  point  of  the  binocular 
cortex.  Simultaneous  stimulation  of  slightly  disparate  points  QL  and 
Q'R  leads  to  excitation  <r  of  somewhat  lesser  amount  and  at  a  some- 
what different  location.  Simultaneous  stimulation  of  widely  disparate 
points  gives  only  subthreshold  excitation,  if  any,  in  the  binocular 
cortex,  though  it  may  lead  to  excitation  in  the  two  cortical  regions 
where  the  retinas  are  separately  represented  (cf.  Bichowsky,  1941; 
Verhoff,  1925).  This  would  involve  at  least  three  visual  areas  in  the 
cortex,  two  monocular  regions,  which  might  be  only  two-dimensional, 
and  one  binocular  region  which  corresponds,  point  for  point,  to  a  cer- 


SOME  ASPECTS  OF  STEREOPSIS  99 

lain  region  of  external  space  containing  the  point  of  fixation. 

There  are  doubtless  other  possibilities,  but  whatever  they  are, 
they  must  provide  for  the  fact  that  a  slight  disparity  or  none  leads  to 
a  unitary  cortical  representation,  different  representations  differing 
according  to  the  magnitude  and  direction  of  the  disparity.  In  other 
words,  the  three-dimensional  character  must  be  present,  whether  in 
a  spatial  fashion  or  otherwise.  The  occurrence  of  anomalous  fusion 
in  some  cases  of  strabismus  may  seem  to  invalidate  this  argument 
(Brock,  1940).  However,  it  can  be  questioned  whether  a  real  fusion 
occurs  here,  and  it  seems  simpler  to  suppose  that  a  quite  different 
mechanism  is  at  work,  with  the  factor  of  conditioning  playing  a  pre- 
dominant role,  much  as  it  must  do  in  monocular  depth  judgments. 

If  to  each  point  of  one  retina  there  is  a  unique  "corresponding" 
point  on  the  other  retina,  the  lines  joining  any  pair  of  these  to  the 
nodal  points  will  generally  be  skew  and  fail  to  intersect  in  any  point 
in  space.  For  any  given  fixation  of  the  two  eyes  the  locus  of  points 
in  space  where  pairs  of  corresponding  lines  do  intersect  is  called  the 
horopter.  There  is  at  least  one  such  point,  the  fixation-point,  and  in 
general  the  horopter  is  a  curve  in  external  three-space  (Helmholtz, 
1896).  Consider  the  situation  in  which,  the  head  being  upright,  the 
visual  axes  are  horizontal.  The  horizontal  plane  containing  the  visual 
axes  does  not  necessarily  contain  any  part  of  the  horopter  except  the 
fixation-point  and  other  isolated  points.  However,  if  we  take  any 
point  QL  on  the  intersection  of  this  horizontal  plane  with  the  left 
retina,  there  will  be  a  point  Q'R  of  the  intersection  of  this  plane  with 
the  other  retina  which  is  closest  of  all  these  intersections  to  the  true 
corresponding  point  QR  ,  and  the  point  Q  in  space  which  projects  QL 
and  Q'R  lies  somewhere  on  this  plane.  The  locus  of  the  point  Q  in  the 
horizontal  plane  of  fixation  we  shall,  for  the  present,  refer  to  as  the 
horopter. 

It  is  evident  that  in  symmetric  convergence  this  horopter-curve 
should  be  symmetric  with  respect  to  the  subject's  medial  plane.  If 
each  eye  were  symmetric  about  the  fovea,  so  that  corresponding  points 
were  at  equal  distances  from  the  two  foveas,  it  is  easy  to  see  that  the 
horopter  would  always  be  a  circle.  Actually  it  is  found  empirically 
that  for  a  suitably  located  fixation-point  the  horopter  is  a  straight 
line,  while  for  nearer  fixation  it  is  an  ellipse,  for  more  distant  fixation 
a  hyperbola,  either  conic  passing  through  the  nodal  points  of  the  two 
eyes  (Ogle,  1938). 

Now  granting  the  existence  of  a  rectilinear  horopter  and  the 
anatomical  fixity  of  the  corresponding  points,  that  the  other  horop- 
ters are  conies  of  the  sort  described  follows  from  elementary  projec- 
tive geometry,  and  furthermore  the  equations  of  these  can  be  deduced 


100   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

with  no  parameters  undetermined  once  the  rectilinear  horopter  has 
been  located  (Householder,  1940). 

If  the  geometric  analysis  is  carried  further  one  can  see  that  on 
placing"  before  one  eye,  say  the  left,  a  cylindrical  size-lens,  axis  90°, 
the  horopter  should  undergo  a  rotation  with  the  result  that  the  en- 
tire visual  field  would  appear  to  have  undergone  a  rotation  in  a  clock- 
wise direction.  This  is  indeed  borne  out  by  experiment,  and  the  amount 
of  the  rotation  is,  as  predicted,  proportional  to  the  increment  in  size 
(Ogle,  1938). 

Now  so  far  as  the  horopter  is  concerned  there  is  no  reason  to 
suppose  that  any  shift  of  the  subjective  space  should  occur  when  the 
cylindrical  size-lens  is  placed  with  axis  180°  instead  of  90°.  Never- 
theless, it  turns  out  that  there  is  a  shift,  of  approximately  the  same 
amount  but  in  the  opposite  direction  (Ogle,  1940).  But  none  of  the 
cues  so  far  mentioned  can  account  for  this,  so  another  one  must  be 
at  work. 

The  experimental  procedure  here  is  to  set  up  before  the  observer 
a  plane  which  is  free  to  rotate  about  a  vertical  axis  and  which  con- 
tains a  few  small  circles  for  fusion  arranged  in  two  horizontal  rows. 
The  subject  is  asked  to  adjust  the  plane  until  it  is  parallel  to  his  own 
frontal  plane.  Best  results  are  obtained  when  the  fusion-contours 
are  "restricted  to  relatively  small  areas  above  or  below  the  center  of 
the  plane"  (Ogle,  1940).  It  is  found  that  the  rotation  is  approxi- 
mately proportional  to  the  magnification-increment  when  this  is  not 
too  great,  but  that  the  effect  breaks  down  when  the  magnification 
exceeds  a  certain  critical  amount. 

It  is  evident  that  the  vertical  disparities,  rather  than  the  hori- 
zontal disparities,  are  producing  the  effect,  and  this  fact,  together 
with  the  fact  of  the  breakdown  of  the  effect  for  higher  magnifica- 
tions, suggests  that  the  subjects  are  locating  the  medial  plane  rather 
than  the  frontal  plane.  Normally  the  medial  plane  would  be  the  locus 
of  objects  whose  retinal  images  are  equal  in  the  vertical  direction, 
and  those  which  produce  unequal  images  would  be  on  the  same  side 
of  the  medial  plane  as  the  eye  having  the  larger  image.  Normally, 
too,  the  medial  and  the  frontal  planes  are  perpendicular.  But  if  a 
lens  giving  vertical  magnification  were  placed  before  the  left  eye,  and 
if  the  magnification  were  not  too  great,  the  points  yielding  equal 
retinal  images  would  be  to  the  right  of  the  true  medial  plane.  On  the 
basis  of  this  cue  any  object  would  therefore  appear  to  lie  closer  to 
the  left  eye  and  farther  from  the  right  than  it  actually  does.  But 
when  the  magnification  is  great  enough,  any  object  whatever  forms 
a  larger  image  on  the  left  retina,  and  in  this  extreme  situation,  if 
not  sooner,  the  localization  would  be  accomplished  by  means  of  other 


SOME  ASPECTS  OF  STEREOPSIS  101 

cues,  say  the  horizontal  disparities  which  are  unaffected  by  the  lens. 
A  quantitative  analysis  bears  out  the  hypothesis  just  outlined 
(Householder,  1943).  If  we  take  for  the  nodal  points,  CL  and  CR  ,  the 
coordinates  (±  1,  0),  with  the  positive  y-axis  extending  forward 
from  the  observer,  and  if  the  magnification  is,  as  before, 

M  =  1  +  u , 

so  that  the  increment  of  magnification  in  the  vertical  direction  is 
100  u  %  ,  then  the  locus  of  points  in  the  horizontal  plane  whose  retinal 
images  are  equal  in  the  vertical  direction  is  the  circle 


/        M2  +  IV  /  2  M     \2 

{*-jF=V+v°={m^1)-  (1) 


or,  very  nearly, 


if  we  set 


/         1  +  u\2  /1  +  uV 


(2) 


u 

1  = ,  (3) 

2(1  +  u) 

then  the  tangent  of  the  angle  of  rotation  of  the  medial  plane  is,  for 
a  fixed  distance  y  of  the  fixation-point, 

x/y  =  Xy(l  +  Py2  +  21iyi---), 

and  for  smajl  angles  this  is  equal  to  the  angle  itself.  This  is  approxi- 
mately proportional  to  the  increment  of  magnification,  to  the  distance 
of  the  point  of  fixation,  and  also  to  the  interocular  distance  since  half 
of  this  distance  is  the  unit  employed.  The  figure  shows  a  comparison 
of  theory  and  experiment  with  two  subjects,  both  left  and  right  eyes. 
It  is  especially  to  be  noted  that  all  parameters  are  determinable, 
these  being  only  the  interocular  distance  and  the  magnification.  The 
breakdown  of  the  effect  occurs  before  the  equality  of  the  images  be- 
comes impossible  at  increments  of  about  6  or  8%.  When  a  spherical 
size-lens  is  employed  the  two  conflicting  effects  of  horizontal  and  ver- 
tical disparities  neutralize  one  another  for  small  magnifications  also 
up  to  about  6  or  8%,  while  for  larger  magnifications  the  horizontal 
disparities  provide  the  effective  cue  and  the  effect  of  the  vertical  dis- 
parities gradually  dies  out. 


102    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


2(f 


A  MB  C.  RIGHT  EVE 
0   M.B.C.  LEFT  EYE 
•  K.N.O.   RIGHT  EYE 
O  K.N  O.   LEFT  EYE 


l<f 


3  4 

PER  CENT 


Figure  1. — Comparison  of  theory  with  experiment:  rotation  of  subjective 
medial  plane  due  to  cylindrical  size-lens,  axis  180°,  before  one  eye.  Curve,  theo- 
retical from  equation  (4);  points,  experimental  (Ogle,  1938).  Abscissa,  magnifi- 
cation-increment produced  by  size-lens;  ordinate,  rotation  of  the  plane  in  degrees. 


PART  THREE 

XIV 

THE  BOOLEAN  ALGEBRA  OF  NEURAL  NETS 

To  the  extent  that  our  neurons  are  representative  of  those  stud- 
ied by  the  physiologist,  our  quantities  s  and  j  must  be  interpreted  as 
time-averages.  The  assumptions  underlying  any  purely  formal  theory 
can  be  justified  only  to  the  extent  that  the  predictions  to  be  drawn 
from  them  are  borne  out  by  experience.  However,  our  theory  is  not 
intended  to  be  a  formal  one  only.  On  the  contrary,  it  is  intended  that 
the  theoretical  neural  structures  will  represent  in  some  sense  the 
essential  character  of  the  actual  anatomical  structures  and  that  the 
postulated  behavior  of  the  constituent  neurons  will  be  representative 
of  the  actual  physiological  activity  of  the  neurons,  or  at  least  of  re- 
curring complexes  of  these,  in  the  real  organism.  Hence  our  postu- 
lated £  and  j  must  be  identified  with  physiological  states  or  events, 
and  the  postulated  laws  of  their  development,  or  something  closely 
approximating  them,  must  be  deduced  from  more  fundamental  prin- 
ciples. 

In  large  measure  this  is  a  statement  of  a  program  for  future  re- 
search. It  is  evident,  of  course,  that  the  e  and  j  are  to  be  correlated 
with  the  "action-spikes"  of  the  neurons.  These  occurrences  occupy 
milliseconds,  whereas  ordinary  psychological  behavior  is  generally  a 
matter  of  seconds  at  least,  which  justifies  a  statistical  averaging  pro- 
cedure. The  first  step  in  the  deduction  has  been  made  by  McCulloch 
and  Pitts  (1943),  and  we  now  outline  their  picture,  microscopic  as 
to  time,  of  the  neuron's  behavior,  a  picture  that  rests  almost  immedi- 
ately upon  direct  observations.  Later,  in  the  next  chapter,  we  indi- 
cate how,  by  carrying  out  a  suggestion  due  to  Landahl,  McCulloch  and 
Pitts  (1943),  this  deduction  might  be  completed.  But  since  the  com- 
plete deduction  of  the  macroscopic  picture,  with  which  we  have  been 
concerned  so  far,  from  the  microscopic  picture,  is  still  wanting,  these 
chapters  constitute  largely  a  digression  from  the  main  trend  of  our. 
discussions  and  will  appear  to  be,  in  some  details,  even  contradictory. 

The  most  striking  feature  of  the  nervous  discharge  as  observed 
in  the  laboratory  is  its  all-or-none  character.  A  change  in  the.  physio- 
logical state  of  the  neuron — e.g.  one  due  to  a  change  in  the.  oxygen, 
tension — may  alter  the  potential  of  any  discharge  the  neurpn  is  ca- 
pable of  making,  but  in  any  given  condition  either  the  maximal  re- 
sponse or  none  at  all  will  occur.  Since  nearly  all  overt  reactions 
have  gradations  the  complete  deduction  must  explain  in  detail  how 

103 


104    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

the  two  facts  are  to  be  reconciled  and  the  greater  part  of  this  mono- 
graph has  been  devoted  to  a  discussion  of  these  gradations.  For  the 
present,  however,  let  us  consider  some  of  the  formal  aspects  of  the 
all-or-none  feature  alone. 

An  afferent  neuron  may  form  any  number  of  synaptic  connec- 
tions with  an  efferent  neuron  and  the  firing  of  one  afferent  alone  may 
or  may  not  be  sufficient  to  elicit  a  discharge  in  the  efferent  neuron. 
In  order  for  a  discharge  to  be  elicited,  it  is  necessary  that  there  be  a 
sufficient  number  of  endfeet  of  discharging  afferents  all  located  with- 
in a  sufficiently  small  region  of  the  stimulated  neuron.  The  discharges 
of  the  afferent  neurons  must,  moreover,  all  occar  within  a  sufficiently 
small  time-interval.  The  minimal  number  of  endfeet  of  discharging 
afferents  required  for  eliciting  a  discharge  in  any  neuron  is  what  we 
now  call  the  threshold,  8  ,  of  the  neuron ;  the  endfeet  may  or  may  not 
all  come  from  the  same  afferent,  but  for  simplicity  we  suppose  them 
all  equal  in  their  effectiveness.  The  maximal  time-interval  within 
which  the  summation  can  occur  is  about  a  quarter  of  a  millisecond. 
There  is  a  delay  of  about  half  a  millisecond  between  the  arrival  of 
the  impulses  upon  a  neuron  and  the  initiation  of  its  own  discharge. 
Compared  to  this  synaptic  delay,  the  time  required  for  the  conduc- 
tion from  origin  to  terminus  is  quite  short. 

In  some  instances  the  arrival  of  a  nervous  discharge  upon  a  neu- 
ron will  have  the  effect,  not  of  initiating  a  discharge  in  it,  but  of  pre- 
venting a  discharge  that  would  otherwise  occur.  No  explanation  of 
this  phenomenon  of  inhibition  has  won  general  acceptance;  neither 
is  it  known  whether  the  inhibition  is  complete  or  only  partial.  But 
the  fact  of  its  occurrence  is  beyond  question,  and  only  the  details  of 
the  schematic  structure  will  be  affected  by  the  assumption  of  partial 
rather  than  complete  inhibition  (McCulloch  and  Pitts,  1943).  To  be 
definite  we  make  the  arbitrary  assumption  that  inhibition  when  it 
occurs  is  complete. 

The  McCulloch-Pitts  picture  involves  a  formal  representation  of 
neural  nets  in  terms  of  Boolean  algebra  as  follows:  Let  the  life-span 
of  the  organism  be  divided  into  elementary  time-intervals  of  common 
length  equal  to  the  period  of  synaptic  delay  and  introduce  this  inter- 
val as  the  time-unit.  Since  each  nerve-impulse  is  followed  by  a  refrac- 
tory period  of  about  half  a  millisecond  during  which  the  neuron  is 
incapable  of  further  activity,  no  neuron  can  fire  twice  within  any  unit 
interval,  and  moreover,  from  the  very  definition  of  the  unit  of  time, 
the  firing  of  a  neuron  within  a  given  unit  interval  can  cause  the  fir- 
ing of  any  efferent  neuron  only  during  the  next  unit  interval,  if  at  all. 
Summation  or  inhibition,  if  either  occurs,  is  only  effective  in  case  the 
summating  or  inhibiting  neurons  fire  within  the  same  unit  interval. 


THE  BOOLEAN  ALGEBRA  OF  NEURAL  NETS  105 

Now  consider  any  net  of  interconnecting  neurons.  Let  these 
neurons  be  assigned  designations  ©»  in  any  manner  and  let  Ni(t) 
denote  the  proposition  that  the  neuron  Ci  has  fired  during  the  t-th 
time-interval.  If  this  is  any  but  a  peripheral  afferent,  fired  directly 
by  a  receptor,  then  the  necessary  and  sufficient  condition  for  Ni(t) 
is  that  some  one  proposition  or  group  of  propositions  among,  perhaps, 
several  possible  such,  of  the  form  Nj(t-l)  shall  be  true,  where  the 
neurons  c,  are  those  afferents  forming  excitatory  connections  with 
Ci ,  and  that  furthermore  the  propositions  Nk(t—1)  are  all  false,  the 
neurons  ck  being  the  afferents  forming  inhibitory  connections  with 
Ci .  Let  a;  represent  the  class  of  subscripts  corresponding  to  any  set 
of  afferents  which  form  excitatory  connections  with  Ci  and  whose 
summated  impulses  suffice  to  excite  Ci  and  let  k;  represent  the  class  of 
all  such  classes  a*  .  Let  pi  represent  the  class  of  all  subscripts  corre- 
sponding to  all  afferents  forming  inhibitory  connections  with  Ci .  In 
conventional  logical  symbolism,  the  negation  of  Nk(t—1)  is  repre- 
sented by  ™Nk(t— 1)  and  the  joint  negation  for  all  kefii  by 

n  ~zv*(*-i), 

fce/3i 

where  the  symbol  s  here  denotes  class  membership.   A  sufficient  con- 
dition for  the  firing  of  neuron  d  is  then 

n  ~Ak(t-i)  n  Nj(t-i), 

kepi  jeai 

and  the  necessary  and  sufficient  condition  is  the  disjunction  of  all 
such  propositions,  or 

n  ~tf*(*-i)  2  n  Njh-i), 

the  symbols  77  and  2  denoting  conjunction  and  disjunction  respec- 
tively.  If  we  introduce  the  functor  5  defined  by  the  equivalence 

SNi(t)  .  =  .Ni(t-l),  (1) 

then  the  activity  of  the  neuron  d  is  completely  described  by  the  equi- 
valence 

Ni(t)  .  =  .S  n  ™Nk(t)    2      nNj(t).  (2) 

To  consider  examples,  suppose  neuron  c3  has  a  threshold  6  =  2. 
If  Ci  and  c2  each  has  a  single  enclfoot  on  c3  (Figure  lc),  then  c3  fires 
in  any  interval  only  if  c1  and  c2  have  both  fired  in  the  preceding  in- 
terval : 

N3(t)  .  =  .SNAt)  .SNAt). 

On  the  other  hand,  if  cx  and  o2  each  has  two  endfeet  on  c3  (Figure  lb) , 


106   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


a 


<-o- 


< 


< 


L^ 


:i3- 


< — ^ 


< 


J7 


-4 

-4 


■^ 


^ 


Zl    -B& 


Figure  1 


THE  BOOLEAN  ALGEBRA  OF  NEURAL  NETS  107 

then  c3  fires  if  either  cx  or  c2  has  fired: 

N3(t)  .  =  .SNAt)vSN2(t), 

the  symbol  v  denoting  disjunction.  Finally,  if  cx  has  two  endfeet  on 
c3 ,  but  c2  forms  an  inhibitory  connection  with  c3  (Figure  Id),  then 
c3  fires  provided  d  has  fired  and  c2  has  not: 

N3(t)  .  =  .™SN2(t)  .SNAt). 

It  may  be  that  c;  is  itself  a  member  of  a  class  cu  .  For  the  mo- 
ment, however,  we  consider  the  case  when  the  net  contains  no  cycles 
so  that  if  we  pass  successively  from  any  neuron  to  any  of  its  effer- 
ents  in  the  net  we  shall  never  pass  twice  over  the  same  neuron.  Then 
associated  with  each  neuron  Oj  except  the  peripheral  afferents  of  the 
net  there  is  a  single  equivalence  of  the  form  (2),  and  neither  the 
assertion  nor  the  negation  of  Ni(t)  occurs  anywhere  on  the  right  in 
this  equivalence.     If  there  occurs  on  the  right  of  (2)   any  N  (t) , 

where  c.r  is  not  a  peripheral  afferent,  then  N.t  (t)  can  be  replaced  by 

the  right  member  of  the  equivalence  associated  with  c .  .    By  continu- 

ing  sequentially  we  shall  find  ultimately  that  Ni(t)  is  equivalent  to  a 
certain  disjunction  of  conjunctions  of  propositions  of  the  form 
SnNk(t)  and  of  the  negations  of  such,  n  being  everywhere  at  least  1, 
and  every  c^  being  a  peripheral  afferent.  Moreover,  since  every  term 
in  the  disjunction  is  a  sufficient  condition  for  the  firing  of  Ci  at  the 
time  t ,  no  term  in  the  disjunction  can  consist  exclusively  of  nega- 
tions. The  set  of  all  equivalences  of  the  type  just  described  consti- 
tutes a  solution  of  the  net,  since  this  set  contains  the  necessary  and 
sufficient  condition  for  the  firing  at  time  t  of  every  neuron  in  the  net, 
in  terms  of  the  firing  and  non-firing  of  the  peripheral  afferents  at 
earlier  times. 

The  right  member  of  an  equivalence  of  the  type  just  described 
McCulloch  and  Pitts  call  a  temporal  propositional  expression  (abbre- 
viated TPE)  and  it  denotes  a  temporal  propositional  function.  A 
TPE  is  any  sentence  formed  out  of  primitive  sentences  of  the  type 
Ni(t)  operated  upon  any  number  of  times  by  the  operator  S ,  the 
sentences  being  combined  by  disjunctions  and  conjunctions,  as  well 
as  any  sentence  formed  by  conjoining  a  sentence  of  the  foregoing 
type  with  the  negation  of  another  sentence  of  this  type.  Otherwise 
put,  a  TPE  is  any  disjunction  of  conjunctions  of  sentences  SnNj(t) 
and  M  SnNk(t),  except  that  no  conjunction  can  consist  exclusively  of 
negations.  The  importance  of  this  notion  of  a  TPE  lies  in  the  fact 
that  any  TPE  is  "realizable,"  which  is  to  say  that  given  any  TPE,  it 
is  possible  to  describe  a  non-cyclic  net  of  such  a  sort  that  this  TPE 


108   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

expresses  the  necessary  and  sufficient  condition  for  the  firing  of  one 
of  the  neurons  of  this  net.  In  other  words,  the  behavior  of  any  non- 
cyclic  net  can  be  described  exclusively  by  TPE's,  and  conversely  any 
TPE  describes  the  behavior  of  a  neuron  in  some  theoretically  possible 
non-cyclic  net. 

We  give,  as  an  illustration  of  this,  an  example  due  to  McCulloch 
and  Pitts,  the  construction  of  a  neural  net  capable  of  giving  the  illu- 
sion of  heat  produced  by  transient  cooling.  In  this  illusion,  a  cold 
object  held  momentarily  against  the  skin  produces  the  sensation  of 
heat,  whereas  if  this  object  is  held  for  a  longer  time  the  sensation 
is  only  of  cold  with  no  sensation  of  heat,  even  initially.  Heat  and  cold 
are  served  by  different  skin  receptors ;  let  ca  and  c2  be  these  neurons. 
Let  c3  and  c4  be  the  neurons  whose  activity  implies  a  sensation  of  heat 
and  cold.  Then  for  the  firing  of  oa  it  is  necessary  either  that  c*  shall 
have  fired,  or  else  that  c2  fired  momentarily  only,  whereas  for  c4  to 
fire  it  is  necessary  that  c2  shall  fire  for  a  period  of  time. 

These  conditions  are  expressible  symbolically  in  the  f  orm 

Na(t)  .  =  .S{NAt)  vS[SN2(t)  .~N2 (*)]}, 
NM)  .  =  .S[SN2(t)  .N2(t)]. 

For  convenience  we  suppose  that  the  threshold  6  =  2  for  each  neuron, 
which  is  to  say  that  two  endfeet  from  active  neurons  must  connect 
with  any  neuron  in  order  that  it  may  be  made  active.  To  construct 
the  net  we  must  exhibit  connections  between  the  peripheral  alferents 
&x  and  c2  ,  and  the  peripheral  eff  erents  c3  and  c4 ,  by  way,  perhaps,  of 
internuncial  neurons,  of  such  a  sort  that  the  equivalences  hold  as 
given  above.  We  start  with  the  sentence  N2(t)  affected  by  the  great- 
est number  of  operations  S  and  construct  a  neuron  ca  such  that 

Na(t)  .  =  .SN2(t). 

This  requires  two  endfeet  from  Ca  upon  cc .  An  endfoot  from  c2  and 
one  from  ca  each  upon  c4  gives 

N,(t)  .  =  .S[Na(t)  .N2(t)].  =  .S[SN2(l)  .N2(t)] 

which  satisfies  the  second  equivalence.  We  next  introduce  c6  having 
upon  it  two  endfeet  from  ca  and  an  inhibitory  connection  from  c2 , 
giving 

Nt(t)  .  =  .S[Na(t)  .~N2(t)].  =  .SlSN2(t)  .~tfa(*)]. 

Finally  a  pair  of  endfeet  from  c&  upon  c3  and  also  a  pair  from  d  gives 
the  disjunction  on  the  right  of 

N3(t)  .  =  .S[NAt)vNb(t)-\, 


THE  BOOLEAN  ALGEBRA  OF  NEURAL  NETS  109 

and  with  the  substitution  from  the  above  equivalence  for  Nb(t)  the 
construction  is  seen  to  be  complete  (Figure  le).  The  other  diagrams 
of  Figure  1  are  discussed  by  McCulloch  and  Pitts  (1943). 

In  the  case  of  non-cyclic  nets,  the  necessary  and  sufficient  condi- 
tions for  the  firing  of  any  neuron  can  be  stated  exclusively  in  terms 
of  the  behavior  of  the  peripheral  afferents,  i.e.,  the  neurons  of  the  net 
to  which  no  neuron  is  afferent.  Morover,  for  a  given  net  the  requisite 
firing  times  of  the  peripheral  afferents  is  determinate  in  terms  of  the 
firing  times  of  the  eff  erents.  The  introduction  of  cycles,  however,  ren- 
ders the  problems  considerably  more  difficult.  For  one  thing,  a  cycle 
may  be  of  such  a  sort  that  when  activity  is  once  initiated  in  a  suffici- 
ent number  of  the  neurons  it  will  continue  indefinitely.  For  the  de- 
tails in  the  discussion  of  this  case  reference  must  be  made  to  Mc- 
Culloch and  Pitts  (1943).  However,  there  is  in  every  cyclic  net  a 
certain  minimal  number  of  neurons  whose  removal  would  render  the 
net  non-cyclic.  This  number  is  called  the  order  of  the  net,  so  that  a 
non-cyclic  net  has  order  zero.  Then  the  behavior  of  the  net  is  deter- 
mined by  the  behavior  of  the  set  which  consists  of  these  neurons  and 
the  peripheral  afferents,  and  the  problem  reduces  to  a  consideration 
of  the  neurons  of  this  set. 

The  most  striking  difference  between  cyclic  and  non-cyclic  nets, 
with  reference  to  the  types  of  propositions  which  occur  among  the 
conditions  for  firing  of  a  neuron,  lies  in  this,  that  universal  and  exist- 
ential propositions  may  arise  for  cyclic  nets.  Thus  the  net  consist- 
ing of  a  neuron  c2  whose  threshold  is  2  ,  and  a  peripheral  afferent 
Cj  in  which  d  and  c2  each  has  a  single  endfoot  upon  c2 ,  realizes  the 
universal  sentence: 

N*(t)  .  =  .  (z)t.SN1(z). 

This  is  a  symbolic  formulation  of  the  assertion  that  for  c2  to  fire  at 
the  time  t  it  is  necessary  and  sufficient  that  cx  must  have  fired  in 
every  interval  z  prior  to  t .  An  objection  is  immediate  at  this  point, 
to  the  effect  that  the  theory  provides  no  mechanism,  therefore  by 
which  the  circuit  can  ever  get  started,  and  it  must  be  admitted  that 
within  the  net  in  question  there  is  no  such  mechanism.  However,  it 
is  to  be  understood  that  the  theory  concerns  the  behavior  of  the  net 
while  free  from  all  disturbing  outside  influences  except  specified  ones 
of  a  specified  type  (stimulation  of  peripheral  afferents).  That  is,  it 
is  a  theory  of  a  system  in  isolation,  as  any  theory  must  be.  Hence  the 
assertion  is  to  be  taken  to  mean  that  the  circuit  being  once  started 
in  an  unspecified  fashion  and  the  system  being  then  placed  in  "isola- 
tion," the  circuit  can  be  maintained  only  by  continued  stimulation  of 


110   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

If  Cj  and  c2  each  has  two  endfeet  upon  c2 ,  the  existential  sen- 
tence is  realized: 

N*(t)  .  =  .  (Ez)t.SNi(z). 

This  is  a  symbolic  formulation  of  the  assertion  that  c^  will  fire  at  the 
time  t  provided  N^  has  fired  in  some  interval  z  prior  to  t .  For  in 
this  case  the  cycle  formed  by  c2 ,  being  once  set  into  activity,  will 
continue  in  a  state  of  permanent  activity.  In  either  case  we  could 
replace  iVj  (t)  by  any  TPE,  and  the  neuron  d  by  the  net  which  realizes 
this  TPE.  Again,  in  the  case  of  the  "existential"  net,  if  instead  of 
the  single  self-exciting  neuron  c2  we  have  a  cycle  of  n  neurons,  then, 
activity  being  once  started,  each  neuron  of  the  cycle  will  be  active 
subsequently  once  every  n  time-intervals,  and  not  in  every  time-in- 
terval. 

It  is  understood,  of  course,  that  some  common  features  of  neu- 
ronal activity  are  left  out  of  the  formalized  picture  given  here.  On 
the  other  hand,  some  of  these  can  be  given  a  formal  representation  in 
these  terms,  such  a  formal  representation  having  by  no  means,  how- 
ever, the  purport  of  factual  description  or  explanation  of  the  phe- 
nomenon in  question.  Thus  learning  could  be  described  in  terms  of 
activated  cycles  and  facilitation  in  terms  of  internuncials  in  the  net. 
Extinction  can  be  described  in  terms  of  relative  inhibition,  and  ref- 
erence has  already  been  made  to  the  fact  that  relative  inhibition  can, 
in  turn,  be  replaced  formally  by  total  inhibition  in  nets  of  somewhat 
different  structure. 


XV 

A  STATISTICAL  INTERPRETATION 

The  time-units  in  the  Boolean  formulation  are  of  the  order  of 
milliseconds,  the  approximate  minimal  separation  between  consecu- 
tive impulses  in  any  neuron  being,  in  fact,  about  half  this  unit.  Hence, 
for  even  the  briefest  of  overt  responses,  which  require  generally  a 
duration  of  seconds  or  longer,  there  is  time  for  the  occurrence  of 
hundreds  or  even  thousands  of  individual  impulses  on  the  part  of 
any  one  of  the  participating  neurons.  It  is  legitimate,  therefore,  and 
necessary,  to  develop  a  statistical  theory  of  the  temporal  distribution 
of  these  impulses  for  application  to  the  temporally  macroscopic  psy- 
chological processes. 

Whether  this  statistical  development  will  lead  immediately  to 
quantities  that  can  be  interpreted  as  being  the  postulated  e  and  j  of 
the  Rashevsky  theory  is  still  to  be  seen.  It  is  quite  possible,  indeed, 
that  £  and  j  must  rather  be  regarded  as  the  statistical  effects  of  im- 
pulses in  certain  groups  of  neurons,  rather  than  individual  neurons. 
As  Rashevsky  (1940,  chap,  viii)  has  emphasized,  while  the  postulated 
development  of  e  and  j  is  suggested  by  direct  observation  of  the  ac- 
tion of  individual  neurons,  it  is  necessary  to  suppose  only  that  there 
are  neural  units  or  groups  of  some  kind  whose  activity  follows  the 
postulated  pattern,  and  the  interpretation  in  terms  of  groups  is 
strongly  suggested  by  the  paucity  of  hypothetical  neurons  required 
to  account  for  many  of  the  psychological  processes  discussed  in  earlier 
chapters  of  this  monograph.  Moreover,  the  predictive  success  of  the 
theory  provides  strong  presumptive  evidence  in  favor  of  the  supposi- 
tion that  the  development  of  e  and  j  as  postulated  does  correspond  to 
some  basic  physiological  processes,  whatever  these  may  be. 

As  a  step  toward  the  deduction  of  the  macroscopic  from  the  mi- 
croscopic theory,  Landahl,  McCulloch,  and  Pitts  (1943)  have  shown 
how  the  propositional  equivalences  can  be  transformed  immediately, 
by  well-defined  formal  replacements,  into  statistical  relations.  To 
obtain  the  rule  for  making  these  replacements,  we  now  let  d  repre- 
sent the  period  of  latent  addition,  the  interval  within  which  the  mini- 
mal number  of  converging  impulses  must  occur  in  order  to  result  in 
a  stimulation.  If  v}  is  the  mean  frequency  of  the  impulses  in  the  neu- 
ron Cj  afferent  to  the  neuron  Ci ,  then  Vjd  is  the  probability  that  an 
impulse  will  occur  in  c,  during  any  particular  interval  6  ,  and  1  —  v}d 
is  the  probability  that  an  impulse  will  not  occur  in  ck .  If  the  im- 
pulses in  the  various  neurons  c;  concurrent  upon  d  are  statistically  in- 
Ill 


112    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

dependent,  the  probability  that  impulses  will  arrive  concomitantly 
(that  is,  within  an  interval  6)  along  any  group  of  r  of  these  neurons  is 
equal  to  the  product  Sr  JJ  Vj ,  there  being  r  factors  Vj .  Likewise  the 
probability  that  no  impulse  will  arrive  along  any  one  of  a  group  of  s 
(inhibitory)  neurons  ck  is  the  product  77(1  —  S  vk),  there  being  here 
s  factors.  And  if  the  group  of  r  neurons  c,  is  sufficient  to  excite  Ci , 
the  probability  that  c,  will  be  excited  in  any  interval  <3  by  the  neurons 
Cj  is  the  product  of  r  +  s  factors. 

d'nvjlia-dn)  (1) 

if  the  s  neurons  c^  include  all  inhibitory  afferents  to  c;  and  if  the  im- 
pulses are  statistically  independent.  If  we  then  form  the  product  such 
as  (1)  corresponding  to  every  group  of  neurons  c,  sufficient  to  fire 
Ci  by  their  concomitant  activity  in  the  absence  of  inhibitory  impulses, 
the  probability  that  d  will  fire  is  given  approximately  by  summing 
all  products  so  formed.  Thus  with  the  same  notation  as  that  em- 
ployed for  the  equivalence  (2)  of  the  preceding  chapter,  we  find  that 
the  frequency  vi  with  which  c,  fires  is  given  by  the  equation 

dvi  =  n  a-svk)  2    *r(ai)  n  vj,  (2) 

where  r(cti)  is  the  number  of  neurons  in  the  set  ai .  But  if  we  com- 
pare this  equation,  and  the  manner  in  which  it  was  formed,  with  the 
equivalence  (2),  chapter  xii,  we  see  that  the  two  sides  of  the  equiva- 
lence become  identical  with  the  two  sides  of  the  equation  when  each 
assertion  N  is  replaced  by  the  corresponding  dv ,  and  each  negation 
°°N  by  the  corresponding  (1  —  dv). 

However,  as  we  have  remarked,  this  expression  is  approximate 
only.  Each  product  6r  77  v}  is  equal  to  the  probability  that  at  least  all 
the  neurons  in  the  particular  set  of  r  neurons  Cj  will  fire  in  the  time- 
interval  S  ,  but  the  possibility  that  additional  excitatory  afferents  may 
also  fire  is  not  ruled  out.  Duplication  is  therefore  possible  and  the 
sum  would  then  be  too  large.  To  give  a  concrete  example,  suppose 
there  are  only  two  afferents,  c1  and  c2 ,  both  excitatory,  and  each  alone 
capable  of  stimulating  c3 .  Then  c3  may  be  excited  by  the  firing  of  (h 
alone,  by  the  firing  of  c2  alone,  or  by  the  concomitant  firing  of  c^  and 
c2 .   Now  by  the  formula  we  obtain  the  probability 

d  vA  =  6  v-l  +  6  Vi  , 

whereas  the  true  probability  is 

d  v3  =  d  V!  ( 1  —  d  v2)  +  <5  v2  ( 1  ~  <5  v-i )  +  b  v1  d  v2 

=  6  v-i  +  6  v-2  —  S2  v-i  v-2  , 

the  first  formulation  exhibiting  in  detail  the  probability  of  the  three 


A  STATISTICAL  INTERPRETATION  113 

exclusive  events  each  sufficient  for  the  occurrence  of  the  firing  of  C3 . 

Hence  a  more  exact  formulation  in  place  of  equation  (2)  would 

involve  including  with  each  product  dr  II  v,  also  the  product  of  all 

factors  of  the  form  (1  —  d  v.  )  corresponding  to  all  excitatory  aifer- 

ents  c    not  included  in  the  set  of  the  c/s .   However ,  since  d  is  a  small 

quantity,  the  terms  omitted  in  equation  (2)  are  terms  of  a  higher  or- 
der and  in  general  negligible.  In  fact,  to  the  same  degree  of  approxi- 
mation we  can  pick  out  the  smallest  among  the  r(ai)  and  neglect  all 
terms  in  the  sum  involving  a  larger  r .  We  must  suppose,  however, 
that  the  sum  of  the  vk  is  large  by  comparison  with  the  sum  of  the  v, 
for  any  set  at ,  since  otherwise  the  effect  of  the  inhibition  also  would 
be  negligible. 

With  this  understanding  we  find,  by  straightforward  induction, 
that  the  following  rules  of  replacement  enable  us  to  transform  a  logi- 
cal equivalence  with  regard  to  neural  nets  into  a  probability-relation: 

(1)  Replace  each  assertion  N(t)  by  6  v(t)  with  the  same  sub- 
script ; 

(2)  Replace  each  negation  ^Nit)  by  [1  —  S  v(t)]  giving  to 
v  the  subscript  of  the  N ; 

(3)  Replace  logical  disjunction  and  conjunction  by  arithmetic 

addition  and  multiplication; 

t  t 

(4)  Replace  the  operators  (z)t  and  (Ez)t  by  77  and  2  respee- 

t=0  (=0 

tively ; 

(5)  Where  a  function  of  t  is  preceded  by  an  operator  Sa,  re- 
place the  argument  t  by  t  —  a  6  . 

The  factor  d  can  be  everywhere  omitted  when  the  period  of  latent 
addition  is  taken  as  the  unit  of  time. 

We  note  in  conclusion  that  with  the  approximation  made  here  the 
equation  (2)  can  be  expanded  to  the  form 

dn  =  <5r(l-<5  2  vk)     2      n  vi9  (3) 

ks^i  aiEKi(r)  jeai 

where  r  is  the  smallest  of  the  r(aO,  and  Ki(r)  includes  only  those 
classes  at  which  contain  r  neurons.  On  removing  the  parentheses  and 
multiplying  out  we  obtain  two  terms,  the  first  essentially  positive,  the 
second  essentially  negative.  The  first  we  can  therefore  interpret  as 
the  e  ,  the  second  as  the  j  of  the  present  theory. 


CONCLUSION 

The  measure  of  success  of  a  predictive  theory  depends  upon  at 
least  three  factors:  the  simplicity  of  the  theory,  the  range  of  its  appli- 
cations, and  its  accuracy.  It  is  quite  possible  for  both  of  two  incom- 
patible theories  to  be  in  common  use  when  one  is  much  simpler,  the 
other  more  extensive  or  more  exact  in  its  applications.  This  is  the 
case  with  the  gravitational  theories  of  Newton  and  Einstein.  While 
the  measure  of  a  theory's  simplicity  is  evident  to  any  competent  read- 
er, the  test  of  its  range  and  accuracy  must  be  the  work  of  the  experi- 
menter. Thus  the  success,  and  also  the  improvement,  of  a  scientific 
theory  rests  quite  as  much  upon  the  experimenter  as  upon  the  theorist. 

The  theories  presented  in  this  monograph  point  two  ways.  On 
the  one  hand,  they  provide  for  the  prediction  of  behavior.  On  the 
other  hand,  they  presuppose  the  presence  of  specific  anatomical  struc- 
tures. Such  a  theory  might  give  a  very  accurate  representation  of 
behavior  without  necessarily  deriving  from  a  correct  representation 
of  the  actual  mediating  stucture.  If  so,  it  is  no  less  useful  in  the  one 
direction  for  having  failed  in  the  other,  though  it  is  to  be  expected 
that  further  investigation  should  lead  to  a  theory  that  is  more  success- 
ful in  both  respects. 

The  advantage  of  having  a  theory  lies  not  only  in  the  fact  that 
the  theory,  if  it  should  prove  accurate  in  its  predictions,  provides 
just  that  much  of  an  increase  to  our  fund  of  knowledge,  or  increases 
by  just  that  much  our  command  over  nature.  A  well-formulated  the- 
ory, in  addition,  gives  direction  and  meaning  to  experimental  work 
even  when  the  experiments  fail  to  confirm  the  theory,  since  the  fail- 
ure must  occur  in  a  certain  direction,  by  a  certain  amount,  and  must 
therefore  provide  the  necessary,  hitherto  lacking,  background  for  the 
construction  of  a  better  theory. 

Furthermore,  theoretical  organization  gives  unity  and  coherence 
to  the  otherwise  multifarious  and  apparently  disparate  phenomena 
of  observation.  It  has  been  remarked  (Williams,  1927)  that  the  ele- 
mentary physics  textbooks  have  diminished  considerably  in  size  dur- 
ing recent  decades  in  spite  of  the  tremendous  increase  in  our  physical 
information,  just  because  details  formerly  treated  separately  are  more 
and  more  brought  together  as  special  cases  of  a  few  general  princi- 
ples. In  our  brief  account  we  have  indicated  here  and  there  (cf.  in 
particular  chap,  vi)  how  quite  different  forms  of  response  could  be 
mediated  by  structures  of  the  same  kind  and  are  thus  to  be  subsumed 
under  the  same  principle.  Doubtless  many  other  instances  could  be 
found,  and  will  be  as  the  work  progresses.  Thus  it  is  to  be  hoped  that 
as  the  formulations  are  improved  the  same  progress  will  be  observ- 

114 


CONCLUSION  115 

able  in  the  textbooks  of  psychology  as  in  those  of  physics,  and  the 
extent  to  which  this  may  occur  will  depend  entirely  upon  the  extent 
to  which  theorist  and  experimenter  cooperate  in  their  efforts  to  bring 
it  about.  On  the  other  hand,  it  is  neither  to  be  expected  nor  to  be 
desired  that  quantitative  formulations  can  in  any  sense  represent  all 
of  the  complexities  of  any  psychological  process,  or  a  fortiori,  the 
true  "inwardness"  of  conscious  life. 

If  this  brief  summary  of  theoretical  developments  to  date  leaves 
the  reader  with  a  feeling  of  incompleteness,  as  of  a  story  interrupted 
before  it  is  well  under  way,  this  is  only  as  it  should  be.  For  the  authors 
the  hardest  task  in  presenting  each  topic  has  been  to  leave  it  incom- 
plete rather  than  to  postpone  the  writing  of  it  while  pursuing  it  fur- 
ther. In  general,  we  have  summarized  each  topic  as  far  as  it  has  yet 
been  developed.  On  the  other  hand,  we  have  had  to  leave  out  many 
things  that  might  have  been  included.  To  mention  a  few  of  these, 
Landahl  (1940)  has  given  an  interpretation  of  factor-loadings  in 
terms  of  time-scores;  and  he  has  suggested  (Landahl,  1939)  a  pos- 
sible solution  of  the  "bottle-neck"  problem  of  the  visual  pathway; 
Coombs  (1941)  has  discussed  the  galvanic  skin  response;  Household- 
er (1939)  has  discussed  a  mechanism  for  a  certain  Gestalt-phenome- 
non;  Householder  and  Amelotti  (1938)  have  outlined  a  theory  of  the 
delay  in  the  conditioned  reflex.  Rashevsky  (1938)  has  outlined  sev- 
eral theories  of  rational  thinking.  Pitts  (1943)  has  given  a  formal 
theory  of  learning  and  conditioning,  and  Lettvin  and  Pitts  (1943) 
have  developed  a  formal  theory  of  certain  psychotic  states.  While 
neither  of  these  last  two  theories  provides  an  explicit  neural  mech- 
anism, both  carry,  by  implication,  fairly  clear  suggestions  for  these 
and  their  construction  is  in  progress.  Finally,  Rashevsky's  work  on 
visual  aesthetics,  referred  to  in  the  introduction,  is  now  the  subject 
of  extensive  experimental  test  with  very  satisfactory  preliminary  re- 
sults. Only  part  of  this  theory  has  been  published  as  yet. 

While  Rashevsky  is  actively  developing  his  theory  of  aesthetics, 
additional  work  is  now  under  way  and  soon  to  be  published  dealing 
with  shock  and  certain  psychotic  mechanisms,  with  apparent  move- 
ment, and  with  further  elaboration  of  the  theory  of  color-vision. 

But  all  this  together— the  work  accomplished  and  the  work  under 
way— touches  only  a  very  small  part  of  the  field  of  psychology,  and 
the  theories  as  we  have  them  now  represent  at  best  a  number  of  pre- 
liminary attempts,  many  of  them  still  untested.  We  do  not  present 
them  as  finished  products.  Quite  the  contrary,  no  one  would  be  more 
surprised — or  disappointed !— than  the  authors  if  a  decade  hence  does 
not  find  every  one  of  these  theories  considerably  modified  if  not 
wholly  discarded  for  a  better. 


LITERATURE 

Bartley,  S.  H.   1941.    Vision:  A  Study  of  Its  Basis.   New  York:  D.  Van  Nostrand 

Co.,  Inc. 
Bichowsky,  F.  Russel.    1941.    The  anatomy  of  the  cortex  as  derived  from  per- 
ception.   Privately  circulated. 
Brock,  Frederick  W.   1940.   The  cortical  mechanism  of  binocular  vision  in  normal 

and  anomalous  projection.   Am.  Jour.  Optometry,  17,  460-472. 
Carlson,  Anton  J.,  and  Victor  Johnson.    1941.    The  Machinery  of  the  Body.   2nd. 

edition.    Chicago:  The  University  of  Chicago  Press. 
Cattell,  J.   McK.    1886.    The  influence  of  the  intensity  of  the  stimulus  on  the 

length  of  the  reaction  time.   Brain,  8,  512-515. 
Chodin,  A.   1877.  1st  das  Weber-Fechnersche  Gesetz  auf  das  Augenmaass  anwend- 

bar?   Graefes  Arch.  Fur  Ophth.  23,  92-108. 
Coombs,  Clyde  H.    1941.    Mathematical  biophysics  of  the  galvanic  skin  response. 

Bull.  Math.  Biophysics,  3,  97-103. 
Douglas,  A.  C.    1932.    The  Physical  Mechanism  of  the  Human  Mind.    Edinburgh : 

E.  &  S.  Livingstone. 
Grant,  Vernon  W.    1942.    Accommodation  and  convergence  in  visual  space  per- 
ception.  J.  Exper.  Psychol.,  31,  89-104. 
Guilford,  J.  P.   1936.   Psychometric  Methods.   New  York :  McGraw-Hill  Book  Co., 

Inc. 
Gulliksen,  Harold.    1934.    A   rational   equation  of  the  learning  curve  based  on 

Thorndike's  law  of  effect.    J.  Gen.  Psychol,  11,  395-434. 
Gyemant,  R.    1925.    Grundziige  der  Kolloidphysik  vom  Standptinkte  des  Gleich- 

gewichts.   Braunschweig:  Friedr.   Vieweg  u.  Sohn. 
Hartline,  H.  K.,  and  C.  H.  Graham.    1932.    Nerve  impulses  from  single  receptors 

in  the  eye.   J.  Cell,  and  Comp.  Physiol.,  1,  277-295. 
Hecht,  S.    1920.   Intensity  and  the  process  of  photoreception.   Jour.  Gen.  Physiol., 

2,  337-347. 

Helmholtz,  H.  von.  1896.  Handbuch  der  physiologischen  Optik.  Hamburg  u. 
Leipzig:  Verlag  von  Leopold  Voss. 

Holway,  Alfred  H.,  Janet  E.  Smith  and  Michael  J.  Zigler.  1937.  On  the  dis- 
crimination of  minimal  differences  in  weight:  II.  Number  of  available  ele- 
ments as  variant.    Jour.  Exper.  Psychol.,  20,  371-380. 

Householder,  Alston  S.    1938a.    Excitation  of  a  chain  of  neurons.   Psychometrika, 

3,  69-73. 

Householder,  Alston  S.    1938b.    Conditioning  circuits.    Psychometrika,  3,  273-289. 

Householder,  Alston  S.  1939a.  Concerning  Rashevsky's  theory  of  the  Gestalt. 
Bidl.  Math.  Biophysics,  1,  63-73. 

Householder,  Alston  S.  1939b.  A  neural  mechanism  for  discrimination.  Psycho- 
metrika, 4,  45-58. 

Householder,  Alston  S.  1940a.  A  neural  mechanism  for  discrimination  II.  Dis- 
crimination of  weights.   Bull.  Math.  Biophysics,  2,  1-13. 

Householder,  Alston  S.  1940b.  A  note  on  the  horopter.  Bull.  Math.  Biophysics,  2, 
135-140. 

Householder,  Alston  S.  1940c.  A  neural  mechanism  for  discrimination  III.  Vis- 
ually perceived  lengths  and  distances.   Bull.  Math.  Biophysics,  2,  157-167. 

Householder,  Alston  S.  1943.  A  theory  of  the  induced  size  effect.  Bull.  Math. 
Biophysics,  5,  155-160. 

116 


LITERATURE  117 

Householder,  Alston  S.,  and  E.  Amelotti.  1937.  Some  aspects  of  Rashevsky's 
theory  of  delayed  reflexes.   Psychometrika,  2,  255-262. 

Householder,  A.  S.,  and  Gale  Young.  1940.  Weber  laws,  the  Weber  law  and  psy- 
chophysical analysis.    Psychometrika,  5,  183-194. 

Kellogg,  W.  N.  1930.  An  experimental  evaluation  of  equality  judgments  in  psy- 
chophysics.    Arch.  Psychol.,  No.  112. 

Konig,  Arthur,  and  Eugen  Brodhun.  1888,  1 889.  Experimented  Untersuchungen 
iiber  die  psychophysische  Fundamentalformel  in  Bezug  auf  den  Gesichtssinn. 
Sitz.  d.  k.  Pr.  Ak.  d.  Wiss.   Berlin,  1888,  917-931;  1889,  641-644. 

Kubie,  Lawrence  S.  1930.  A  theoretical  application  to  some  neurological  prob- 
lems of  the  properties  of  excitation  waves  which  move  in  closed  circuits. 
Brain,  53,  166-177. 

Landahl,  H.  D.  1938a.  A  contribution  to  the  mathematical  biophysics  of  psycho- 
physical discrimination.   Psychometrika,  3,  107-125. 

Landahl,  H.  D.  1938b.  The  relation  between  the  intensity  of  excitation  and  the 
number  of  neurons  traversed.   Psychometrika,  3,  291-295. 

Landahl,  H.  D.  1939a.  Contributions  to  the  mathematical  biophysics  of  the  cen- 
tral nervous  system.   Bull.  Math.  Biophysics,  1,  95-118. 

Landahl,  H.  D.  1939b.  A  contribution  to  mathematical  biophysics  of  psycho- 
physical discrimination  II.   Bull.  Math.  Biophysics,  1,  159-176. 

Landahl,  H.  D.  1940a.  Discrimination  between  temporally  separated  stimuli. 
Bull.  Math.  Biophysics,  2,  37-47. 

Landahl,  H.  D.    1940b.   Time  scores  and  factor  analysis.   Psychometrika,  5,  67-74. 

Landahl,  H.  D.  1940c.  A  contribution  to  the  mathematical  biophysics  of  psycho- 
physical discrimination  III.    Bull.  Math.  Biophysics,  2,  73-87. 

Landahl,  H.  D.  1941a.  Studies  in  the  mathematical  biophysics  of  discrimination 
and  conditioning  I.    Bull.  Math.  Biophysics,  3,  13-26. 

Landahl,  H.  D.  1941b.  Studies  in  the  mathematical  biophysics  of  discrimination 
and  conditioning  II.  Special  case:  errors,  trials,  and  number  of  possible  re- 
sponses.  Bull.  Math.  Biophysics,  3,  63-69. 

Landahl,  H.  D.  1941c.  Theory  of  the  distribution  of  response  times  in  nerve 
fibers.    Bull.  Math.  Biophysics,  3,  141-147. 

Landahl,  H.  D.  1943.  Studies  in  the  mathematical  biophysics  of  discrimination 
and  conditioning:  III.  Bull.  Math.  Biophysics,  5,  103-110. 

Landahl,  H.  D.  and  A.  S.  Householder.  1939.  Neuron  circuits;  the  self-exciting 
neuron.    Psychometrika,  4,  255-267. 

Landahl,  H.  D.,  W.  S.  McCulloch  and  Walter  Pitts.  1943.  A  statistical  conse- 
quence of  the  logical  calculus  of  nervous  nets.  Bull.  Math.  Biophysics,  5, 
135-137. 

Lettvin,   Jerome   Y.   and   Walter   Pitts,   1943.    A   mathematical   theory   of   the 
affective  psychoses.   Bull.  Math.  Biophysics,  5,  139-148. 

Lorente  de  No,  Rafael.  1933.  Vestibulo-ocular  reflex  arc.  Arch.  Neur.  and  Psy- 
chiatry, 30,  245-291. 

McCulloch,  Warren  S.  and  Walter  Pitts.  1943.  A  logical  calculus  of  the  ideas 
immanent  in  nervous  activity.   Bull.  Math.  Biophysics,  5,  115-133. 

Macdonald,  P.  A.,  and  D.  M.  Robertson.  1930.  The  psychophysical  law. — III.  The 
tactile  sense.    Phil.  Mag.,  10,  1063-1073. 

Matthews,  Bryan  H.  C.  1933.  Nerve  endings  in  mammalian  muscle.  J.  Physiol., 
78,  1-53. 

Ogle,  Kenneth  N.  1938.  Die  mathematische  Analyse  des  Langshoropters.  Arch, 
f.  d.  ges.  Physiol,  239,  748-766. 

Ogle,  Kenneth  N.   1940.   The  induced  size  effect.   Jour.  Opt.  Soc.  Am.,  30,  145-151. 


118   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 

O'Leary,  James  Lee.  1937.  Structure  of  the  primary  olfactory  cortex  of  the 
mouse.    J.  Comp.  Neur.,  67,  1-31. 

Pecher,  C.  1939.  La  fluctuation  d'excitabilite  de  la  fibre  nerveuse.  Arch.  In- 
ternat.  Physiol,  49,  129-152. 

Peddie,  W.    1922.    Colour  Vision.    London:  Edward  Arnold  &  Co. 

Pieron,  H.  1920.  Nouvelles  recherches  sur  l'analyse  du  temps  de  latence  sen- 
sorielle  et  sur  la  loi  qui  relie  le  temps  a  1'intensibe  d'excitation.  L' 'Annie 
Psydhologique,  22,  58-142. 

Pitts,  Walter.  1942a.  Some  observations  on  the  simple  neuron  circuit.  BulL 
Math.  Biophysics,  4,  121-129. 

Pitts,  Walter.  1942b.  The  linear  theory  of  neuron  networks:  The  static  prob- 
lem.  Bull.  Math.  Biophysics,  4,  169-175. 

Pitts,  Walter.  1943a.  The  linear  theory  of  neuron  networks:  The  dynamic 
problem.    Bull.  Math.  Biophysics,  5,  23-31. 

Pitts,  Walter.  1943b.  A  general  theory  of  learning  and  conditioning.  Psycho- 
metrika,  8,  1-18,  131-140. 

Rashevsky,  N.  1938.  Mathematical  Biophysics.  Physicomathematical  Founda- 
tions of  Biology.    Chicago:  The  University  of  Chicago  Press. 

Rashevsky,  N.  1940.  Advances  and  Applications  of  Mathematical  Biology.  Chi- 
cago: The  University  of  Chicago  Press. 

Rashevsky,  N.  1942a.  Suggestions  for  a  mathematical  biophysics  of  auditory 
perception  with  special  reference  to  the  theory  of  aesthetic  ratings  of  com- 
binations of  musical  tones.   Bull.  Math.  Biophysi.es,  4,  27-32. 

Rashevsky,  A.  1942b.  An  alternate  approach  to  the  mathematical  biophysics  of 
perception  of  combinations  of  musical  tones.   Bull.  Math.  Biophysics,  4,  89-90. 

Rashevsky,  N.  1942c.  Further  contributions  to  the  mathematical  biophysics  of 
visual  aesthetics.    Bull.  Math.  Biophysics,  4,  117-120. 

Rashevsky,  N.  1942d.  Some  problems  in  mathematical  biophysics  of  visual  per- 
ception and  aesthetics.    Bull.  Math.  Biophysics,  4,  177-191. 

Riesz,  R.  R.  1928.  Differential  intensity  sensitivity  of  the  ear  for  pure  tones. 
Physical  Rev.,  31,  867-875. 

Sacher,  George.  1942.  Periodic  phenomena  in  the  interaction  of  two  neurons. 
Bull.  Math.  Biophysics,  4,  77-81. 

Shaxby,  John  H.  1943.  On  sensory  energy,  with  special  reference  to  vision  and 
colour-vision.    Phil.  Mag.,  34,  289-314. 

Stanton,  Henry  E.  1941.  A  neural  mechanism  for  discrimination  IV:  Monocular 
depth  perception.  Bull.  Math.  Biophysics,  3,  113-120'. 

Thurstone,  L.  L.  1927.  A  law  of  comparative  judgment.  Psychol.  Rev.,  34, 
273-286. 

Thurstone,  L.  L.  1935.  Vectors  of  Mind.  Chicago:  The  University  of  Chicago 
Press. 

Urban,  F.  M.  1908.  The  Application  of  Statistical  Methods  to  the  Problem  of 
Psychophysics.   Philadelphia:  Psychol.  Clinic  Press. 

Verhoeff,  F.  N.  1925.  A  theory  of  binocular  perspective.  Am.  Jour.  Physiol.  Op- 
tics, 6,  416-448. 

Williams,  H.  B.  1927.  Mathematics  and  the  biological  sciences.  The  Gibbs  Lec- 
ture for  1926.   Bull.  Am.  Math.  Soc,  33,  273-293. 

Woodrow,  H.  1914.  The  measurement  of  attention.  Psychol.  Rev.  Mono.,  17, 
No.  5. 

Young,  Gale,  and  Alston  S.  Householder.  1941.  A  note  on  multi-dimensional 
psychophysical  analysis.    Psychometrika,  6,  331-333. 


INDEX 


Absolute  discrimination,  66,  72 
Absolute  thresholds,  65 
Accessibility,  10,  11,  12 
Accessible,  10,  31 
Accommodation,  48,  97,  98 
Action-spikes,  103 
Activity 

continuous,  26 

maximal,  11 

nervous,  vii 

permanent,  22,  29 

steady-state,  7,  67,  81 
Activity  parameters,  17,  19 
Adaptation,  42,  72 

to  muscular  stretch,  50 
Aesthetic  judgments,  74 
Afferent  chains,  74 
Afferent  neuron,  firing  of,  104 
All-or-none  character,  103 
Amelotti,  E.,  115 
Analysis 

factor,  75 

multidimensional    psychophysical, 
74,76 
Anatomy,  vii 
Anomalous  fusion,  99 
Apparent  depth,  96 
Apparent  size,  96,  97 
Applied  stimulation,  11 
Applied  stimulus,  9 
Asymptotic  value,  7,  25 
Asymptotically  exciting  neuron,  5 
Asymptotically   inhibiting   neuron,   5, 

6,  9 
Auditory  data,  62,  69 
Auditory  sensations,  68,  69,  70 
Auditory  stimuli,  40 
Awareness  of  stimuli,  16 

Bartley,  S.  H.,  45,  48 

Behavior,  vii,  1 

Berger,  G.  O.,  39,  55 

Bichowsky,  F.  R.,  98 

Binocular  cue,  71,  97 

Binocular  disparity,  98 

Binocular  stereopsis,  97 

Binocular  vision,  97 

Binocular  visual  field,  98 

Blind  alley,  85 

Boolean  algebra  of  neural  nets,  103 

statistical  interpretation  of,  111 
Brightness 

constant,  45 

judgment  of,  6 

relative,  48 
Brock,  F.  W.,  99 
Brodhum,  E.,  68 


Carlson,  A.  J.,  90 


Categorical  judgment,  58,  60 
Cattell,  J.  McK.,  39,  55 
Centers,  inaccessibility  of,  12 
Chain  of  neurons,  8,  9,  11,  12,  22,  31, 

35 
Chain  of  two  neurons,  38,  41,  46 
Chains,  afferent,  74 
Chains,  interconnected,  56,  74 
Chains  of  neurons,  7 
Chains,  two-neuron,  79 
Chodin,  A.,  71 
Circuit,  24,  28,  31,  81 

two-neuron,  25 
Circuits,  66,  78,  79,  83 

self -exciting,  12 

simple,  22 
Color-center,  91,  92 
Color-contrast,  13 
Color-perception,  21 
Color-pyramid,  92 
Color-vision,  theory  of,  90 
Colors,  92 

discrimination  of,  72,  91 
Common  synapse,  49 
Common  terminus,  13 
Complete  inhibition,  104 
Complex  stimulus-object,  74 
Conditioned  stimuli,  83 
Conditioned  stimulus,  81 
Conditioning,  79,  81 
Conduction  time,  22,  31,  33,  104 
Conjunction,  105 
Connections,  excitatory,  105 
Constancy  of  perceived  size,  97 
Constant  brightness,  45 
Constant  stimulus,  42,  48,  49,  54 
Continuous  activity,  26 
Contraction,  muscular,  1,  4 
Convergence,  97,  98 
Coombs,  C.  H.,  115 
"Correct"  judgments,  72 
Correct  response,  58,  61,  83,  87 

probability  of,  58,  59,  89 
Corresponding  points,  98 
Cortical  image,  97 
Critical  flicker-frequency,  45,  46 
Critical  frequency,  44,  46 
Critical  value,  8 
Crossing  excitation,  65,  78 
Crossing  inhibition,  65,  74,  76,  78 
Cutaneous  receptors,  64 
Cycles,  107 
Cyclic  net,  109 
Cylindrical  size-lens,  100,  102 

Delay  of  reaction,  39,  40 
Depth,  apparent,  96 
Discharge,  nervous,  104 
Discriminal  response,  75 


119 


]20    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


Discriminal  sequence,  64,  74 
Discriminal  sequences,  56 
Discriminal  stimulus  sequences,  21 
Discriminating  mechanism,  91 
Discrimination 

absolute,  66,  72 

intensity,  69,  70 

of  colors,  72,  91 

of  lengths,  69,  71 

of  lifted  weights,  73 

of  weights,  71 

psychophysical,  56 

relative,  66,  72 
Disjunction,  105 

neural,  95 
Disparate  point,  98 
Disparities 

horizontal,  100,  101 

vertical,  100 
Disparity,  98 
Distance-cue,  96,  97 
Distance,  interocular,  101 
Distinct  response,  68 
"Doubtful"  judgments,  59,  60 
Douglas,  A.  C,  82,  94 
Duration 

of  reaction,  41 

of  response,  41,  42 
Dynamics,  22 

intra-neuronal,  3 

neuronal,  viii 

of  simple  circuits,  22 

of  single  synapse,  37,  48 

trans-synaptic,  1 

Effective  stimulus,  5 

Effector,  vii,  7,  13,  78,  90 

Effectors,  1 

Effects,  modulating,  22 

Endfeet,  104 

"Equality",  judgment  of,  59,  61 

Equilibria,  fluctuating,  35 

Equilibrium,  27,  28 

fluctuating,  29 

stable,  23,  26,  35 

unstable,  24,  35 
Errors,  number  of,  85,  88 
Excitation,  2,  9,  21,  81 

crossing,  65,  78 

functions,  higher  order,  8 
Excitatory  connections,  105 
Excitatory  factor,  83 
Excitatory  neuron,  17,  19,  28 
Excitatory  neurons,  15,  57 
Excitatory  state,  2,  4 
Existential  net,  110 
Existential  sentence,  110 
Expression,  temporal  propositional, 
107 

realizable,  107 
Eye  movement,  71 

Facilitation,  110 
Factor,  excitatory,  83 


Factor  analysis,  75 
Ferry-Porter  law,  45 
Final  common  path,  80 
Final  pathways,  90 
Firing  of  afferent  neurons,  104 
Firing  of  neuron,  109 
Flicker-frequency,  critical,  45,  46 
Fluctuating  equilibrium,  29 
Fluctuations  in  response,  22 
Fluctuations  of  threshold,  53,  57 
Frequency,  critical,  44,  46 
Frequency  of  response,  49,  50 
Frontal  planes,  100 
Fusion,  99 

anomalus,  99 

of  images,  98 
Function,  temporal  propositional,  107 

General  neural  net,  30 

Gestalt-phenomena,  21 

Golgi  preparation,  30 

Graham,  C.  H.,  3 

Grant,  V.  W.,  97 

"Greater",  judgments  of,  61 

"Greater  than"  judgments,  59,  61 

Group  of  stimuli,  86 

Guilford,  J.  P.,  75 

Gulliksen,  H.,  84,  85 

Gustatory  stimuli,  40 

Gyemant,  R.,  53 

Hartline,  H.  K.,  3 
Hecht,  S.,  43 

Helmholtz,  H.  von,  90,  91,  99 
Higher  order  excitation-functions,  8 
Holway,  A.  H.,  73 
Horizontal  disparities,  100,  101 
Horopter,  99 
Horopter  rotation,  100 
Householder,  A.  S.,  8,  22,  26,  66,  70, 
75,  96,  97,  100,  101,  115 

Idealized  model,  vii 
Image 

cortical,  97 

retinal,  96,  97 
Images,  fusion  of,  98 
Impulse,  nervous,  vii,  2,  31 
Impulses,  13 

statistical  effects  of,  111 

statistically  independent,  112 
Inaccessibility  of  centers,  12 
Inaccessible,  10,  11,  31,  33 
Inadequate  stimulus,  81 
Incompatible  responses,  16,  74 
Inhibiting  effect,  5 
Inhibition,  4,  104 

complete,  104 

crossing,  65,  74,  76,  78 

partial,  104 

relative,  110 

total,  110 
Inhibitory  neurons,  10,  11,  15,  17,  19, 
28,  57,  67,  81,  87 


INDEX 


121 


Intensity,  92 

of  sound,  63 

of  stimuli,  61 

of  stimulus,  39,  40,  55,  84 
Intensity-discrimination,  69,  70 
Interaction  of  perceptions,  14 
Interaction  of  transmitted  impulse,  14 
Interconnected  chains,  56,  74 
Interconnecting  neurons,  105 

inhibitory,  57 
Interdependence  of  perceptions,  13 
Intermediate  neurons,  12 
Intermittent  stimulation,  45,  46 
Internuncial  neuron,  30 
Internuncial  neurons,  108 
Interocular  distance,  101 
Intra-neuronal  dynamics,  3 

Johnson,  V.,  90 

Judgment,  categorical,  58,  60 

Judgments 

aesthetic,  74 

"correct",  72 

"doubtful",  59,  60 

"greater  than",  59,  61 

"less  than",  59 

of  brightness,  60 

of  "equality",  59,  61 

of  "greater",  61 

of  "less",  61 

of  loudness,  63 

"wrong",  72 

Kellogg,  W.  N.,  60,  62,  63 
Kinaesthesis,  95 
Konig,  A.,  68 
Kubie,  L.  S.,  22 

Landahl,  H.  D.,  8,  16,  18,  22,  26,  52, 
54,  57,  61,  62,  64,  83,  85,  86, 
88,  89,  97,  103,  111,  115 

Lashley,  K.  S.,  84 

Latent  addition,  period  of,  111,  113 

Law  of  Plurality  of  Connections,  30 

Law  of  Reciprocity  of  Connections,  30 

Learning,  79,  85,  88 

Length,  discrimination  of,  70,  71 

"Less",  judgments  of,  61 

"Less  than"  judgments,  59 

Lettvin,  J.  Y.,  115 

Lifted  weights,  72 

discrimination  of,  73 

Light-dark  ratio,  45 

Lorente  de  No,  R.,  22,  30 

Loudness,  judgments  of,  63 

McCulloch,  W.  S.,  viii,  2,  103,  104,  107, 

108,  109,  111 
MacDonald,  P.  A.,  70 
Macroscopic  picture,  103 
Magnification-increment,  101 
Matthews,  B.  H.  C,  3,  49,  50 
Maximal  activity,  11 
Mean  value  of  threshold,  54 


Measure  of  region,  20 
Mechanism,  32,  37,  49,  56,  64,  72,  76, 
86 

discriminating,  91 

neural,  13,  90,  92 
Medial  plane,  100 
Memory,  79 

Microscopic  picture,  103 
Mixed  neuron,  26 
Mode  of  presentation,  62 
Model,  viii,  1 

idealized,  vii 
Modulating  effects,  22 
Monocular  vision,  94 
Movement,  eye,  71 
Movements,  spontaneous,  22 
Muller's  law  of  specificity,  90 
Muscles,  1 

Muscular  contraction,  4 
Muscular  stretch,  adaptation  to,  50 
Multidimensional  psychophysical 
analysis,  74,  76 

Near-threshold  stimulation,  65 
Negation,  105 
Nerve-impulse,  104 
Nervous  activity,  vii 
Nervous  discharge,  104 
Nervous  impulse,  vii,  2,  31 
Net,  30,  66,  83,  86,  105,  107 

cyclic,  109 

existential,  110 

non-cyclic,  107,  109 

order  of,  109 

steady-state  of,  36 

symmetrical,  57 
Nets 

neural,  95,  113 

boolean  algebra  of,  103 
Neural  chains,  13 
Neural  circuits,  79 
Neural  complex,  viii 
Neural  disjunction,  95 
Neural  mechanism,  13,  90,  92 
Neural  net,  general,  30 
Neural  nets,  95,  113 

boolean  algebra  of,  103 
Neural  pathway,  4 
Neural  pathways,  segregation  of,  90 
Neural  structures,  vii,  7,  37 
Neuron,  5,  22,  25,  33,  41,  49,  53,  54, 
61,  62,  107 

asymptotically  exciting,  5 

asymptotically  inhibiting,  5,  6,  9 

effects  of  ions  on  excitability  of, 
53 

excitatory,  17,  19,  28 

firing  of  afferent,  109 

inhibitory,  17,  19,  28 
single,  24 

internuncial,  30 

minimal  region  of,  53 

mixed,  26 

self-stimulating,  22,  26 


122    MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


single,  11,  31,  38 

transiently  exciting,  6 

transiently  inhibiting,  6 
Neuronal  dynamics,  viii 
Neurons,  vii,  1,  7,  13,  37,  62,  83,  103, 
112 

chain  of,  8,  9,  11,  12,  31 

chain  of  two,  38,  41,  46 

chains  of,  7,  22 

excitatory,  15,  57 

inhibitory,  10,  11,  15,  57,  67,  81, 
87 

interconnecting,  105 

intermediate,  12 

internuncial,  108 

parallel,  13 

interconnected,  21 

several,  49 
Nodal  points,  97,  101 
Node  of  Ranvier,  53 
Non-cyclic  net,  107,109 
Non-neural  events,  7 
Number  of  errors,  85,  88 
Number  of  possible  choices,  86 
Number  of  trials,  85,  88 

Ogle,  K.  N.,  99,  100,  102 

O'Leary,  J.  L.,  22 

One-dimensional  array  of  synapses,  90 

Order  of  net,  109 

Origin,  2,  3,  22,  33,  35 

Organism,  1,  13,  103 

response  of,  30 
Overlearning,  89 

Parallel  neurons,  13 

interconnected,  21 
Parameters,  17 

activity,  17,  19 
Partial  inhibition,  104 
Path,  final  common,  80 
Pathways 

final,  90 

segregation  of  neural,  90 
Pecher,  C,  53,  54 
Peddie,  W.,  90 

Perceived  size,  constancy  of,  97 
Perception 

color,  21 

of  space,  94,  95 
Perceptions,  13,  14 

interaction  of,  14 

interdependence  of,  13 
Perceptual  space,  95 
Period  of  latent  addition,  111,  113 
Permanent  activity,  22,  29 
Physiological  stimulation,  2 
Physiology,  vii 
Pieron,  H.,  40 
Pitts,  W.,  viii,  2,  9,  31,  33,  35,  36,  103, 

104,  107,  108,  109,  111,  115 
Pleasant  response,  83 
Plurality  of  Connections,  law  of,  30 
Possible  choices,  number  of,  86 


Presentation,  mode  of,  62 

Primary  receptor,  92 

Probability  of  correct  response,  58,  59, 
89 

Probability  of  wrong  response,  58,  84 

Probability-relation,  113 

Prompting,  88 

Proposition,  105 

Propositions,  universal  existential,  109 

Psychological  scale,  75,  76 

Psychophysical   analysis,   multidimen- 
sional, 74,  76 

Psychophysical  discrimination,  56 

Punishment,  84 
strength  of,  88 

Ranvier,  node  of,  53 

Rashevsky,  vii,  ix,  3,  4,  15,  16,  18,  19, 

21,  26,  79,  81,  111,  115 
Rate  of  decay,  49 
Reaction 

delay  of,  39,  40 
duration  of,  41 
Reaction-time,  38,  42,  43,  51,  52,  54 
Reaction-times,  variation  in,  55 
Recall-learning,  89 
Receptor,  7,  13,  78,  90 

primary,  92 
Receptors,  1 

cutaneous,  64 

retinal,  91 

stretch,  49 
Reciprocity  of  Connections,  law  of,  30 
Recognition-learning,  89 
Recovery  time,  43 
Reflex,  39 

scratch,  26 
Region,  measure  of,  20 
Reisz,  R.  R.,  69 
Relative  brightness,  48 
Relative  discrimination,  66,  72 
Relative  inhibition,  110 
"Relative  interval",  69 
Relaxation,  1 

Replacement,  rules  of,  113 
Response,  vii,  viii,  13,  16,  17,  37,  38, 
41,  43,  51,  56,  57,  61,  66,  78, 
80,  83 

correct  58,  61,  83,  87 

discriminal,  75 

distinct,  68 

duration  of,  41,  42 

fluctuations  in,  22 

frequency  of,  49,  50 

of  organism,  30 

pleasant,  83 

probability  of,  58,  59 

probability  of  correct,  58,  59,  89 

probability  of  wrong,  58 

steady,  44 

unpleasant,  84,  86 

wrong,  58,  84,  87 
Responses,  incompatible,  16,  74 
Response-time,  54,  55 


INDEX 


123 


Retinal  images,  96,  97 

Retinal  receptors,  91 

Retinas,  97,  98 

Reward,  strength  of,  84,  86,  88 

Robertson,  D.  M.,  70 

Rosette,  33,  36 

Rotation 

horopter,  100 

of  subjective  medial  plane,  102 

visual  field,  100 

Sacher,  G.,  26 

Saturation,  92 

Scale,  psychological,  75,  76 

Scratch  reflex,  26 

Segregation  of  neural  pathways,  90 

Self-exciting  circuits,  12 

Self-stimulating  neuron,  22,  26 

Sensations 

auditory,  68,  69,  70 

disparate,  64 

spontaneous,  22 

tactile,  68,  69,  70 

visual,  69,  70 
Sequence,  discriminal,  64,  74 
Sequences,  discriminal,  56 
Shape,  13 

visual  illusions  of,  13 
Shaxby,  J.  H.,  90 
Simple  circuits,  22 
Simultaneous  stimulation,  64 
Simultaneous  stimuli,  17 
Size 

apparent,  96 

constancy  of  perceived,  97 
Size-lens,  101,102 

cylindrical,  100,  102 

spherical,  101 
Smith,  J.  E.,  73 
Space 

perception  of,  94 

perceptual,  95 
Spontaneous  movements,  22 
Spontaneous  sensations,  22 
Stable  equilibrium,  23,  26,  35 
Stanton,  H.  E.,  97 
Statistical  effects  of  impulses.  111 
Statistically  independent  impulses,  112 
Steady  response,  44 
Steady-state,  11 

activity,  7,  67,  81 

of  net,  36 
Stereopsis,  94 

binocular,  97 
Stimulation 

applied,  11 

intermittent,  45,  46 

near-threshold,  65 

physiological,  2 

simultaneous,  64 

total,  11 
Stimuli,  13,  16,  78 

auditory,  40 

awareness  of,  16 


conditioned,  83 

group  of,  86 

gustatory,  40 

intensity  of,  61 

simultaneous,  17 

strong,  16 

two  competing  for  attention,  16 

unconditioned,  83 

visual,  39 
Stimulus,  viii,  5,  7,  10,  13,  22,  30,  31, 
37,  38,  49,  56,  57,  66,  78,  83, 
90 

applied,  9 

conditioned,  81 

constant,  42,  43,  49,  54 

discriminal  sequences  of — and  re- 
sponse, 21 

effective,  5 

inadequate,  81 

intensity  of,  39,  40,  55,  84 

subliminal  auditory,  65 

testing,  43 

total,  9 

unconditioned,  80,  81 

unpleasant,  83 

warning,  51,  52 
Stimulus-density,  18 
Stimulus-object,  75,  76,  77 

complex,  74 
Strength  of  punishment,  88 
Strength  of  reward,  84,  86,  88 
Stretch  receptors,  49 
Strong  stimuli,  16 
Structure,  30,  50,  79,  81,  83 

symmetric,  64 
Structures,  viii,  37 

neural,  vii,  7,  37 

of  subjective  space,  94 
Subjective  medial  plane,  rotation  of, 

102 
Sub-threshold,  14 
Summation,  104 
Symmetric  structure,  64 
Synapse,  vii,  6,  31,  33,  35 

common,  49 

single, 

common,  33 
dynamics  of,  37,  49 
Synapses,  91 

one-dimensional  array  of,  90 

three-dimensional  array  of,  91 
Synaptic  delay,  104 

Tactile  data,  70 
Tactile  sensations,  68,  69,  70 
Talbot  law,  48 

Temporal  propositional  expression, 
107 

realizable,  107 
Temporal  propositional  function,  107 
Terminus,  2,  3,  7,  18,  22,  33,  35,  67 

common,  13 
Testing  stimulus,  43 
Theory,  viii,  ix 


124   MATHEMATICAL  BIOPHYSICS  OF  THE  CENTRAL  NERVOUS  SYSTEM 


Three-dimensional  array  of  synapses, 

91 
Three-receptor  hypothesis,  91 
Threshold,  3,  9,  10,  24,  38,  41,  43,  53, 
61,  67,  104 

fluctuation  of,  57 

fluctuations  of,  53 

mean  value  of,  54 

modification  of,  82 

variations  in,  54,  55 
Thresholds,  13,  17,  19,  57 

absolute,  65 
Thurstone,  L.  L.,  75 
Time-unit,  31,  33 
Total  inhibition,  110 
Total  stimulation,  11 
Total  stimulus,  9 
Transiently  exciting  neuron,  6 
Transiently  inhibitory  neuron,  6 
Transmission,  14 
Transmitted   impulse,   interaction   of, 

14 
Trans-synaptic  dynamics,  1 
Trials,  number  of,  85,  88 
Two-neuron  chains,  79 
Two-neuron  circuit,  25 
Two  parallel  chains  with  crossing  in- 
hibition, 64 

Unconditioned  stimuli,  83 
Unconditioned  stimulus,  80,  81 
Universal  existential  propositions,  109 
Universal  sentence,  109 
Unpleasant  response,  84 
Unpleasant  stimulus,  83 
Unstable  equilibrium,  24,  35 


Urban,  F.  M.,  58,  59,  61 

Value,  asymptotic,  7,  25 
Variations  in  reaction-times,  55 
Variations  in  Threshold,  55 
Verhoeff,  F.  N.,  98 
Vertical  disparities,  100 
Vision 

binocular,  97 

monocular,  94 
Visual  axes,  97 
Visual  cues,  95 
Visual  data,  61,  68 
Visual  field,  binocular,  98 
Visual  field  rotation,  100 
Visual  illusions  of  shape,  13 
Visual  sensations,  68,  69,  70 
Visual  stimulus,  39 

Warning  stimulus,  51,  52 
Weber  ratio,  42,  48,  68,  72 
Weights 

discrimination  of,  71 

lifted,  72 

discrimination  of,  73 
Williams,  H.  B.,  114 
Woodrow,  H.,  51,  52 
"Wrong"  judgments,  72 
Wrong  responses,  58,  61,  84,  86,  87 

probability  of,  58,  84 

Young,  G.,  75 
Young-Helmholtz  theory,  90 

Zigler,  M.  J.,  73