Marine Biological Laboratory 7_, I 33 Pt A R Y AUG 1 4 1947 WOODS HOLE. MASS. Mathematical Biophysics Monograph Series, No. 1 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM By ALSTON S. HOUSEHOLDER AND HERBERT D. LANDAHL SttJV-ULTPS The Principia Press, Inc. bloomington, indiana COPYRIGHT 1945 BY THE PRINCIPIA PRESS COMPOSED AND PRINTED BY THE DENTAN PRINTING COMPANY COLORADO SPRINGS, COLORADO PREFACE Since the proposal of the two-factor dynamical model of neural activity by Professor Rashevsky in his book, Mathematical Biophysics, a great deal of work has been done in the formal development of the theory as well as in the applications to specific psychological problems. It is only natural that these developments by various authors over a period of time will be lacking somewhat in coherence and continuity. With this in view, it seems appropriate to pause at the present time to review the work which has been done so far to explain sys- tematically the techniques, to summarize and describe such structures as have been devised and the applications made, to suggest promising directions for future development, to indicate types of experimental data needed for adequate checks, and also to present new material not published elsewhere. It is the hope that this perspective of past achievements and, still more, of future prospects, will be of benefit to those theorists and experimenters alike who are interested in con- tributing to the understanding of some of the mechanisms which un- derlie psychological processes. While perhaps the names which most commonly run throughout the monograph are those of the authors themselves, this is largely only a reflection of their dominating interests in its preparation; and neither the monograph itself nor the papers of the authors and of many others herein referred to would ever have seen the light of day had not the general procedures and fundamental postulates been pre- viously developed by Professor Rashevsky. For this, and for other reasons too abundant to enumerate, the authors owe to him their foremost debt of gratitude. Their thanks are due also to Dr. Warren S. MeCulloch, Dr. Ger- hardt von Bonin, and Dr. Ralph E. Williamson for many helpful sug- gestions made during the preparation of the manuscript, and to Mr. Clarence Pontius for preparing all the original drawings ; to Mrs. Gor- don Ferguson and Miss Helen De Young for typing of the manuscript, to Miss Gloria Robinson for final preparation of the manuscript, proofreading, and preparation of the index. For help with the latter thanks are also due to Mr. Richard Runge. The authors also wish to thank the Editor of The Bulletin of Mathematical Biophysics for permission to reproduce Figures 1 and 2 of chapter vi ; Figure 2 of chapter vii ; Figures 3, 4, 9, and 10 of chapter ix ; Figures 2, 3, and 4 of chapter xi, Figure 1 of chapter xiii and Figure 1 of chapter xiv. To the Editors of Psychometrika, their thanks are due for permission to reproduce Figures 1, 6, 7, and 8 of chapter ix, and to The University of Chicago Press for permission to iii reproduce Figures 2 and 5 of chapter ix, from Rashevsky's Advances and Applications of Mathematical Biology, and Figure 1 of chapter xi from his Mathematical Biophysics. Finally, the authors wish to express their gratitude to the Prin- cipia Press and to the Dentan Printing Company for their unfailing efforts involved in publishing the book. Alston S. Householder Herbert D. Landahl Chicago, Illinois October, 19U IV TABLE OF CONTENTS PAGE Introduction vii PART ONE CHAPTER I Trans-synaptic Dynamics 1 II. Chains of Neurons in Steady-State Activity - - 7 III. Parallel, Interconnected Neurons 13 IV. The Dynamics of Simple Circuits 22 V. The General Neural Net 30 PART TWO VI. Single Synapse: Two Neurons 37 VII. Single Synapse : Several Neurons 49 VIII. Fluctuations of the Threshold 53 IX. Psychological Discrimination 56 X. Multidimensional Psychophysical Analysis - - 74 XI. Conditioning ------- 80 XII. A Theory of Color-Vision 90 XIII. Some Aspects of Stereopsis 94 PART THREE XIV. The Boolean Algebra of Neural Nets - - - - 103 XV. A Statistical Interpretation Ill Conclusion 114 Literature 116 Index 119 61131 INTRODUCTION This monograph is directed toward the explanation of behavior by- means of testable hypotheses concerning the neural structures which mediate this behavior. We use the word behavior, for lack of a better term, in a very broad sense to cover any form of response to the en- vironment, internal or external, whether it is acting or only perceiv- ing, and whether the response occurs immediately or after long delay, providing only the response is governed by nervous activity initiated by occurrences in the environment. We are seeking to develop a theory of the nervous system as the determiner of behavior. Data of anatomy and physiology are altogether inadequate — un- less in the case of a simple spinal reflex — for tracing in detail the progress of a nervous impulse from its inception at a receptor, through its ramified course in the nervous system, to its termination at the effector. We know pretty well where the fibers lead from the retina; we known even in some detail where the different retinal areas are mapped on the cortex ; we know a good deal about the inter- action of one region of the cortex upon another, and we know many details about the functioning of neural units. But how the neural units are combined in the visual area to enable the organism to locate an object seen and to act accordingly, how the nervous discharges from the two retinas are shunted this way and that to combine and emerge at the appropriate effectors, is not explained by existing ex- perimental and observational technique. For solving such problems it is necessary to create testable hypotheses, to be revised, replaced, or expanded according to the outcome of the tests. The task of developing this theory is three-fold. First, an ideal- ized model of the elementary units must be constructed in terms of postulates governing their individual behavior and their interactions. The model must be simple enough to permit conceptual manipula- tion. The units we have designated neurons. Our neurons are defined by the hypotheses we impose upon them, and it may turn out that not the single neuron of the physiologist and anatomist, but some re- curring complex of these is most properly to be regarded as its pro- totype. The junction of neurons we refer to as a synapse, and where the impulses from two or more neurons are able to summate in pro- ducing a response in one efferent neuron, or in each of several, we have also, briefly, referred to the set of these junctions as constitut- ing a single synapse. Such usage, in harmony with that of Rashevsky vii MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM (1938, 1942) , seems to simplify the terminology, since for us the junc- tion is primarily of dynamical, not anatomical, significance. We de- vote chapter i to the development of the postulates of our system, and to the elaboration of a few elementary consequences. The second stage in the development of the theory consists in the investigation of the properties of complexes of specified structure, these properties being deduced from the postulated properties of the units and from their interrelations in the structure. The bulk of Part 1 deals in a purely abstract manner with a number of different struc- tures which are in some sense typical of those required by the appli- cations, and considers the general problem of the reciprocal determi- nations of the structural form and the dynamics. The final stage is the comparison of prediction with fact. A neu- ral complex of particular structure is assumed to link the stimulus to the response in a given class of cases. From this assumption, a certain quantitative functional relation between stimulus and re- sponse can be deduced. To the extent to which experience verifies the prediction, we have confidence in our initial assumption and are jus- tified in extending the range of our predictions. In general the func- tional relations involve variables and parameters, each capable of assuming values over a certain range, so that any such relation yields predictions well beyond the actual range of verification. Whether or not verification occurs over some range, there must somewhere occur a failure. This is because both our units and our structures are of necessity over-simplified. But the failure is itself instructive, for the trend of the deviations can yield insight into the nature of the complications required for increasing the realism and extending the range of applicability of our model. This is the theme of Part II, in which deductions made on the basis of special struc- tures are compared with data, where data are available. Unfortunate- ly, even when data of a kind are available, these are not always well adapted to our special purpose. The test of a specific theory gen- erally requires the imposition of specific conditions upon the conduct of the experiment, and when the theory is not available to the experi- menter it is largely chance if these conditions are satisfied. Hence some comparisons can be made only in the light of special assump- tions, and too often no quantitative comparison at all is possible. It is our hope in publishing this monograph that more experiments will be planned to make these tests. In Part III we present the basis for an alternative development, as laid down quite recently by McCulloch and Pitts (1943). The neu- ronal dynamics as postulated by these authors is much more realistic, but the deductions from them of laws of learning and conditioning, viii INTRODUCTION of response-times, and of discrimination, remains largely a program for the future. Their laws are temporally microscopic, as opposed to those of Part I, which are temporally macroscopic. It is therefore to be hoped that the macroscopic laws can be deduced from the microscopic ones as approximations valid at least for certain com- monly occurring neural complexes, and some steps in this direction are outlined in the concluding chapter. In general we have sought, within the available space, to sum- marize and systematize the most important methods and results to date. We have passed lightly over most of the results already reported in Rashevsky's "Advances and Applications of Mathematical Biology," and we have omitted all reference to Rashevsky's aesthetic theory. On the other hand, many pages of Part I have been devoted to formal dis- cussions making no immediate contact with experience. While those whose interest lies only in the applications may wish to skip this ma- terial, the theoretically minded will recognize in these pages the groundwork for the further elaboration of what we hope will become a comprehensive and unified theory of the operation of the central nervous system. IX PART ONE TRANS-SYNAPTIC DYNAMICS The performance of any overt act, by any but the most primitive of organisms, is accomplished by the contraction and relaxation of groups of specialized cells called muscles. Normally these contractions and relaxations, by whatever mechanisms they may be effected, are at least initiated by prior events occurring at the junctions with these muscles of certain other specialized cells called neurons. Whatever the nature of these prior events, and by whatever mechanism they are effected, they are themselves initiated by a sequence of still prior events in the neurons themselves, and these in turn by yet earlier events at the junctions of other neurons with these. Thus regressing, step by step, we conclude that apart from possible cycles, pools of perpetual activity, the whole sequence was started by an initial set of events at the points of origin of an initial set of neurons. And finally this ultimate set of initial events — the set, or any member of the set, according to convenience, being called a stimulus — was brought about by or consisted in some physical or physiological occur- rence in the environment or within the organism. Doubtless there are often and perhaps always countless other accompanying events occurring within the organism and interacting to a greater or lesser degree with those events here mentioned, but no scientific theory can account for everything, and still less for everything all at once. We wish, therefore, to define our schematic reacting organism as one consisting solely of receptors (sense-or- gans) , effectors (muscles) , and a connecting set of neurons, the whole and the parts being affected by the physical or physiological environ- ment only insofar as this acts as a stimulus via the receptors. We wish to consider to what extent behavior can be accounted for in terms of such a model. In undertaking such an inquiry, we freely and expressly acknowledge that much is left out, and we emphatically refuse to make any claim in advance as to the range of the behavior that can be so accounted for. This is an empirical question to be ex- perimentally decided. But a hypothesis cannot even be refuted until it is clearly formulated. The structure of a neuron is fairly complicated and its behavior is hardly less so. Consequently, to make progress the neurons, too, must be schematized. Structurally there is a cell body and two or more 2 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM threadlike processes, but the terminations of these processes are of two sorts. A termination of the one sort we shall call an origin, one of the other sort, a terminus. When appropriate stimulation of suffici- ently high degree is applied at an origin, there is conducted along the neuron a "nervous impulse" all the way to the various termini. This nervous impulse, arriving at a terminus, may contribute to the stimu- lation of any neuron which has an origin at the same place. Doubtless in all strictness the impulse does not simply jump from neuron to neuron but passes by way of some intermediary process set up in the synapse which is the junction between the two neurons. From our point of view, it is largely a matter of convenience whether we postulate such an additional process or not. The nervous impulse manifests itself as a localized change in electric potential, its duration at any point is about half a millisecond, and it is transmitted at a rate that, though low in some neurons, in others may equal or exceed 10* cm sec-1. Moreover, in physiological stimulation, if the stimulation is maintained, the impulses are repeat- ed and may reach a frequency which is of the order of 102 sec-1. The more intense the stimulation, the more frequent the impulses, but there is an upper limit for any given neuron which varies somewhat from neuron to neuron. When we have occasion to take account of it, we shall suppose this upper limit to be a fixed characteristic of the neuron. McCulloch and Pitts (1943) have developed a theory of the "quantized" dynamics of the neuron which takes account of the in- dividual impulses and we shall return to this later at the end of this monograph. For the present, however, we shall schematize further by doing some statistical averaging and by fixing our attention upon the synapse rather than upon the neuron itself. We shall choose the alternative of supposing that the impulses of the afferent neurons are not the immediate stimuli for the efferent neuron, but that these im- pulses start or maintain at the synapse an intermediate process which is the immediate stimulus. To have a concrete picture, one may imag- ine that some chemical substance is released by the impulses and dis- sipated or destroyed as a monomolecular breakdown. However, it is by no means implied that this is the case, and furthermore, we shall not speak in such terms but shall speak merely of an "excitatory state," and denote the state or its intensity by s . More briefly we shall speak of the excitation e. The amount by which the impulses increase s in unit time is pre- sumably proportional to the frequency of these impulses, and the fac- tor of proportionality is taken to be a characteristic of the fiber. We make the simplest assumption as to the rate of dissipation of e and TRANS-SYNAPTIC DYNAMICS 3 assume it to be representable by the term ae . If we then take cuf> , proportional to the frequency, to represent this rate of increase of e by the impulses, we obtain the equation (Rashevsky, 1938) de/dt = a( — e) (1) which we assume to describe the development of e . We take e to be a measure of the stimulus acting upon any neuron which originates at the synapse in question. Note that by equation (1) we pass, in a sense, directly from origin to terminus of the neuron, compressing into the function 6 our only reference to the intra- neuronal dynamics. When = 0 the impulses have zero frequency, i.e. do not occur, and we shall say the neuron is at rest. Nevertheless, £ is not necessarily zero, and in fact, after the neuron has been active, e vanishes asymp- totically only according to equation (1) in which = 0 . Now $ is proportional to the frequency of the generating im- pulses, and this is, as implied, an increasing function of the applied stimulus with, however, a finite asymptotic value. Hence we may write $=(S), (2) where S is a measure of the applied stimulus. However, in order for the impulses to occur, S must exceed a certain minimal value, called the threshold, which is characteristic of the neuron in question and which we shall denote by h . Hence <£ (S) is zero for S = h , and for S > h , ( h are the following (Rashevsky, 1938; in this connection cf. Hartline and Graham, 1932, and Matthews, 1933) : <£ = 4>o[l-e-Q], (3) «/>0 (S-h)d + h = log , (4) log<5 S where 6 is small and 0 is the asymptotic value of . For not too large values of S either function may be approximated by an expression of the form = a(S-h), (5) and the second by 4> = fi\og(S/h). (6) In any case, for S ^ h , = 0 . 4 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM Now ^ is a function of S , and S may well be a function of t , in which case $ is a function of t . The complete solution of equation (1) is then given by e = e-at[e0 + a(teaTct>(T)dT], (7) J 0 where £0 is the initial value of e . When, in particular, S , and there- fore also , is constant with respect to time, this becomes e = e-** e0 + <£(1 - e-at). (8) Thus £ approaches the value asymptotically, the approach being in all cases monotonic, and either increasing or decreasing according to whether exceeds or is exceeded by e.0 . If we were now to introduce an assumption to relate the muscu- lar contraction with the applied e , we should have a system of for- mulae to be evaluated sequentially along any neural pathway from receptor to effector, for relating the time and the intensity of the re- sponse to the temporal form of the stimulus. But this would obvious- ly provide only a very incomplete picture. A given stimulus not only leads to the contraction of one set of muscles ; it leads also to the re- laxation of the antagonistic muscles. Any effective movement involves both components, of contraction and the inhibition of contraction. Thus we are inevitably led to extend our picture to include the phe- nomenon of inhibition. There are many ways in which such a phenomenon could be in- troduced into our schematic picture, but the simplest way seems to be to suppose that at least some neurons have the property of creating, as the result of their activity, an inhibitory state of intensity j , (briefly, an inhibition j) , antagonistic to the excitatory state e , and to suppose that the production of j follows the same formal law as that for e: dj/dt = b{y>-j). (9) The function y> is of the same type as <£ and it is only as a matter of convenience that we introduce a separate symbol. Rashevsky (1938, 1940) commonly assumes that in general the activity of any neuron leads to the production of both s and j , al- though for particular neurons, the one or the other may be negligible in amount. Evidently we may always replace a single neuron devel- oping both e and j by a pair, one developing £ alone and one j alone. It is useful, however, to consider some of the characteristics of a "mixed" neuron of the Rashevsky type. Since £ and j are antagonistic, we are now supposing that and y are themselves constant. Asymptotically £ and j approach and \p , respectively, so that the neuron is asymp- totically exciting or asymptotically inhibiting according to the rela- tive magnitudes of

. However, the initial rates of increase of £ and j are equal to a<}> and to by , respectively, so that an asymptoti- cally exciting neuron — for which cf> > ip — would be momentarily in- hibiting in case by> > a , and vice versa. Thus the transient and the asymptotic effects of such a neuron would be quite opposite. Furthermore, suppose, for definiteness, that the neuron is asymp- totically inhibiting, \p > , and consider the effect following the cessa- tion of its own stimulus, when the neuron, as a result, comes to rest. We suppose for simplicity that the constant stimulus is maintained long enough for the asymptotic state to be reached. Then, on re- moval of the stimulus, <£ and y.> both drop to zero so that £ and j de- cline exponentially to zero. If b > a , the decline of j is more rapid than that of £ and a transient exciting effect may, and in fact always does, ensue while the neuron is thus at rest. To summarize all possible cases of this sort: A neuron is a) Asymptotically exciting whenever 4> > ip . /* K f^ . — ~"": \ \ // ' J \ \ / / \ \ / / \ \ / / 1 / a^ 1/ if \ \^^ 1/ •v — ^_ Figure 1 6 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM Furthermore, when i) a < b , cuj> > by it is always exciting in activity and transiently exciting at rest; ii) a < b , 04 < by it is transiently inhibiting in activity, tran- siently exciting at rest (Figure 1) ; Figure 2 iii) a > b , a > by it is always exciting in activity, transiently inhibiting at rest (Figure 2). The case a > b , cuj> < by is inconsistent with <£ > y> . b) Asymptotically inhibiting whenever Furthermore, when b , cuj) < by it is always inhibiting in activity and tran- siently inhibiting at rest; ii) a > b , cuf> > by it is transiently exciting in activity, transient- ly inhibiting at rest; iii) a < b , a$ < by it is always inhibiting in activity, transiently exciting at rest. The case a < b , a > by is inconsistent with <£ < \p . We have tacitly assumed that the be its terminus, and so sequentially, the terminus of the last neuron of a chain of n neurons being sn . If a stimulus S is applied to N0 at s0 , it may come from a receptor or from a neuron or neurons not in the chain. If Nn-X develops o- at sn , this may act upon an effector or upon a neuron not in the chain. That is immaterial. Suppose that the neurons of our chain are all of the simple excitatory type. Suppose, further, that only a negligible time is required for the o-(= e) produced by any neuron to reach its asymptotic value (/> when a constant stimulus S is applied. In other words, we are now considering the chain only in its asymptotic state after stimulation by a constant stimulus. Then if u0 = S is the total stimulus acting at s0 upon N0 , it follows that o"i — o (o"o) is the a produced by N.0 at sx , where fo is the ^-function of N0 . If no receptor or neuron outside the chain introduces any S or a at s1 , then (T2 = <£i((Ti) = <£l[<£o(0"o)] is the o- produced by Nx at s2 . Thus we can calculate sequentially all the o-i . 8 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM Define the functions i,v(o) by recursion formulas ^>i,v+i(c7) = 0i+v+i [i,v(j,v \Oi) , (2) so that each function i>v gives the a produced at si+v+] in terms of that present at Si . If the derivatives exist, then 0'i,v+i('i+v<-i[<£i,v('i+v[^);,v-i (cr)] • • • ' i (i(o) is monotonic, so is each 4>'i,v('i,v(a). Further, if each ■/>'< vanishes asymptoti- cally, so does each 'itV . Hence these "higher order" excitation-func- tions are functions of the same type as the ordinary ones, and we may always replace any such chain of neurons by a single one, at least if we are interested in the asymptotic behavior alone. If every '» (a) ^1 for all a , then the sequence Co , ffi , (J? , • • • , trn is decreasing. In fact, if hi is the threshold of Nt , then <7i+l < CTj ~~ Aij , and if h is the lowest threshold in the chain, 'i(a) > 1 for small values of a , this is not necessarily the case. It has been shown (Householder, 1938a) that when all the neu- rons are identical, and the chain is long, the a, will then either di- minish to zero, or approach a certain positive limit characteristic of the chain, according to whether 1 , ah/ (a — 1). When the second inequality is reversed but the first holds, the pro- gression consists of negative terms which increase numerically until some as linear until the stimulus reaches a certain maximal value, and constant at the upper limit thereafter. This representation, though no doubt less accurate than the functions (3) and (4) of chapter i, is at any rate a fair first approximation, and is much more easily handled. Let St be the applied stimulus at Si , and define the quantities £i = Si — hi , rji = £i + i(^i) (7) where 4>i (v> ) — 0 when rji ^ 0 , i (rji) = a» rji when 0 < rji < H% , (8) « (rji) = an Hi when rji ^ #; . The coefficient a* we shall call the activity-parameter of 2V< , and it may be positive, for an excitatory, or negative, for an inhibitory neu- ron, but is not zero. Our problem is the following: supposing Si , 52 , • • • , Sn fixed, to express rjn as a function of rj0 — S0 — K , and, more generally, to express rji+v as a function of rji when Si+1 , ••• , S^v remain fixed. Since each rji+1 varies linearly with rjx when the latter occupies a certain restricted range, and rjin is otherwise constant, it is at once apparent that the same is true of the variation of any iji+v with rji . Again, since any & may be so large that rji always exceeds Hi , or so small that rji is always negative, it is evident that rjuv may remain constant for all values of rji . In such a case si+v is said to be inac- cessible to Si (Pitts, 1942a) ; otherwise it is accessible. More explic- itly, if, as rji varies over all values from — oo to + oo while Si+1 , ••• , Si+v remain fixed, the value of rji+v remains constant, then si+v is inac- cessible to Si . Clearly si+1 is always accessible to Si . If si+v is inac- cessible to Si , then si+v is also inaccessible to any s , and further , i—v' any s is inaccessible to Si . Finally , it is clear that if Si+V is acces- i+v+v' sible to Si , then where rj^v varies with rji , it decreases if there is an odd number, increases if an even number of inhibitory neurons (with negative a's) between Si and Si+V . The above conditions for inaccessibility may be phrased thus: // any of the four following conditions holds: a* > 0 , S^ + aiHi^O, ai >0, £i+l = "t+1 > (S), very similar results must hold in general. Ill PARALLEL, INTERCONNECTED NEURONS Color-contrast and visual illusions of shape provide well-known examples of the almost universal interdependence of perceptions. Phy- siologically no stimulus occurs in the absence of all others, and the response to any stimulus depends in part upon the nature of the back- ground against which it is presented. One may wish to say, indeed, that the true stimulus to the organism is the whole situation, but since we cannot discuss any whole situation, and since the whole situ- ation is never duplicated, such terminology does not seem to serve any useful scientific purpose. If two stimuli differ only in degree, it may be true that the re- sponses which they evoke differ only in degree, the stronger stimulus evoking the stronger response. But in many instances there is a com- plete change in the form of the response, and in others it is the weaker stimulus, and not the stronger, which brings forth the strong- er response. In our schematic reacting organism, such phenomena are easily understood in terms of suitable interconnections between parallel neu- rons. We are reserving Part II for the precise formulations neces- sary to make quantitative predictions, so that we content ourselves here with a few qualitative results to indicate in general how this comes about. In the barest terms, if two stimuli which differ only in degree lead to responses which differ in form, then there are pathways — neural chains from receptor to effector — which can be traversed when the impulses are initiated by a stimulus within a given range of in- tensities but not when these are initiated by a stimulus lying outside this range on the scale of intensities. The simplest neural mechanism having this property consists merely of two neurons, Ne excitatory and Ni inhibitory, having a common origin and a common terminus (Landahl, 1939a). Let he and hi be the thresholds of Ne and Ni , re- spectively, and let h be the threshold of some neuron N originating at the common terminus of the two neurons. Suppose hi> he ,y>(oo) > (hi) >h. Then if S is sufficiently near to hi in value (Figure 1), asymptotically (£) ~y(S) >h and N will become excited, whereas a somewhat larger S will result 13 14 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM Figure 1 in a negative a at the origin of N , and a somewhat smaller one will yield a sub-threshold, though positive, a . There will be some range h', h" , therefore, which contains h, , and within which S must lie if transmission is to occur. These are not the only possible relations that will limit the range over which transmission may occur. We could have, for example (Figure 2), - v hi= he Figure 2 hi = he, ' (hi) > y>'(hj), y(oo) > (oo). Then for an S within a limited range, exceeds y> , which is all that is required except that h must be sufficiently small. Suppose, then, one has a number of such sets, Ni , Ne and N , all with a common origin, and each possessing a characteristic range. If these ranges do not overlap, and if each set is connected through a chain to a par- ticular effector, then any S will excite only the effector corresponding to the particular range on which S lies. As a first step in discussing the interaction of perceptions (strict- ly the interaction of the transmitted impulse), and as a kind of gen- PARALLEL, INTERCONNECTED NEURONS 15 eralization of the mechanism just discussed, consider the following (cf. Rashevsky, 1938, chap. xxii). The neurons N„ and N22 are ex- citatory, originating, respectively, at s1 and s2, terminating, respec- tively, at s\ and s'2 . The neurons N12 and AT.21 are inhibitory, originat- ing, respectively, at sx and s2 , terminating, respectively, at s'2 and s\ . Let us restrict ourselves here to a range of intensities over which the linear approximations to the functions and y> are adequate. Then, stimuli Si and S2 being applied at s3 and s2 , we have <*i — «n (Si — /in) + a21 (S2 — h21), a2 = ai2 (Si — /i-i2) + a22 (S2 — /l22) , (1) when the quantities within the parentheses are all positive. When any of these quantities within parentheses is negative, however, the term is deleted. The conditions for the excitation of 2v\ and N2 are, respectively, a1 > hx and h? . A geometric representation of these conditions is easily obtained on the (Si , S2) -plane. The graph of the relation hi is that to the right of and below the broken line. Like- wise the region defined by or2 > Ju consists of those points above and to the left of a certain broken line which consists of a horizontal ray extending to the left, and a ray of positive slope. If these regions overlap (Figure 3), it is possible to have both Nx and N2 acting simul- taneously. Otherwise it is not possible. Figure 3 These regions necessarily overlap if the determinant of the co- efficients in equation (1) is positive: 16 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM an a21 O.X2. tX22 >0, (2) for then the line o-i = hx is steeper than a2 = h2 , so that for large Sx and S2 the magnitudes can be so related that both the inequalities, ax > hx and «r2 > hn , are satisfied. The case when condition (2) fails is of some interest (Rashev- sky, 1938, chap. xxii). Now, if neither corner lies in the other re- gion, the rays do not intersect, and it is never possible, with any Si and S2 , for both Nx and N2 to be excited at the same time. In fact, even with strong stimuli, the point (Sx , S2) may be outside both re- gions and neither Nx nor 2V2 is excited. On the other hand, if either corner does lie in the other region, the rays do intersect, and there is a finite region of overlap as illustrated in Figure 3. Points (Si , S2) beyond the intersection and between the two rays represent pairs of stimuli which, though strong, fail to excite either Nt or N2 . The analytic condition for this is the simultaneous fulfilment of the two inequalities an (h12 — &n — fei/flu) — a21(/^i — h22 — fh/a22) > 0 , (3) — /a22) > 0 , together with the failure of equation (2). The situation here described may be thought of as that of two stimuli competing for attention. When conditions (3) hold and (2) fails, there are moderate stimuli which lead to excitation of both 2V"i and 2V2 (awareness of both stimuli) , while with more intense stim- uli, unless one is sufficiently great as compared with the other, each stimulus prevents the response appropriate to the other and no re- sponse occurs. For the special case in which the mechanism is altogether sym- metric (Landahl, 1938a; cf. also chap, ix) we may set a11 = a22 = a, — a12 = — a21 = P, (4) If the two responses are incompatible in nature the parameters might be so related that the two conditions (3) cannot be satisfied. The failure of these reduces to the single inequality h"^a(h'-h) (5) which holds necessarily in case we have h' ^ h . If the relation (5) is replaced by an equation, we have a kind of discriminating mechan- PARALLEL, INTERCONNECTED NEURONS 17 ism by which the stronger of two simultaneous stimuli elicits its ap- propriate response and prevents the other response. If, further, a1 = S1- S2+ (h' -h), and if, finally, h' and h are very nearly equal, the transmitted stimu- lus approximates the absolute value of the difference between the two stimuli. More generally, let the excitatory neuron NH connect s, with s'i (i== 1 , ••• , n) (Figure 4), let the inhibitory neuron Na(i ¥= j) con- N11 N, S^^—^ , ^^.jt> So> ' 2 — ><^^—""« *-""Va . ->3^^_?j3. Figure 4 nect St with s'j and let a , — ft , h , li and ft" be the activity parameters and the thresholds of the various neurons. If all neurons of the first level are active, Vi = a(Si-h) -/J2 (Sj-hf) , (6) and the conditions for excitation of N\ (originating at s\) is oi > h". If h < h', then for values of the Sj between h and h' the excitatory but not the inhibitory neurons are excited. If, further, a(h' -h) >h" , (7) then for values of the Sj near h' the neurons Ni are all excited. But with n > 1 + a/0 , (8) when all the Si are equal, the m -> " = Om+1* (10) a + p In particular if Oi == O2 — — • • • — — o«i — — o , &in+l Orft+2 * * * on O , then these relations are equivalent to [a-/?(ra-l)] S' - P(n-m)S" > ah- (n-l)ph' + h" (11) ^ [a - P(n-m-l)] S" - fimS' . In either case the m stimuli Si produce their response and prevent the occurrence of the response to the other n—m stimuli (cf. Rashev- sky, 1938, chap, xxii ; Landahl, 1938a) . Receptors in the skin and the retina are far too numerous to be treated by the simple algebraic methods so far employed. Here we must think in terms of statistical distributions. The receptors, or at least the origins of the neurons to be discussed, may have a one-, a two-, or a three-dimensional distribution. According to the case, let the letter x stand for the coordinate, the coordinate-pair, or the co- ordinate-triple of the origin of any neuron. Let x' represent the co- ordinate, the coordinate-pair, or the coordinate-triple of any terminus of one of these neurons. Running from the region x , dx (consisting of points whose coordinates fall between the limits x and x + dx) to the region x', dx' may be excitatory or inhibitory neurons, or both. If we consider only the linear representation of the functions and tp , each neuron is characterized by the two parameters a and h . To consider a somewhat more general type of structure than the one just discussed for the discrete case, let N (x , x' , a , h) dx dx' do. dh represent the number of neurons originating within the region x , dx , terminating in x' , dx' , and characterized by parameters on the ranges a, da and h , dh . Then, S(x) being the stimulus-density at x , the a-density which results from these neurons alone at x' is N(x , x' , a , h) [S(x) — h] dx da dh PARALLEL, INTERCONNECTED NEURONS 19 provided h < S (x) . Hence the total a-density, obtained by summing over the entire region (x), over all values of a (positive and nega- tive), and over all values of h < S(x) is a(x') =J(xJ^Jso^aN(x,x',a,h)[S(x) -K] dhdadx. (12) Corresponding expressions can be derived, of course, on the suppo- sition that and xp are non-linear of any prescribed form (Rashev- sky, 1938, chap. xxii). Instead of writing the special form of expression (12) for the strict analogue of the discrete case considered above, let us suppose next that the inhibitory neuron Na , rather than passing from Si to s'j , passes from s\ to s'j (Rashevsky, 1938, chap. xxii). The net a at any s'j is then equal to the a produced by excitatory neurons terminating here diminished by the amount of inhibition produced by the inhibi- tory neurons originating at the other s'(- , whereas it is this net a at the s'i which acts as the stimulus for these inhibitory neurons. We have therefore to solve an integral equation in order to determine the net a. In the continuous case, let a(x) be the gross a-density produced by the excitatory neurons terminating in the region x , dx . For the inhibitory neurons, let — p = a represent the activity parameter. Let N(x' y x , /} ,'h) dx' dx dp dh represent the number of inhibitory neu- rons which run from the region x' , dx' to x , dx , and have activity parameters and thresholds limited by the ranges p , dp and h , dh . Then a(x) =o(x) — (13) f x> lo IoiX,) PN(X' 'X'P'W&ix') ~ &] dhdfidx'. This is an integral equation, but one in which the unknown func- tion enters as one of the limits of integration. If we interchange orders of integration, we have as a form equivalent to equation (13) a(x) =7(x) — (14) J"/00/ 0N(x' ,x,p,h)[h Then if, as a particular case, all inhibitory neurons have the same ft and the same h , this becomes a(x) =a(x) -ft J N(X' ,x)[ h , we could substitute expression (16) for a into the integral in expression (17), integrate, solve for / , and finally place this value in relation (16) to obtain a . Not knowing these limits, we proceed as follows. Since o- and a differ only by a constant, the limits of the region can be defined by the equation «{x)=[i, (18) for a suitable ii . Leaving /u for the moment undetermined, we carry through the steps as outlined except that the range of the integration is to be defined by o- > n . We first obtain fa(x')dx' - hM(fx) / ii . But since a > ii and h define the same region, it follows from (16) that it-XI(ii)=h. (20) Hence if we solve this equation for /a , then we find, by equations (16) and (20), that (x) =7(x) +h- p. (21) si It is evident that the procedure here outlined is applicable, with obvi- ous modifications, in case AT is a function of x' but is independent of x . From the fact that a and cTdiff er only by a constant, certain prop- erties of the solution are at once apparent. If o- anywhere exceeds h , then o- must somewhere exceed h . For if a nowhere exceeded h , then I = 0 , a = a , and we have a contradiction in the fact that o- itself somewhere exceeds h . Further, o- is a decreasing function of X . For if an increase in X led to an increase in a, then by relation (17) / would increase and by relation (16) a would decrease, which is a con- tradiction. To suppose that a decreases as h increases leads likewise to a contradiction, so that (£ + e) - e]. (2) 22 THE DYNAMICS OF SIMPLE CIRCUITS 23 If the outside stimulus is constant, the only case we shall consider, this differential equation can be solved by a quadrature, t/ £1 de (£ + £> -e to obtain t as a function of e : t = T(e) = at , (3) (4) We must then solve this equation for £ as a function of t . However, certain properties of this solution are obtainable directly from a con- sideration of the form of equation (2). We recall that for £ + £ positive, and its first derivative are positive, with the derivative decreasing monotonically to zero. For £ + £ negative $ is identically zero. Suppose first that <£'(0) = 1 . Then, since £ is always non-negative, the equation £ = <£(! + £) (5) has always a single root s0 which may be zero (Figure 1). For £ > £0 the right member of equation (2) is negative and £ is decreasing; for £ < £0 the right member of equation (2) is positive and £ is increas- ing. Hence £ = £0 represents a stable equilibrium. Whenever £ ^ 0 , £0 == 0 . Whenever I > 0 , then £0 > 0 , and enhancement of a in the amount £„ results, but after withdrawal of the stimulus, when £ = — h , the neuron comes to rest. If (j>' (0) > 1 , equation (5) has a single root £ == e0 > 0 when I > 0 , the single root s = 0 when I < 0 and numerically large, two positive roots besides the root e = 0 when £ < 0 and numerically small (Fig- ure 2), and one positive root besides the root £ = 0 when £ = 0 . If 24 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM £<0 4>'io) >1 Figure 2 e is eliminated from equation (5) and *'U + e)=l , and the resulting equation solved for £ = £<, , then £0 < 0 , and for £o < I < 0 we have the case for which equation (5) has two distinct positive roots. Let £0 represent the greatest of the roots (possibly zero). Then Co represents a stable equilibrium of equation (2). Since we can have £0 > 0 even for £ < 0 (if also I > £0) it is possible for the activity to persist even after the withdrawal of the stimulus when £ = — h , provided —h > £0 , and provided the initial value £ = £i at the time of withdrawal of the stimulus exceeds the smaller, unstable, positive equilibrium obtained from equation (5) when £ — — h . But whatever the value of '(0), if bx exceeds the threshold h at the time the stimulus is withdrawn, some activity will continue for a time, if not permanently. In order to account for learning in terms of activated circuits, the continuation must be permanent or nearly so (cf. chap. xi). Very likely a number of circuits would be involved in any act of learning, in which case forgetting could be accounted for as a result of the gradual damping out of one after another because of extraneous inhibition. In order to determine the period of the continuation where it is not permanent, it is necessary to know something about how the applied stimulus S disappears. If S suddenly drops to zero, then the time required for the activity to die out is given by equation (3) with £ = — h and the upper limit £ of the integration equal to +h . But if S is itself an £ from another neuron, a new set of equations must be written down and solved. If a circuit is formed by a single inhibitory neuron, the behavior is described by the equation THE DYNAMICS OF SIMPLE CIRCUITS 25 dj/dt = b [y(£- j) - ?] . (6) Then y> > 0 only if | > 0 , but in this case there is always a single, stable, equilibrium. The result is that the applied 5 is decreased at equilibrium by a certain amount j.0 . Also j(> increases as S increases, although if \p has a finite asymptotic value, j0 cannot exceed this, whatever the value of & . We may note, however, that the presence of additional circuits of this kind, with higher thresholds, which add their effects with increasing S , would provide an effective damping mechanism over an arbitrarily large range. Consider next a two-neuron circuit, with one neuron passing from s2 to s2 , and the other from s2 to Sj . Suppose, first, that these are both excitatory, and, for simplicity, that they are identical in character. Let |t be the excess of Si over the threshold of the neuron originating at sx , let £i represent the excitation produced here by the other neuron, and let |2 and e2 represent the corresponding quantities at s2 . Then, still neglecting the conduction time, we have del/dt = a [» ei=l£i + £1). (8) 26 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM For any point to the right of the first curve sx is decreasing, and for any point above the second curve e2 is decreasing. In case <£'(0) ^ 1 , the two curves have always one and only one intersection ; at this both e's are positive only if at least one £ > 0 and the other is not too small, and they define a stable equilibrium. In case ^'(0) > 1 there will be one or three intersections. If there are three, one of these is always the origin, and this is always a stable equilibrium; if there is only one, this equilibrium is always stable and it may be the origin. In the former case, the intersection farthest from the origin is also stable. In particular, continuous activity following the withdrawal of the outside stimuli can occur only in circuits for which ^'(0) > 1 , and then only in case at least one of the initial e's has become suffici- ently large. More detailed discussions of this type of circuit in which an exponential form is assumed for the functions , but these are not assumed to be identical, have been given by Rashevsky (1938), and Householder (1938b). Rashevsky (1938) has introduced these circuits in his theory of conditioning [cf. chap. xi]. Circuits containing both excitatory and inhibitory neurons are somewhat more interesting, because of the possibility of periodical phenomena (Landahl and Householder, 1939; Sacher, 1942). The scratch-reflex is one of numerous examples of a repetitive or fluctuat- ing response to a stimulus. Consider the case of a single inhibitory and a single excitatory neuron, both of which originate and termi- nate at the same place, s . A self-stimulating neuron of mixed type may be regarded formally as a special case. For a mixed neuron, it is common to assume (Rashevsky, 1938) that the functions of and \p have a constant ratio for all values of their common argument, and we shall make this assumption for simplicity. Then the equations may be written de/dt = AE(£ + e - j) -as, dj/dt = BE(§ + s-j)-bj, (9) E — 04/ A = by/B . By dividing out a suitable factor from E and incorporating it into A and B , we may suppose without making any restrictions that £7(0) =1. (10) Now if it should happen that a = b , we could subtract the sec- ond of these equations (9) from the first, replace e — j everywhere by +&- (A - B)e1] x (17) + lab - (bA - aB)el] x + ~- = 0 with terms of second and higher degree in x omitted. Now at the value o-0 considered, the slope of the left member of equation (14) must be less than one, and this, in view of the expan- sion (15), means that the coefficient of x in equation (17) is positive. Hence the characteristic roots of the linearized equation (17) are either complex or else real and of the same sign ; if, further, a + b> (A -B)elf (18) the real parts are both negative ; and if, finally, (\M - \rB)2e1 (|2 - j) - e] , dj/dt=bbp(i-i + e) -n. (20) There is always a single equilibrium obtained by equating to zero the right members of these equations (Figure 4) . Let e,0 , j0 represent Figure 4 the values at equilibrium, and let neither of them vanish. Then if we set X = £ - £0 , y = j - jo and expand, equations (20) have the form x' = — a(x + ay) +•••, y' = b(fix-y) +..., (21) where —a and /? are the derivatives of 4> and of \p at /<> and at e0 , re- spectively. The characteristic equation is A2 + (a + b)X + ab (1 + a 0) — 0 . (22) Since all parameters are positive, the real parts of the characteristic roots are always negative and the equilibrium is stable. If, further THE DYNAMICS OF SIMPLE CIRCUITS 29 (a- b)2 < 4 abaft , the roots are complex and the approach to equilibrium is fluctuating with a frequency v satisfying -16ji2r2= (a- b)2 - 4 abaft . In this circuit it is plain that permanent activity is only possible when |2 > 0 . Thus the simplest circuits which exhibit fluctuation are those consisting of one excitatory and one inhibitory neuron, and a circuit so constituted can maintain permanent activity in the ab- sence of external stimulation only if both neurons originate and termi- nate at the same synapse and A/a > B/b . This is, of course, quite evident intuitively. THE GENERAL NEURAL NET If the response of the organism can be expressed as some func- tion of the stimulus, this function must depend upon whatever para- meters are required for describing the structure of the nervous sys- tem. The psychologists can tell us much about the empirical charac- ter of this function but nothing about the parameters. The anat- omists and physiologists can tell us many things about the para- meters. Our hope is for a synthesis of the results of both lines of endeavor. If we knew all about the structure, we might hope to devise meth- ods for deducing the function. Actually, with complex structures, this becomes exceedingly difficult, though we have done this for struc- tures of some very simple types. If we knew all about the function, empirically, we might hope to deduce some of the characteristics of the structure. However, there is never a single, unique structure, but many possible ones, all leading to a function of the same empirical characteristics. And, of course, we do not know all about either the structure or the function, but only some things about each. Certainly the structure of the complete nervous system can be no less complex than the behavior which is an expression of it, and any Golgi preparation of a section from the retina or the cortex abun- dantly exhibits such complexity. We have already mentioned one of two general principles concerning this structure first stated by Lo- rente de No in a paper which appeared in the Archives of Neurology and Psychiatry, Vol. 30 (1933), pp. 245 ff. His statement of these is as follows: "Law of Plurality of Connections. — If the cells in the spinal or cranial ganglia are called cells of the 'first order' and the following ones in the transmission system cells of the second, third to ••• nth. order, it can be said that each nucleus in the nervous system always receives fibers of at least n and n + 1 order, and often of n , n + 1 and n + 2 order." "Law of Reciprocity of Connections.— If cell complex A sends fibers to cell or cell complex B , B also sends fibers to A , either direct or by means of one internuncial neuron." In chapter ii we assumed the function or y> for each neuron to be linear with S between the threshold and a certain maximal value characteristic of the neuron, and elsewhere constant. The coefficient 30 THE GENERAL NEURAL NET 31 of the linear variation we called the activity-parameter. We found that if a chain of n neurons leads from a synapse s,0 to a synapse sn , and if fixed stimuli &,•••,£» are applied at each synapse sx , ••• , sn , then the total excitation yn + y" , obtained in chapter ii for a chain, differs formally from that for a single neuron only by the presence of the term 3 . But if we set Z = Sn + 3, (2) we have more simply yn = Z + Ay0 (3) when y0 lies between the stated limits, and when it does not the near- est limit appears in this equation in place of y,0 • The similarity to the behavior of a single neuron is now complete, the term Z cor- responding to the stimulus applied at the terminus. However, this term, as well as the limits y and y" , depend here upon the particular stimuli applied at the various synapses of the chain. In the discussion of more general types of net, only the occur- rence of circuits can present essential complications and hence we limit ourselves to these, considering first the case of a simple circuit of n neurons. Such a circuit is obtained by closing a chain of n neu- rons, bringing s„ and s0 into coincidence. But if sn is inaccessible to s0 before the closure then the closure makes no change in the value of yn . Hence we suppose sn accessible to sc . Following Pitts (1942a) in substance, we find it convenient to employ semi-dynamical considerations, taking into account the con- duction-time. Let us introduce as the time-unit the time required for a nervous impulse to traverse the circuit completely. Having defined in chapter ii the 5 and 4 employed in equation (1), we shall have no further occasion to refer to the parameters of the individual fibers, or to the y at any point except s0 = sn , wherefore it is legitimate to drop all subscripts as designations of neurons and synapses. Fur- ther, it increases somewhat the generality without adding essential complications to allow the stimulus at this point during the interval 0 = t < 1 to be different from the constant value to be assumed there- 32 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM after. Then, if y0 is now taken to represent the value of y during this initial interval, we have y(t)=y0 (0^t 1 , the interval between y(t) and y:X , wherever the latter may be, continues to increase while y(t) lies on the interval, and after having passed either limit — which will occur in a finite time — it re- mains constant. If A < — 1 , the interval y — y^ increases numeri- cally but with alternating sign until, after a finite time, one limit or the other is passed. Thereafter, if equation (8) is satisfied without the equality, (iii) occurs, while if equation (8) fails or if an equality holds, then (ii) occurs. When A = — 1 , alternation between fixed values starts immediately if relation (8) is satisfied without the equal- ity, and otherwise (ii) occurs. Finally if A = 1 , it is evident from the difference-equation (4) itself that the y(y) form an arithmetic progression until one limit is passed, the later terms being identical. Additional essential complications are involved in the discussion of nets consisting of two or more circuits. However, certain simplifi- cations can be performed at once. We wish to determine y at each THE GENERAL NEURAL NET 33 synapse. But if any synapse is the origin of only one and the termi- nus of only one neuron, the two neurons constitute a chain, and after the E is determined for this chain, this synapse requires no further consideration. Again, let a neuron N form a synapse with two or more neurons Nx , N2 , • • • . The results are the same if we suppose N replaced by two or more neurons N7, N2', • •• , with identical prop- erties all originating at the origin of N , but N* forming a synapse with Nx alone, 2VY with iV2 alone, ••• (Pitts, 1942b). Thus the only synapses requiring separate consideration are those at which two or more neurons terminate. Each of the synapses of the set under consideration is the terminus of two or more chains which originate at other synapses of the set, and, the distribution of stimulation be- ing fixed, each chain is characterized by the values of its set of four parameters. In case the terminus of any of these chains is inacces- sible from its origin, the a which it produces is calculable indepen- dently of the value of y at its origin, and we may delete this chain and add this a to the 5 at the terminus. If there happens to be only one other chain terminating there, this can be combined with the chain or the chains originating there and the synapse dropped out of the set being considered. We therefore suppose that each synapse is accessible to the origin of every chain which terminates there. There is also the possibility that when the S applied at any synapse is increased by the maximum a that can be produced together by all the chains which terminate there, this is still below the y', or that the S increased by the minimal a of all together is above the y" of some chain which originates there. If so, the a produced by this chain can be calculated at once, the result added to the S applied there, and the chain deleted. It is clear, of course, that these dele- tions, which are made possible by the inaccessibility of one synapse to another, will be different for different distributions of stimulation. Having performed these simplifications, we suppose, before pass- ing on to the most general case, that only one synapse remains. The resulting net, which consists of a number of circuits all joined at a single common synapse, we shall call a rosette (Pitts, 1943a), and the common synapse, we shall call its center. Now if the conduction- time is not the same around all these circuits, we may nevertheless, with sufficient accuracy, regard these times as commensurable, and we shall use their common measure as the time-unit. Let n be the number of circuits, and let //, be the conduction-time of the i-th circuit. Now consider the contributions of the i-th chain to the stimulus y at s at any time. If y(t) is the total stimulus at time t , then the contribution at time t + m is Si + Aiy(t) when y{ ^y(t) ^yj', 34 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM the y(t) in that expression being otherwise replaced by the nearest limit. Let us introduce quantities at(t), fti(t) denned as follows: 0Li(t)=l when jf(f)=Vo uj(O=0 when y(t)yi". Then the contribution of the i-th chain to y(t + //;) may be written Si + A^imWfiiit) y(t) + tl-aiW] y{ + [1 -&(*)] 1/i" }. If we introduce the operator E defined by Ey(t) = y(t + l), let fi represent the largest of the jut , and set P§ = (i — fij , then we have finally E*y{t)=Z + ^AiE» {*dt) pi(t)y(t) + [1 - a,(t)] yi (11) + ii-fawiyn. The functions a and /? are constant except when y crosses one of the boundaries y' or y" associated with the corresponding chain. Hence the difference-equation (11) can be solved on the assumption that the a's and /3's are constant, and the solution is valid as long as it lies on the particular interval associated with the assumed values of the a's and /3's. If the numbers y,' and y" are arranged in order, they limit at most 2n + 1 intervals (two of them infinite), and each interval is associated uniquely with a particular set of values a, , fit . No other set of the a; , /?/ is possible. Associated with each of these sets a* , /J» , is a unique value of y satisfying [l-2A»ai0i] y-^ + 24i [(1-002//+ (l-fii)yn (12) provided the coefficient of y is non-null. This defines a possible equi- librium of the difference-equation. However, if this value of y does not lie on the associated interval, then no equilibrium for the gen- eral equation (11) exists on this interval. If, for the set a, , /3; , the solution y of equation (12) does lie on the associated interval, the solution y(v + t) of the difference-equation (11) corresponding to the a; , /?, equal to these constant values differs from this constant value y by a sum of terms of the form p(v)Av, where p(v) is a poly- nomial in v multiplied, possibly, by a sine or a cosine, and a is a real THE GENERAL NEURAL NET 35 root or the modulus of a complex root of the equation a* ~ 2 Ai ai fr xp* = 0 . (13) As before, v is an integer for which t = V + T (O^T^l). Hence the equilibrium is unstable unless every root of equation (13) has a modulus less than unity. In case for any of the intervals the coefficient of y in equation (12) vanishes, this equation has no solution unless the right member also vanishes. But then equation (13) has a root unity and the corre- sponding solution of (11) involves a simple polynomial of non-null degree, so that no stable equilibrium occurs. Thus, in brief, in order for any interval to possess a stable equilibrium, it is necessary and sufficient that the solution y of equation (12) obtained from the associated set a, , pt , shall lie on this interval, and that the equation (13) shall have every root of modulus less than unity. Fluctuating equilibria of the sort met with in the simple circuit are here possible, and also another sort arising from possible complex roots of the char- acteristic equation (13) and leading to terms involving sines and cosines. In the general case, let the synapses Si and the chains CK be sep- arately enumerated, and let us define two sets of quantities PjK and QjK as follows : PjK = 1 if Sj is the origin of CK , Pjk = 0 if Sj is not the origin of CK , QiK = 1 if Si is the terminus of CK , QiK = 0 if s» is not the terminus of CK . All simplifications as described above having been made, we can de- fine for each synapse s< a quantity Zi = Si + ^QiKEKt and for each chain CK the sets of quantities aK , pK with values 0 or 1 according to the value of y , at the origin CK , relative to yK' and yK". Then the difference equations satisfied by the yt have the form (Pitts, 1943a) E* yi = Zi + 2 Q^ AKE»«{(1- aK)yK' (14) + (1 -Pk)v" t- Ok & 2^*2//}. If there are n, chains originating at Si , the a's and p's associated with these chains are able to take on at most 2/1* — 1 different sets of val- ues, and there are therefore at most U (2nt — 1) sets of values for 36 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM them altogether. Each set of values is associated with a region in y-spa.ce which may contain a single point whose coordinates yi rep- resent a steady-state of the net. For this to be so (the equilibrium being stable) two conditions must be fulfilled: The constant y-x de- fined by equations (14) when the aK and /5K are given these values and the operator E is taken to be the unit operator must define a point which lies in this region ; and a certain algebraic equation (the char- acteristic equation of these difference-equations) must have only roots whose moduli are less than one. In principle, therefore, the steady- state activity of nets of any degree of complexity can be determined, though admittedly the procedure could become exceedingly laborious. Thus given three synapses, joined each to each by a total of six chains, 27 regions in ?/-space may exist and require separate consideration as possible locations of equilibria. Moreover, persistent fluctuations may arise, no steady-state being approached at any time. While the solution of the direct problem of describing the output of any given net is complete, at least in principle, the general inverse problem is still open. However, in the special case where the output function is such that the a's and /5's remain constant, Pitts (1943a) has shown how to construct a rosette to realize this function. This concludes our purely formal discussion of neural structures, and we turn now to some special structures and their possible rela- tion to concrete types of response. PART TWO VI THE DYNAMICS OF THE SINGLE SYNAPSE: TWO NEURONS Thus far we have been concerned with the formal development of methods for determining the activity of structures composed of neurons. We shall now attempt to make application of the theory and method to experimental problems. Two paths are open to us. We could, on the one hand, examine specific neural structures, seeking' to determine for each the response which it mediates as a function of the stimulus, or we might start with this function and attempt to construct a suitable mechanism. In this and in succeeding chapters we follow the first course. In the final chapters of this Part II we follow the second course. The immediate problem is considered solved if, from the theoretical structure, quantitative relations are derived which agree with the experimental data within a suitable margin of error for some range of the variables in question, and if the number of parameters is not too large. Now many of the parameters may be explicit functions of certain other variables which have been kept constant throughout the experiment. Thus in many cases, the struc- ture will have different properties when the constants of the experi- mental situation are changed. In these cases we may say that the structure studied makes predictions regarding activity outside the domain of activity intended to be covered. Such predictions may sug- gest an experimental approach not otherwise evident. If the predic- tions are borne out, the theory is immediately extended in its applica- tion. If not, the structure must be extended or revised in such a man- ner as to include the old as well as the new properties. Thus on whichever course we set out, whether working from mechanism to behavior, or from behavior to mechanism, we are led finally into the second course when extensions are required. In this we are guided by a consideration of the elements without which there could be no correspondence between the activity of the structure and the activity observed. If certain of the observed elements interact, then the elements of the structure must be inter-connected. If the action is unilateral in the experimental situation, a unilateral con- nection may suffice. If the observed activity depends on the order of the events of the past, the structure must contain elements which ex- hibit this property of hysteresis. Thus one is limited to a consider- able extent in the choice of mechanisms to be studied. 37 38 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM For first application we choose the simplest structures, working gradually to those of increasing complexity. We shall find in the present chapter how a very simple mechanism will serve for the inter- pretation of such superficially different sorts of data as those con- cerning the occurrence and duration of a gross response, just-dis- criminable intensity-differences, adaptation-times, and fusion-frequen- cies in vision and perhaps other modalities. In general, even where the structure is relatively simple, it is not possible to solve in closed form the equations resulting from this structure. Thus certain re- strictions upon the parameters may have to be introduced in order to obtain a workable, even if approximate, solution. As the choice of the restrictions is somewhat arbitrary, one should keep in mind that other equally plausible restrictions could lead to different results and increase both the accuracy and the scope of the theory. The simplest structure which can be studied is a single neuron, and the simplest assumption that can be made about its activity is that it is of the simple excitatory type. Its activity is determined when we have evaluated e(t) for any 5 . However, one does not ob- serve e but some response R . Thus the simplest structure in which we can deal with observed quantities is a chain of two neurons, the first being acted upon by some stimulus S , and the second, which may be a muscular element, capable of producing some response R . The response R is produced as soon as e reaches the threshold of the sec- ond neuron. Hence if we set h = s(t) and solve for I , then since the function s(t) depends upon S through the function $(S) (chap, i, equation 1), we obtain the reaction-time t1(S) as a function of the intensity S , this time being measured from the application of the stimulus until e reaches the threshold. For this purpose we use as given by equation (6) of chapter i and assume 1 h tl = --\0g (1 ), (1) a ft log S/h, where hx is the threshold of the afferent neuron. This relationship should apply to an experiment on a simple re- flex in which a stimulus of intensity S produces a response R after a time tr . However, the total time tr from the application of the stimulus until occurrence of the response as registered by the timing- device involves, in addition to t1(S), also a time t0 which measures the time for conduction plus the time required for the muscular re- sponse to effect the recording instrument plus any other time of delay which does not depend appreciably upon S . We may then expect the equation SINGLE SYNAPSE : TWO NEURONS 39 tr = to lOg (1 a h -) piogS/K (2) to represent the experimentally determined relation between tr and S . The extent to which this does so in some cases for which data are available may be seen in Figures 1, 2, and 3 where experimental data BERGEN ANO CATTELL VISUAL DATA O SUBJECT B • SUBJECT C STIMULUS INTENSITY S- Figure 1. — Comparison of theory with experiment: dependence of delay of reaction upon intensity of the stimulus for visual stimuli. Curves, theoretical predictions by equation (2) ; points, experimental. (Visual data from Cattell, 1886.) Abscissa, intensity (on logarithmic scale) of stimulus; ordinate, interval between presentation of stimulus and occurrence of response. (points) and theoretical predictions (curves) are shown for each of a number of rather different types of stimuli. The details are given in the legends. In general, we cannot expect that the chain from receptor to effector involved in the reflex will contain so small a number of ele- ments. But certainly this demands first consideration since it is the simplest possible mechanism. And even if the chain were known to contain a larger number of neurons, the slowest synapse in the series will tend to govern by itself the temporal form of the response, so that if the remaining synapses are relatively fast, the equation just deduced will still provide an adequate description of the experimental situation. If a stimulus S is presented for too short a time t , e will not 40 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM have reached the threshold h at the end of this time. But for a given S there is a minimal t = f at which time e = h . This i (S) is the minimal period of stimulation with the intensity S which just suffices to produce the response. We obtain for this an equation similar to STIMULUS INTCNSITY S - Figure 2. — Comparison of theory with experiment: dependence of delay of reaction upon intensity of the stimulus for auditory stimuli. Curves, theoretical predictions by equation (2) ; points, experimental. (Auditory data from Pieron, 1920). Abscissa, intensity (on logarithmic scale) of stimulus; ordinate, interval between presentation of stimulus and occurrence of response. STIMULUS INTENSITY S - Figure 3. — Comparison of theory with experiment: dependence of delay of reaction upon intensity of the stimulus for gustatory stimuli. Curves, theoretical predictions by equation (2) ; points, experimental. Gustatory data from Pieron, 1920.) Abscissa, intensity of stimulus; ordinate, interval between presentation of stimulus and occurrence of response. SINGLE SYNAPSE : TWO NEURONS 41 equation (2) , but with t0 = 0 and tr replaced by t . Thus from a con- sideration of a chain of two neurons one should expect that if all other conditions remain unchanged the same relationship should hold in both cases, except that t0 would be absent in this case. Since the two cases are experimentally distinct, it may be that the results from the two types of experiments are widely divergent. If so, it may be necessary to assume that there are several neurons in the chain or even circuits in the structure. In any case, the kind of disagreement may suggest the nature of the change to be made in the neural net. Let us consider another special case of a chain of two neurons. Let the afferent neuron be of the mixed type with <£ = y> , a > b , and threshold hx . A constant S > hx applied to such a neuron results in a a (= e — j) which is positive, but which vanishes asymptotically. Then, the stimulus being presented at t = 0 when e = j = 0 , if S is large enough, and h not too large, a will first reach the value h at some time tf . From this one can determine a relation tr(S) similar to that of equation (1). If S is maintained at a constant value for a sufficient time, t + t let S be replaced by S + AS, where A S may take on any positive value. Negative values of A S would be of in- terest only if tr < t* < tr + r . Then for t > t* > t r + T we may write + 4>(S + A S) [e-&<<-'*> - «?-«(«-«•>] . If A S is large enough, a will again reach h at some time t — t* + tr' and the response will again be initiated. By setting a = h in equa- tion (3), we obtain tr'(AS, S, t*) from the smaller root, t. For t* > > 1/6 and tr' < < 1/6 , we may obtain an equation for tr' which shows the time V to depend only upon A S/S , and not upon hx . At some time t = t* + tr' + t the response will again cease. Using the larger root of a = h in expression (3) we may determine the duration T'(AS ,S ,t*) of the response. For t* > > 1/6 and r > > 1/a, we may write T' = -log[/log(l + zlSyS)] -tf', (4) 6 / being a constant. For fairly large values of A S/S we may neglect tr' in equation (4). The equation thus makes a definite testable pre- diction as to the nature of the relation between the duration of the response and the relative increase of the stimulus. 42 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM If we let t* be large and if, for a fixed S , we restrict A S to the least value it can have while U remains finite (i.e., the two roots tr' and tr + t of equation (3) coincide), we obtain, if we use equation (6) of chapter i, AS/S = d = d0, (5) da being a constant. That is, when a constant stimulus of intensity S > hx has been applied for a long time, the smallest additional stim- ulus A S necessary to produce the response must be a constant frac- tion of the intensity of the original stimulus S . On the other hand, US < hi we have 6= (S0 + l)hJS- 1. (6) Thus d(S) decreases hyperbolically from oo to S0 as S varies from zero to hx ; thereafter S is a constant. The quantity <3 of equations (5) and (6) is essentially a Weber ratio, and its variation with 5 as de- scribed in the above equations has the chief qualitative properties of the experimental relation for most types of stimuli. This problem will be discussed in more detail subsequently (chap. ix). Suppose that instead of replacing S by S + A S at time T, we remove S for a time t' after which only A S is presented. This is the experimental technique for studying the processes of adaptation and recovery. Then for t > t* > t' , CT=z<£(S) [e-°<*-'*> - r*<*-**} + e-bt - e-at] (7) At the time t*, we shall suppose a < h . Hence at some time t = t* + t' + t" r , if A S is large enough, o- = h , and the response occurs. From this relation, together with equation (7), we can determine the reac- tion-time t"r(AS , S , t*, V) from the smaller root, and the duration t" (AS , S , t\ t') of the response from the larger root, at any stage of the process of recovery following preadaptation to the intensity S . We can further determine the minimal AS required for stimulation as well as the minimal interval of exposure at a given AS . ' If t* > > 1/b , t' > > 1/a , t' > > tr" and tr" < < 1/b , we have cl>(AS) (1 - e-"*r") - ^(S) e-w - h = 0. (8) Thus tr" = - -log( 1 - *±-L—— , (9) so that the reaction time tr" increases with S but decreases with both A S and t' . SINGLE SYNAPSE : TWO NEURONS 43 Equation (9) makes a definite prediction as to the relationship between the reaction time and the variables S , A S , and t' and is thus subject to experimental test. If, for fixed S and t', we now restrict A S to the smallest value for which the response can still take place, we obtain a relation of the form AS S log = e-bt'log — , (10) h' K where log K = log hx + h/(i . Thus, except for minimal if, the loga- rithm of the testing stimulus A S is an exponentially decreasing func- tion of the time t' of recovery. As the time t' becomes infinite, A S approaches h'. The intensity S determines by how much the ordinate is multiplied in the graph of A S against P. The type of relationship between A S and t' of equation (10) for the case of visual stimuli is found in the work of various investigators (cf. S. Hecht, 1920). Suppose next, still assuming that = y> , that any constant stim- ulus has been applied for a long time and that at t = 0 the stimulus is increased at a rate such that cUf>/dt = X . After a time t , we find that X X = t/' • Let a stimulus S be given for a period of time rT followed by no stimulus for a period of time (l—r)T. Let this be repeated indefinitely. For each successive interval we can determine e and j from the differ- ential equation together with the requirement that £ and j both be continuous. The value of a(t) during the interval 0 < t < rT of stimulation after a large number of repetitions may be obtained as follows. Let e„-t represent the value of e at the beginning of the n-th resting period, and e'n-i the value at the beginning of the n-th period of stimulation, where e0 — 0 . We use equation (8) of chap- 44 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM ter i, replacing c0 by c'M_i and t by rT to calculate e»_i , and we use the same equation to calculate e'„ by setting = 0 , replacing s0 by £„_i , and £ by (1— r)T . When we do this we find by simple induction that e'B = 0(l - e-^) (1 - e-naT)e-a^T/(l - e-"T) , and as n becomes large the exponential containing it can be neglected. The expression for j'n , similarly defined, is the same with b replacing a. To obtain j(t) during the interval in question, we need only re- place j0 by j'n and b by a in the same equation (8) . After taking the difference e — j and performing elementary algebraical simplifica- tions, we obtain finally the desired expression: — A p-bt a = <},e / 1 - e-hil-r)T \ / 1 - e-a^-r)r \ Now a reaches a maximum at t = t* given by t* = log a a(l- c*<1-r>f) (1 - e-bTl 6(1 - e-^-'^Ml - rT , in which case the maximum value of a is a(rT). Set a = h in equation (12) with t replaced by rT or t* according to the case. Then for given r and T , that value of 5 which satisfies the equation is the least stimulus that will produce a steady response when repeated in this manner. For this r and S , let T* be the particular value of T employed. Then f = 1/T* is a critical frequency separating response from no response for the value of 5 in question. Then, if = ft log S/hi and H = h/0, for rT* < t* or f > r/t*, I 6-°' V ~~1 brT* _ g-br* -bT* irT* (>-aT* \ S log- = H, 1 - e-aT* I K (14) and for rT* > t* or /* < r/t*, b 1 -- e~aT* b -a(l-r)T* -. -a- 1 - h', there results a response to flick- ering (intermittent) illumination, whereas when S < h' there results no response, h' being a constant which is the effective threshold. Thus a plot of /*(log S/hx) begins at (log h'/hlf 0), rises vertically at first, then flattens off while approaching a final slope which depends upon r . The relationship between f* and r is generally determined for a constant apparent brightness given by S' = Sr . Using the approxi- mate expression (16) with S' = constant, we find that f*(r) rises rapidly from zero to a maximum for r < ^(LDR < 1) and then falls to zero for r = 1 (Figure 4). The position and the height of the .4 Figure 4 .8 1.0 maximum depends upon S". However, when r is near zero or unity, the approximations break down. Furthermore, equation (16) holds only for large enough /*. Hence the exact relation f*{r) may be of considerable complexity. For the various experimental relationships, one may consult Bartley (1941). Most of the results quoted agree with the above prediction that for constant S', f*(r) is decreasing in 46 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM r when r is just less than one-half (LDR just less than unity) . More generally, if instead of alternating a stimulus 5 with no stimulus, we had alternated S + AS with S , we should have obtained the same results with log S/lh. replaced by log(l + A S/S) unless S < hx . From this it is clear that for a constant S + A S , an increase in S > h^ decreases the critical frequency f*. Similarly, an equal in- crease in both S and S + A S decreases f*. That is, an illumination added to both phases, as from stray light, decreases the critical flick- er-frequency. Although we have referred to visual phenomena only, one may well expect that analogous properties of some other modalities also could be accounted for roughly by just such a simple mechanism as the one considered here. We have assumed throughout that a > b and

, and let R$ = xp , with R a fraction having a value between zero and one, we find, that for constant S , o- increases to a maximum and decreases to a constant value (1—R)$. This corresponds to behavior of the continuously acting elements of the retina. We proceed to consider some properties exhibited by a neural element of this latter type. We have now a chain of two neurons, the afferent member of which is of the mixed type with — xp/R , 0 < R < 1 . Let the stimu- lation again be intermittent, of frequency f = 1/T and fractional stimulation r . Let the intermittent stimulation be continued indef- initely. The value of a at the end of each period of stimulation, that is, oo and t = rT(mod T) , can be determined in the manner described above. If 6 is that value of a divided by a ( co ) for r = 1 , 6 is essentially the ratio of a at the end of a period of stimula- tion for a particular / and r , to the value which a would have if a stimulus of the same intensity were applied continuously. 0 would be better defined as the ratio of (a - ^)/(tr00 - h), but we shall neg- lect the threshold as compared to a . We may then write SINGLE SYNAPSE: TWO NEURONS 47 6 = 1- R 1 - e*rT -brT -uT R o-bT (17) Equation (17) gives a relation between the relative net excitation 8 , the fraction r of time of stimulation, and the frequency / = 1/T . For / = o , 6 = 1 f or all r . But for / = oo , 6 = r . Furthermore, 6 in- creases with / for small / and the height is greater for small r . The type of relationship between 0 and / for various values of r is shown m e 48 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM in Figure 5. Notice that the maximum moves to the right with in- creasing r . However, if instead of the 8 used, we had taken the aver- age value over the interval rT , we should have obtained essentially the same results, but with the maximum moving to the left. The equation corresponding to expression (17) is, however, much more complicated. The quantity 6 suggests immediately, in the terms of the visual field, the relative brightness during flicker. Experiments by Bartley (1941) show essentially the same type of variation as ex- pression (17) but there is no significant change in the position of the maximum with r on a range from \ nearly to 1. From what has been stated above, one could probably find a simple average which would give this result. If cj) is proportional to S over the range to be considered, then multiplying S by 1/6 would make the responses equal. For the fre- quency / very large, 6 = r , that is, S must be increased to S/r to ap- pear the same as a continuously applied S . This is just a statement of the Talbot law. From this brief consideration of the dynamics of two-neuron chains we have been able to derive equations predicting quantitative relations among various experimental variables. These include the relation between reaction-time and intensity of stimulation, for given change in stimulation and period of accommodation. Similarly, the duration of response is determined in terms of these same variables. A Weber ratio is determined as a function of accommodation-time. Furthermore, a relation is determined connecting flicker-frequency, light-dark-ratio and intensity. And finally, relations are determined between relative brightness and the light-dark-ratio and frequency. In some cases, quantitative agreement with experiment is exhibited. In others, general qualitative agreement is obtained. It is well worth noting at this point that while, on the one hand, to the extent that the formulae are verified these manifold relations are all brought within the scope of a single unifying principle, on the other hand the dis- cussion explicitly introduces a great many problems which the experi- ments only vaguely suggest. VII THE DYNAMICS OF THE SINGLE SYNAPSE: SEVERAL NEURONS When a pair of afTerents, instead of the single one assumed in the preceding chapter, form a common synapse with a single efferent, the resultant a at this synapse is capable of varying with time in a much more complicated manner. We shall consider briefly two pos- sible applications of such a mechanism, one in which both afferents are supposed to be affected by the same stimulus, one in which the stimuli are assumed to be different. Consider first the very special case in which both afferents are stimulated by the same constant stimulus. Let one of the afferents be of the simple inhibitory type with the associated yu and 6T . Let the other be of the mixed type with 2 , xp2 , a2 and b2 . Let a2 > > bx or b2 and let 2 — y\ — y^ = h , the threshold of the efferent. These assumptions are made to reduce the number of parameters. We em- ploy equation (8) of chapter i, with its analogue for j . Then, a2 be- ing large, the term e-a*[ quickly dies out so that except for very small t , a - h = ?/>! e~bit + y<2 e^1 . (1) But the frequency v of response (chap, i) in the efferent neuron is proportional to a — h when this is not too large. Since y, and y>2 are arbitrary, w e may replace a — h by v . We can then attempt to inter- pret equation (1) as giving the variation with time of the frequency of the response of an efferent neuron when a constant sustained stim- ulus is applied to its afferents. Equation (1) is shown in Figure 1 (curve) for particular values of the constants while in the same figure are shown the results of experiments by Matthews (1931) (points). In these experiments the stretch receptors in muscle were stimulated by means of attached weights and the variation with time of the fre- quency of discharge was determined. Since experimentally the weights tend to sink with time, one might separate the stimulus into two parts, a constant part acting on the second neuron and a variable part acting on the first, making this one excitatory with large a. The variable part would presumably be roughly an exponentially decreas- ing function whose decay-constant must be equal to 6, of equation (1). This, also, would lead to equation (1). Whether or not the ac- tual rate of decay corresponds sufficiently with the experimentally determined decay-constant is then a question of fact. Formally, we 49 50 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM u LU T O x >- u z h^. and S2 > h2 be constant stimuli applied at times t± = 0 and t2 to the neu- rons Nx and N2 respectively. Let h be, as before, the threshold of the efferent neuron. Also let ex , e2 and j2 be zero initially. Then the effer- ent neuron is excited at the time t' when a = ex + e2 — j2 = h , or [1 -&2< t'-h)]=hm (2) ^(1 - e-^') + & [1 - e^»(*'-*.)] - xp2 The value of t' depends on Sx , S2 and £> • To simplify the problem further, let <£2 = y*2 and let o-2 be always less than h . This can be done readily by restricting S2 or requiring that cj>2 < h . Then no response can occur prior to the moment t = 0 , even if t2 < 0 . If we set tw = t' — t2 we obtain by solving equation (2) SINGLE SYNAPSE: SEVERAL NEURONS 51 t' = --log a, h , fa(S2) ^ (e-w, -a2 1 fa(Sx) 0i (50 ■) (3) Finally if we set £r = t0 + £', where t0 is a constant as in equation (2) of chapter vi, we can determine the total reaction time tr as a function of S2 through fa , Si through fa , and of tw , which differs by t0 from the time by which S2 precedes the initiation of the response. As S2 is a stimulus which precedes Sx and affects the response time to Si , but is itself incapable of producing the response, it may be considered a warning stimulus. Hence we may take equation (3) to predict the kind of results to be obtained in an experiment in which a particular stimulus of intensity Si has been preceded by a warning stimulus S2 and produces a response in a time tr(S1} S2 , tw) depending on the strength of the warning stimulus as well as upon the manner in which Sx and S2 are spaced in time. For the particular case in which a fixed Si and S2 are used and for tl0 > > tr , we may write equation (3) as tr = t0' log [1 + D(e-6*<- - e-°»'»)] a, (4) in which U' = t0 log (1-h/fa) > D = fa/ (fa - h) 12 16 20 2* PREPARATORY INTERVAL IN SECONDS tw — * Figure 2. — Comparison of theory with experiment: effect of time of occur- rence of warning stimulus upon the reaction-time. Curve, theoretical predictions by equation (4); points, experimental (Woodrow, 1914). Abscissa, interval be- tween presentation of warning and effective stimuli; ordinate, interval between presentation of effective stimulus and occurrence of response. 52 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM In Figure 2 is shown a comparison of equation (4) (curves) with experimental data (points) by H. Woodrow (1914) in which the effect of the interval between warning stimulus and stimulus proper on the reaction-time was measured. The upper curve was obtained under the condition that the successive values of tw in the experiment were mixed randomly; the lower curve was obtained under the condition that the value of tw was kept the same for a number of trials before being changed to another value. As the conditions are different in the two cases, one might expect that the parameters would also be differ- ent. Further details are given in the legend of Figure 2. For a dis- cussion of a mechanism which can differentiate between these two conditions, the reader is referred to Landahl (1939a). VIII FLUCTUATIONS OF THE THRESHOLD It has been assumed thus far that the threshold of a neuron is a constant which does not depend on time. Actually it varies from moment to moment and when we speak of the threshold as a constant we must understand by this some mean value of a group of measure- ments of the threshold. A more complete description would give also a measure of the variability. The threshold may vary with many changes in the organism. These variations would generally be rather slow. But within the neuron and in its immediate surroundings there occur rapid minute fluctuations in the concentrations of the various metabolites. The work of C. Pecher (1939) indicates very strongly that it is these fluctuations in concentration that are responsible for the variations in the thresholds of the peripheral fibers with which he experimented. His calculations showed that as few as some thous- and ions was sufficient to produce excitation. From the kinetic theory, one should then expect that the per cent variation in the threshold should be one hundred divided by the square root of the number of ions necessary for excitation (Gyemant, 1925). This value in terms of the coefficient of variation is of the order of a few per cent and is comparable with the values obtained experimentally. We may make the calculation of the variation as follows. In order for excitation to occur, it is necessary to stimulate a minimal region of a neuron. Suppose this to be a node of Ranvier. Let the width of the node be d oo 104 cm, the radius of the fiber r oo 10-4 cm. The effect of an ion is small at distances of a few diameters. Thus ions a few diameters removed from the cell surface will have little influence on the surface. Let this distance of influence be d oo 107 cm. Then the volume within which the ions affect the excitability of the neuron is 2nrdd . If C oo 10-5 is the molar concentration of the ions, and if N is Avogadro's number, the total number of ions influenc- ing the excitability is 2nrddCN and thus the per cent fluctuation is given by 100 /\/2jirddCN oo 2%. Had we used the area of an end- foot (c\3l0-7), the same sort of result would have been obtained. But because of the variations in these quantities one cannot exclude the possibility that rather large variations may occur. The calculations only indicate that the fluctuations about the threshold may be appre- ciable. As long as the range in variation is not comparable with the threshold itself, the kinetic theory requires that the fluctuations be distributed normally to a high degree of approximation. That this 53 54 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM is the case for single nerve fibers is illustrated in the data by C. Pecher (Landahl, 1941c). If for a particular neuron the mean value of the threshold is h , the coefficient of variability is v , and if p(C) represents a normal curve of unit area, then the probability P of a response in the absence of a stimulus is given by the integral of p(C) from 1/v to oo. If t is the least time for a fluctuation to have effect, then after a time t/P one could reasonably expect a chance response. The mean frequency of such responses would then be given by P/r per second. These re- sponses would not be periodic. If t is taken to be of the order of mag- nitude of 10~3 seconds, then for v = 30% the mean time between re- sponses would be a few seconds, while for v = 20%, the mean time be- tween responses would be a number of hours. From this we see that the probability of a chance response, even over a considerable period of time, becomes negligible rapidly as v becomes much smaller than one-fifth. But one should consider also the slower changes in the en- vironment of the neuron which not only changes the threshold but also its degree of variation. Variations in the threshold would cause the response-times and other measurable variables to be distributed in some manner about the value corresponding to the mean value of the threshold. As an illustration, let us estimate the dependence of a measure of the varia- tion in response-times on the intensity of the stimulus. We consider the case of a simple excitatory afferent stimulated by a constant stim- ulus S and acting on an efferent of threshold h . Suppose that tx is the value of t for which e = h . Then if h is decreased by an amount vh , e = h is satisfied by t = U. — a. Then 1 / vh \ a = -log[ 1+ (D a V <£ — h J is an average variation in the reaction-time due to the variation in threshold. In general, since one must consider more than one chain, one may suppose that variations, essentially independent of , are introduced at other synapses and at the end-organ. Let a0 be a meas- ure of the total effect of this variation. Then the measured variation in the response-time will be given by the square root of the sum of the squares of these two variations. As v is generally quite small a = vh/((p — h)a and thus at = Veto2 + vya*{4>/h - l)2, (2) and we have a relation between a measure of the variation in re- sponse-times and the stimulus-intensity in terms of ^ . We may com- FLUCTUATIONS OF THE THRESHOLD 55 pare this result with the results of experiments by Berger and Cattell (1886) cited in chapter vi, in which the mean variations of the re- sponse-times were measured. We may use the same parameters ex- cept for (70 which is arbitrary for this curve. Thus we have one para- meter to determine this relation. The data are incomplete in the re- gion of the threshold, but the comparison in Figure 1 is made for .040 .030 - O.020 .010 10 100 STIMULUS INTENSITY S— * 1000 Figure 1. — Comparison of theory with experiment : the variation in reaction- times as a function of stimulus-intensity. Curve, theoretical predictions by equa- tion (1); points, experimental (Cattell, 1886). Abscissa, intensity of stimulus; ordinate, mean variation in reaction times. illustrative purposes primarily. Nothing has been said of the type of distribution one would expect for a particular stimulus-intensity. This would require a more detailed analysis. We wish only to indi- cate the kind of effects due to the variations in the threshold. In the next chapter we shall show how they provide a possible basis for the distribution of judgments in situations that require some form of discrimination. IX INTERCONNECTED CHAINS: PSYCHOPHYSICAL DISCRIMINATION In chapter iii we dealt with the general problem of the interac- tions among interconnected parallel chains of neurons, and more es- pecially with the mutual reduction of the a's developed by the simul- taneously stimulated chains when the interconnections are inhibitory. We also exhibited a mechanism capable of transmitting excitation when the intensity of the stimulus lies on a limited range only. In this chapter we shall apply these considerations to the interpreta- tion of sequences of a stimulus and a response of a type often stud- ied by psychologists, which we shall speak of as discriminal se- quences. By a discriminal sequence we shall mean any sequence in which the response is one of a limited set of qualitatively different possible responses, while that feature of the stimulus that deter- mines which response of the set is to occur is either its absolute in- m m ti \J4 X S. Jz \ti ft I t % 1 ^ Ml V r> 5 £; $ Figure 1 56 PSYCHOPHYSICAL DISCRIMINATION 57 tensity or its intensity relative to that of some other specified stimulus. We consider, then, parallel neurons or chains interconnected by inhibitory neurons (Figure 1), and we impose here the further re- strictions that their inhibitory effect is numerically the same as the excitatory effect due to the neuron with the same origin. Then a stim- ulus Sx may produce a response Ri , and S2 a response R2 , if the stim- uli are presented separately. But if the stimuli are presented to- gether, then the response Rx will be produced if Sx exceeds S2 by an amount which depends upon the thresholds of the efferent neurons. If the difference between Sx and S2 is too small, neither response oc- curs. If the thresholds are negligibly small, the response Rx occurs alone if 51 > S, and R2 if S2 > S1 . Because of fluctuations of the type discussed in the preceding chapter, the values of s-l and e2 produced by the afferent neurons will not generally be exactly equal even when Sx = S2 . If St is slightly greater than S2 , there is a certain finite probability, less than one- half, that the response R2 will be given instead of Ri , the probability decreasing as the difference between Sx and S2 is increased. Suppose that fluctuations occur only in the thresholds of the afferent neurons. The fluctuation of the threshold of any neuron causes fluctuation of the o- produced by this neuron, and we shall pos- tulate the distribution of o- rather than that of h . Furthermore, be- cause of the interconnections between the neurons, an increase in o- at the terminus of one afferent has the same effect as a decrease in a at the other, so that formally we may regard the fluctuation as oc- curring at only one synapse (Landahl, 1943). Thus we shall assume that the thresholds are constant but that at synapse sx , h , response Ri is given ; 58 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM if h i? £i — £.2 — £ = — h , there is no response ; if — h > £i — £2 — £ , response R2 is given. If p(£)d£ is the probability that £ has a value in the range £ to £ + d£ , then the probability that response Rx occurs is obtained by integrating p(£) with respect to £ from minus infinity to (e^ — e2 — h). This becomes evident when we see that if £ is any value less than £1 — £2 — h , response Rx is produced. If we let f\ be the probability of response Rx , P2 the probability of response R2 , and P0 the prob- ability of neither response, and if we define then we may write P(x) = f* p(£)d£, (2) J -"50 P^ = P(e1-B2-h), (3) P2 = P(-e1 + e2-h), (4) Po = l-P*-Pz. (5) If Si = S2 , response Ri may be considered the correct response and R2 the wrong response. In this case Pr = Pc , the probability of a correct response, and P2 = Pw , the probability of a wrong re- sponse. Any failure to respond, or any response other than cor- rect or wrong such as "equal," "doubtful," could be included in the proportion to be identified with P0 • It is commonly the case that when a categorical judgment is required, so that either Rx or R2 is made at each trial, the subject must lower his criteria for judgment. We may interpret this with reference to the structure studied by assum- ing a lowered threshold. For this case we set h = 0 , whence P0 = 0 and Pi + P2 = 1. Thus from a knowledge of only the standard error of the probability distribution, one is able to calculate the probabil- ities of the various responses to any given pair of stimuli when the judgments are categorical; the additional parameter h enters when "doubtful" judgments are allowed. Since complete symmetry has been assumed, it follows that Px = P2 for Si = S2 . In general, this is not true for the observed pro- portions. The amount by which the observed proportions differ from equality is a measure of the bias of the subject. The simplest inter- pretation of the bias is that the afferent thresholds are not exactly equal so that the mean values £x and £2 are not equal for Sx = S2 , but £x(£) = s2(S) + x0. Although x.0 will depend on Si, we shall not consider this any further, preferring rather to incorporate #0 into £ , so that p(£) has a mean value of x0 instead of zero. Thus modified, the mechanism may be applied to the experimental data by F. M. Urban shown in Table 1 (Urban, 1908). These are the average re- PSYCHOPHYSICAL DISCRIMINATION 59 suits from observations made on seven subjects. The first entry in the table, .012, gives the proportion Pg of judgments that a weight of 84 gms is heavier than the standard, which is 100 gms in every case. This is for the case in which judgment of "equality" is permitted, the pro- portion being Pe = .027. On the same line in the last column is given the proportion Pi = 1 — Pg — Pe of times the weight of 84 gms is .50 Experiment a /.CO Figure 2. — Comparison of theory with experiment: distribution of judgments of relative weights. In 2a the abscissa of each point is the experimental value (Urban, 1908) of the proportion of judgments of the indicated type and the ordi- nate is the proportion theoretically predicted by equations (l)-(5). Thus perfect agreement would be indicated if all points lay on the straight line. In 2b the curve is theoretical, the points, experimental. The abscissa of each point is the proportion of "greater than" or of "less than" judgments as the case may be, the ordinate, the proportion of "doubtful" judgments when comparing the same "vari- able" stimulus with the standard". 60 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM n ID ./> ro o >£> CO ro 00 lO f) n iT) i" CD r S330NVD fc)3l3W Nl 30N3fc)3JJI0 AHSN31NI 1HOH Figure 3. — Comparison of theory with experiment: distribution of judgments of relative brightness. Curve, theoretical from equations (l)-(5) ; points, experi- mental (Kellogg, 1930). The ordinate is the difference between the intensities of the "variable" and the "standard" stimuli, in meter-candles. The abscissa is in every case the proportion of judgments of the type in question: for the solid circle, "greater than" categorical judgment; for the open circle, "greater than" with "doubtful" judgments permitted; for the crosses, the proportion of "doubt- PSYCHOPHYSICAL DISCRIMINATION 61 judged to be lighter than the standard. Directly below the first entry- is the number .020 which is the proportion of times the weight of 84 gms is judged heavier than the standard when the judgment of "equal- ity" is ruled out, and so on for the other entries. In Table 1 (p. 72) are given the corresponding probabilities com- puted by equations (1) through (5), using for the standard deviation of the distribution 5.7 gms, for the threshold h = 2.1 gms, and for the bias or constant error x = 2.75 gms. These values were determined from the three values in parentheses on the left, whence the corresponding values on the right are the same. It is to be noted that the parameters are all measured in grams. This is done for convenience only, as ac- tually there is an unknown constant which must multiply each para- meter to give the values in terms of e and j . Furthermore, a linearity between S and e is implied, which is nearly the case in the small range considered. In both the Table and Figure, the proportions Pg , and Pi are used. These are respectively the proportions of judgments of "greater," and "lesser." Their relation to Pc and Pw is evident. The agreement between the theory and experiment is illustrated in Figure 2. Complete agreement would be indicated if all points fell on the line of slope one. The relation between the proportions of judgments of "equality" and the proportions of the correct and wrong responses is also shown in Figure 2. If these results are plotted on probability paper, the predicted results will be simply three parallel lines. The experimental data confirm this rather well (Landahl, 1939b) . The results from each of the seven subjects showed the same trend. When one considers the visual data by F. M. Urban, one finds that the averaged data for several subjects, as well as those for individ- uals, cannot be so simply interpreted. With judgments of "equality" allowed, neither the proportions of correct responses nor those of wrong responses follow the integral of the normal distribution with- in the limits of experimental error; with judgments of "equality" ex- cluded the proportions are not peculiar. This suggests that the thresh- old h is not constant but is affected by the value of S2 must be supposed the same as that of S2 > &i if St and S2 are simply inter- changed. We may suppose that the lowering of the threshold h , due to the change of experimental situation when judgments of "equality" are ruled out, is the result of the activity of some outside group of ful" judgments has been added to the proportion of "greater than" judgments. In the inset the curve represents the values of h predicted by equation (6) plotted against the difference of the intensities of the two stimuli and the points repre- sent the values computed directly from the data. 62 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM neurons. If a neuron of negligible threshold having afferent synapses; at sx and s2 tends to excite this outside group of neurons, then the threshold h will decrease with the absolute value of ux or S2 is not the same as that for which S2 > 0 , we obtain the curves shown in Figure 4. The points are the experimental values obtained by averaging the results from a number of subjects. The inset shows the relationship between h and log Si/S2 which is decidedly asymmetric. If this be considered significant, one might attempt to correlate the asymmetry with the mode of presentation or perhaps with the modality. The. PSYCHOPHYSICAL DISCRIMINATION 63 o o _j o l" O O O -I _J U >■ o a < .SinOAIllllAI Nl A1ISN31NI QNOOS ^t * ■<* rO '-D 1 i i 1 o rvj 1 o iO 1 fi) 1 oS 1 1 o '" * 1 i" 0» O) 00 o 0> o 00 o r- o I* o z S°- o o _ o rf> - AJ % 001 Figure 4. — Comparison of theory with experiment: distribution of judgments of relative loudness (data from Kellogg, 1930). The representation is the same as that in Figure 3 except for the use of the logarithmic scale of intensities, necessitated by the large relative range in the stimuli. 64 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM asymmetry appeared fairly clearly in the results from each of the individuals taking- part in the experiment. On the basis of the mechanism considered, it is essential that the stimuli be presented simultaneously. However, a complication of the mechanism has been considered for which simultaneous presentation is not necessary. Essentially the same results may be obtained if the stimuli are presented in succession (Landahl, 1940a). Since the integral of the normal distribution cannot be given in closed form, it is convenient to introduce an approximation by which closed solutions can be obtained. This is especially desirable when one wishes to use the results in other situations. With the distribu- tion (Landahl, 1938a) p(C)=-e-fci &.., log2Pw + k(ei-e2) =0 (8) to determine the probability of a wrong response when a categorical judgment is required. The applicability of this approximation has been tested by a comparison of theory and experiment made else- where (Landahl, 1938a). While our definition of a discriminal sequence rules out the con- currence of the alternative responses and hence requires that the in- hibitory neurons connecting the parallel excitatory chains shall have activity-parameters at least as great as those of the excitatory chains themselves, it is natural to consider also the case where this restric- tion is removed. We noted in chapter iii, in the case of two parallel chains with inhibitory interconnections, what qualitatively different effects might follow the simultaneous stimulation of both chains as different relations are imposed upon the parameters of the constitu- ent neurons. We select for further consideration here only the sym- metric structure consisting of two parallel chains with crossing in- hibition where a > /5 (cf. equations (4) chapter iii), in which case concurrent transmission along the two paths will occur when the two stimuli are sufficiently strong and not too greatly different. These chains may lead from neighboring cutaneous receptors, from neighboring retinal elements, or from organs of two disparate sensations. The responses which they occasion may be overt bodily movements or they may be merely awareness of the sensations. The mechanism has a possible application wherever there is interference by a stimulus of one type with evocation of a response by another, and reciprocally. It is at once apparent that the interference by the PSYCHOPHYSICAL DISCRIMINATION 65 one stimulus with the other's response may occur even though the first stimulus would be inadequate, if presented alone, to produce its own response. Thus if the application were made to the interaction of auditory with visual perception, the mechanism provides that even a subliminal auditory stimulus would raise the absolute threshold for visual stimulation. With appropriate modifications — crossing excita- tion instead of crossing inhibition — the possibility of a mutual lower- ing of threshold could be similarly treated. To link the mechanism with crossing inhibition more substan- tially with possible experimental results, we consider the effect of threshold-fluctuations, or, what is more convenient and mathemati- cally equivalent, random variations in the o-'s at sx and s2 . Since a^j?we cannot, as before, represent the combined effect by varia- tions at only one synapse, but any variations occurring also at Si and s2 could be formally accounted for by suitably modifying the distribution-functions at the first two synapses alone and we suppose, for simplicity, that with this modification the resulting distributions are identical at the two synapses. Denote these functions by p(C), C being the random addition to either synapse and having zero as its mean. The mutual influence of the stimuli upon absolute thresholds can be determined from an investigation of near-threshold stimulation where the functions and xp can be represented linearly: (S) =aS-a',y>(S) =ps-p, a > p . (9) Corresponding to equations (1), chapter iii, we have a1 = a(S1 + Cx) -a'-/?(S2 + C2) + /J', (10) 0, (11) aS2- pS1 + a^~ PCi~h>0, (12) where h = a'- p + h' , (13) and each efferent from sx' and s2 has the threshold h' . Let (RR), (RO) , (OR) and (00) denote the occurrence of both responses, the first only, the second only, and no response, respectively. The prob- abilities of these events, where Si and S2 have given values, are P(RR)= f p(Ci) J f»(f.)(Ci) JV(C2)<#3<#i, -oo -oo -Si+fe/(o-|3) /3(fi+S!)/a+Va-S2 -S1+7i/(a-(3) (3(fi+S1)/a+ft/a-S2 P(00)=J 2?(d) J"p(C*)#»#i. (17) -oo a(£i+Si)/0-/!/0-S2 Other expressions for P(RR) and for P{00) can be obtained by interchanging- subscripts. These four P's are functions of Sx and S2 whose values are experimentally determinable; their sum is unity so that only three are independent. If p(C) is given they depend upon the parameters a , /5 and h; if the distribution 29(C) is assumed to be normal there is an additional parameter, the standard deviation. Pear- son's tables of tetrachorics can be utilized for determining these para- meters from the empirical frequencies. The quantities St and S2 are not the intensities of the external stimuli but some monotonic func- tions of these; however, at the near-threshold level it is permissible to regard these functions as linear. In the preceding paragraphs we have considered the case of two stimuli simultaneously presented to the organism, with two alterna- tive responses permitted. In chapter vi we considered the case where a response may follow the sudden increase in the intensity of a stim- ulus previously maintained at a constant level. We turn now to a process of more complex form in which each of a group of stimuli differing only in intensity elicits a distinct response. This involves absolute discrimination though a similar mechanism may be at work when relative discrimination occurs. The important point to notice is that an increase in the intensity of the stimulus does not merely change the strength of the response or bring into activity additional elements, but may so alter the response that none of the elements involved before the change is included among those active after the change. Consider the net of Figure 5 which is a parallel interconnected structure containing also circuits (Householder, 1939b). Let an affer- PSYCHOPHYSICAL DISCRIMINATION 67 ne Figure 5 ent neuron form synapses with a number of neurons ne whose thresh- olds differ but which are otherwise equivalent. Let all those which have the same threshold be brought together to act on a single neuron. Only two of an indefinite number of final neurons are shown in the Figure. Thus all the neurons ne having the threshold hk are brought together to act on a single neuron ne3 . Let f(h)dhhe the number of these neurons having thresholds between h and h + dh . Consider- ing only the stationary-state activity, and using the linear relation- ship of equation (5), chapter i, with ft = 1 , we may write s(S,h) = (S-h) f(h) (18) as the value of the excitation at st , the terminus for the neurons ne with the threshold h . Let equivalent inhibitory neurons of negligible thresholds origi- nate at each synapse s» and terminate at every other synapse s, with- out duplication. If a(S,h) is the value of a at the synapse Si at which the neurons of threshold h terminate, then the value of j(S) at this synapse, due to the activity of the neurons ni terminating there, is j(S)=Xfo(S,h) f(h)dh, (19) X being a constant measuring the activity of the inhibitory neurons. Thus j(S) so that 0 . That is, the neurons ne3 corresponding to thresh- olds in the range hx to h2 are excited. Thus we may write 68 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM j(S) = X fh*/• 002 • .000 1 i i - . i ■ , .005 .01 .02 .05 e .10 .20 .50 Figure 9. — Comparison of theory with experiment: discrimination of lengths of line-segments, visually perceived. Curve, theoretical, based on equations (25)- (28); points, experimental (Chodin, 1877). Abscissa, visual angle of shorter segment; ordinate, just-discriminable angular difference. 72 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM Our discussion has dealt mainly with two mechanisms, and it is important to distinguish the experimental situations to which they were applied. The first mechanism was applied to relative discrimina- tion between stimuli simultaneously presented, and the distribution of "correct" and "wrong" judgments was predicted. The second was ap- plied to absolute discrimination, each stimulus being presented alone, and a Weber ratio was deduced. A third mechanism was first intro- duced in chapter vi, and could be applied to sensitivity following adap- tation. Still a fourth was suggested in chapter iii, but only as illus- trating the possibility of a mechanism failing to transmit under in- tense stimulation. No application to concrete quantitative evidence was made of this. The last mechanism discussed here will be extended in chapter xii to provide a mechanism for the discrimination of colors. TABLE I LIFTED WEIGHTS EXPERIMENTAL THEORETICAL s r, Pe Pi PB Pe Pi Grams P r o p o r t i o n s 84 0.012 0.027 0.961 0.004 0.021 0.975 .020 .980 .010 .990 88 .021 .082 .897 .025 .077 .898 ( .053) .947 .053 .947 92 .096 (.181) .723 .103 (.181) .716 .185 .815 .179 .821 96 .275 .266 .459 .284 .265 .451 .420 .580 .409 .591 LOO .502 .267 .231 .551 .250 .199 ( .683) .317 ( .683) .317 L04 .842 .103 .055 .796 .140 .064 .920 .080 .880 .120 L08 .915 0.065 .020 .932 0.054 .014 0.963 0.037 0.966 0.034 PSYCHOPHYSICAL DISCRIMINATION 73 WEIGHT DIFFERENCE AW IN GRAMS Figure 10. — Comparison of theory with experiment: discrimination of lifted weights for three observers Qi,%, and 03 , left hand L, right hand R, and aver- age for both hands M. Curve, theoretical, based on equations (25) -(28) ; points experimental (Holway, Smith, and Zigler, 1937). Abscissa, lesser weight in grams, ordinate, just-discriminable difference in grams. X INTERCONNECTED CHAINS: MULTIDIMENSIONAL PSYCHOPHYSICAL ANALYSIS In some cases, for example in the making of aesthetic judgments, the stimulus-objects are complex and may provide stimulation in any number of distinct modes. Then if a statement of preference is called for — one of two incompatible responses — the sequence of stimulus and response may be regarded as a discriminal sequence as defined in the preceding chapter provided we regard each of the two stimuli as a resultant of the components in the various modes. The composition, we may suppose, is effected in some way within the organism through the concurrence at some point of the afferent chains leading from the several receptors. The simplest scheme for representing the neural processes which mediate a discriminal sequence of this type is the following. Suppose that each complex stimulus-object, CP(p = 1 , 2), provides stimuli of intensities CP;(i = 1 , ■■• , n) in the n modalities, and that these stim- uli send impulses independently along discrete afferent chains to the synapse Si where they occasion the production of o- = SPi , the SPi for each p combining additively to yield the SP hitherto employed: sP = zs Pi Each Spi is then some function of Cpi alone; still regarding only the near-threshold range we may take these functions to be all linear, SP = ^LiCpi-M. (1) We may now use either of the procedures introduced in the previous chapter, with a = /5 , according to the choice of location for the ran- dom element. In either case there are but three, or, with more special assumptions, only two, functions P . The functions remain, however, functions of the SP ; for any pair of stimuli d and C2 , which provide some (unknown, since the L, are unknown) gCL2> relabeling the objects if necessary. For determining the four quan- tities A(i)j, only three equations are available. Hence we may make the assumption r±l -f±2 ) subject to possible later revision. Finally, we may suppose that Ax{2) < A2(2), since we can only separate but not identify the two modes, and we have Ax<2> = [£<°2> - £<12>]/2 , A2<2> = [S<02) + S<12>]/2 , A^ = A^ = S(01> /2 . Now consider any object A(3). If either £(03) =£(01) _|_ £(13) —£(02) _|_ £(23) or else £(03) = £(01) _ £(13) — £(02) _ £(23) then £(03)— ^(3) + 42<3) and the quantities on the right are indeterminate. If neither, there are at least two independent equations involving Aa(3) and A2(3). If there are three, a two-dimensional representation is impossible, but if for every fourth object A(3) there are at most two new equations, two dimensions are sufficient. Thus, apart from the arbitrariness indicated, with a sufficiently large number of stimulus-objects A all the Ai can be determined. It is perhaps clear enough from the above how one must proceed when more than two dimensions are required. The quantities S may be regarded as distances in the representative space, and the space is metric but not Euclidean. The assumption of linearity imposed upon the mechanism is not highly restrictive, in principle, since, if the 78 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM stimuli are all low in intensity, the linear expressions may be re- garded as the first terms of Taylor expansions. Additional cross- connections could be introduced into the mechanism, either inhibitory or excitatory in character, but the essential features would not be changed nor the metric of the space greatly modified. In particular, it is important to note, the arbitrariness is inherent in the nature of the experiment and can only be removed by further experiment or observation of a different kind. In chapters vi and vii we considered the process at a single syn- apse terminating one or more afferent chains, and related this to various sequences of stimulus and response. In all these sequences there was but a single response, however complex in form, dependent upon the variation of the single a , evoked by a single, simple or com- plex stimulus, and perhaps modified, in degree or in the time of its occurrence, by other stimuli. In this and the preceding chapters we have considered two or more synapses with as many afferents, ter- minating afferent chains from various receptors. The structures were associated with classes of sequences of stimulus and response of the following sort. Each of a group of stimuli, simple or complex, when presented in isolation can evoke a certain characteristic response, whereas the concurrence of these stimuli modifies the separate re- sponses by enhancing, reducing, or even preventing them. We have by no means exhausted the possible applications of the various struc- tures ; by varying the assumed relations among the parameters an al- most countless variety of sequences is suggested. For example we have been considering only the case of crossing inhibition and have mentioned only in passing the possibility of having crossing excita- tion, a mechanism that would mediate sequences of a quite different sort. Numerous possible complications are easily suggested. The re- sponse to a given stimulus might be modified, not directly by another external stimulus but by the response to that stimulus, this calling for a connection running from the second effector back to some point in the afferent chain leading to the first effector. Circuits of the type discussed in chapter iv might be introduced at various points and their effects studied and related to observable sequences. The procedure of starting with the simpler structures and seek- ing applications thereof has this decided advantage, that we can feel assured that the postulated mechanism is not more complicated than necessary for mediating the adduced sequence. Thus if one stimulus can in any way modify the response evoked by the isolated occurrence of another, then some connection must lead from the first receptor to the effector for that response, whether the connection is direct, through the spinal cord only, or indirect, through the thalamus, or MULTIDIMENSIONAL PSYCHOPHYSICAL ANALYSIS 79 elsewhere. It is highly unlikely that the actual mechanism mediating any of the stimulus-response sequences here outlined is as simple as the one postulated for it, but by comparing the deductions from the simpler postulates with laboratory data we can take note of the devia- tions and be guided thereby in our endeavor to improve the picture. While this procedure, from mechanism to suggested application, could be pursued indefinitely through increasing degrees of complex- ity, we turn instead, in the following chapters, to the reverse proce- dure, considering certain forms of activity and attempting to con- struct mechanisms capable of mediating these. XI CONDITIONING A most important property of neural circuits is that their ac- tivity may continue indefinitely after cessation of the stimulus. The possible application of this property to memory is evident, but to conditioning it is much less so. We now suggest a mechanism for ex- plaining conditioning and learning. Consider first a few properties of the structure of Figure 1 i x Figure 1 (Rashevsky, 1938) . For the present we shall ignore the presence of the dotted neurons. This structure consists of two neuron-chains, leading through a final common path to a response R , together with a uni- lateral interconnection and a simple circuit C . The chief character- istic of conditioning is that a particular response R , normally pro- duced by the "unconditioned" stimulus Su but not by the stimulus Sc , may after the repeated concurrence of the stimuli Sc and Su become capable of being evoked by Sc alone. This may require one or more concurrences of Sc and Su , and while Sc and Su need not be presented at exactly the same time, the time between them cannot be too long. Suppose, only for the sake of simplicity, that all the neurons are of the simple excitatory type, and let <£o represent the maximum value of for any Sc . Then for the net of Figure 1, let 80 CONDITIONING 81 <£<, < h' < 0 + e0 , (1) ,(, < h" < .0 + .0tt > (2) o« < ^ , (3) where £0 is the value of the excitation at sc due to the circuit when in steady-state activity, and where the 's refer to the afferent neurons. If the unconditioned stimulus Su exceeds the threshold of /„ suf- ficiently, the response R may be elicited. But, as we assume that the circuit C is not active initially, the stimulus Sc cannot produce the response R because 4>0 < h' . Furthermore, neither Su nor Sc can bring C into activity when there is too long a time between their oc- currence. Thus Su alone can produce R but Sc cannot. Now suppose that Su and Sc are applied together for a sufficiently long time. Though simultaneous presentation is not a necessary condition, it will be con- sidered here to simplify matters. Because of condition (2) , the thresh- old of the circuit will be exceeded and the circuit will pass over into a state of steady activity. If, now, a large enough Sc is applied alone for a long enough time, the threshold h' will be exceeded because of condition (1). Thus the response R may now be produced by the hitherto inadequate stimulus Sc alone, and the structure exhibits one of the principal features of conditioning. If we add now the inhibitory neurons III' and III, the resulting structure will exhibit another feature important in the phenomenon of conditioning. Whenever Su is applied, the effect of neuron III is blocked by III'. But if Su is not applied, then the continuous or re- peated application of Sc may cause III to produce enough inhibition at s' to block the action of Sc if conditioning has previously taken place. This corresponds to the loss in effectiveness of the conditioned stimulus which occurs when it is applied repeatedly without rein- forcement by the unconditioned stimulus. If, instead of a single circuit C , we assume that there are a num- ber of them having different thresholds, we should be able to show that a more intense Su and Sc would tend to produce a more intense response. Furthermore, by considering repeated applications of Su and S , it is possible to determine the effect of the number of repeti- tions on the conditioning. By combining these extensions of the struc- ture, N. Rashevsky (1938, chap, xxv, equation 44) obtains an expres- sion eB = A(il-e°») (4) for eR , the excitation tending to produce the response R when S is ap- plied as a function of the number, n , of repetitions. The constant A in- creases with the intensity of the conditioned stimulus, while the con- 82 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM stant a increases with both Su and S and depends on the time between repetitions and the time of stimulation at each repetition. This, too, is in qualitative agreement with results of experiments on conditioning. This conditioning mechanism requires the continued activity of some circuits and the objection could be raised that the great sta- bility of well-established memory-patterns, which resist disruption by either such major disturbances as shock and narcosis or by the cumulative effect of countless minor vicissitudes over a period of time, is inconceivable in terms of so vulnerable a structure. To this ob- jection at least two replies are possible. One is that a quantitative theory is useful in proportion to the extent of the phenomena for which it can account, and it is not less useful for failing to account for others, However, the objection can be met in a more positive way by supposing that the relatively rapid changes permitted by the above mechanism lead in some fashion to more permanent structural changes. It is quite possible that the usual intermittent rise and fall of e and j at a given synapse would have no physiological effect be- yond the exciting and inhibiting effects which we have postulated, whereas the maintenance at some sufficiently high value of either or both would cause permanent changes involving, among other things, a modification of threshold (cf. Douglas, 1932). The theory of the process of conditioning would then hold as outlined, but would re- quire a supplement to account in detail for the observed stability. But regardless of the mechanism, the facts of conditioning re- quire some change in e with successive trials, leading to a change in response. The simplest assumption possible is that s is proportional to the number of trials, at least when the number of trials is small, and this is the essential content of equation (4), with Aa the constant Figure 2 CONDITIONING 83 of proportionality. The theory could be developed formally regard- ing- Aa as a purely empirical constant with no definite physiological significance. But the model enables us to relate this constant to such variables as the strengths of the conditioned and the unconditioned stimuli, temporal factors, and the like, and so provides the possibility of relating a larger group of variables in a single formulation. For interpreting some experimental results in these terms, we consider the net shown in Figure 2 (Landahl, 1941). Let a stimulus Sc normally produce a response Rc , and let a pleasant response R1 always follow Rc in the experimental situation. Let Sw normally pro- duce Rw , which, in the experimental situation, leads to an unpleas- ant stimulus, less pleasant stimulus or to an equally pleasant stimulus but after a longer time. Let the circuits M and C each represent a large group of circuits of different thresholds. Let the part of the structure composed of neurons III, III', IV, IV be equivalent to the corresponding part of Figure 1 of chapter ix. We shall consider only simultaneous presentation of SK and Sc . On the first trial, neither C nor C can become active, and thus we have acting at sc a quantity e,oc; and similarly at slc a quantity e,ow . Then, if one of the two re- sponses must be made, the probability Pc of the response Sc may be given by the approximate equation (8) of chapter ix, with £! — e2 replaced by £oc — £ow • After Sc and Sw have been presented together n times, the re- sponse Rc will have been made, say, c times and the response Rw , w times. We shall refer to n as the number of trials, c as the number of correct responses, and w as the number of wrong responses. Then Pc , the probabilty of a correct response, may be identified with the proportion of correct responses, so that approximately Pc = dc/dn; (5) and similarly Pw = dw/dn , (6) Thus Pc + Pw = l, c + w = n. (7) When a stimulus Sc is presented, a certain group of the circuits M are brought into activity. Then, each time there is a response Rt , conditioning may take place in some circuits of C and the amount, as measured by the increase in the excitatory factor Aec at sc , will not be dependent upon the time tc between the presentation of Sc and response Ri . But, if the circuits M are acted upon by inhibitory neu- rons from various external sources, or if the circuits are replaced by single neurons, the activity will decay with the time, tc , roughly ex- ponentially. Thus, we may obtain an expression for Aec similar to 84 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM equation (4). To a first approximation, equation (4) becomes in this case Aec = ncb (Sc , tc) where b = Aa depends on Sc and tc , and where c is the number of repetitions of Sc and Ri together and thus replaces n . If Sc and tc are constant, the total e at sc is given by ec = s0c + be . (8) We can obtain a similar expression for ew at sw . This is the case when the final response is pleasant, but if the final response is unpleasant we should expect the effect of the conditioning to be the opposite. That is, the centers C are such that they have inhibitory fibers lead- ing to sw . Then the coefficient corresponding to b will be some nega- tive quantity — /S . Thus ew = eow — pw, (9) as w is the number of repetitions of wrong response leading to R2 . Let us apply these results to the particular experimental situation which arises when Lashley's jumping-apparatus is used. Here an ani- mal is forced to jump toward either of two stimuli. Choice of one leads to reward, choice of the other may lead to punishment. For simplicity, we assume that the times tc and tw , respectively, from presentation of the stimuli to the reward and punishment, are con- stants. If we then introduce equations (8) and (9) into equation (8) of chapter ix, and eliminate Pw and c by means of equations (6) and (7), we obtain a differential equation in w and n . From this, with the initial condition w = 0 for n = 0 , we obtain 1 2bekle°c~e°w) w — fog QO) k(b - ft) & 2&e*(e"-£°«> - (b - /S) (1 - e-kbn) for ft^jS. For 6 = >5 , the result is a rising curve which approaches a limit exponentially. In terms of the mechanism, we may consider the experiment as requiring a discrimination between two stimuli whose values, in effect, change in successive trials. The correct stimulus be- comes effectively larger due to the conditioning while the wrong de- creases. Thus, the probability of a wrong response diminishes. In Figure 3 is shown a comparison between the theory and the experimental data by H. Gulliksen (1934). The lower and upper curves were obtained respectively by setting eoc — eow — 0 , kb = .0121 , ft = 0 and k(eoc — cow) = -.46 , kb = .0229, § = 0 . Besides giving a quantitative relation between w and n, our considerations actually give a great deal more. From the results of the preceding paragraphs, we can obtain a function b (Sc , Ri , tc), that is, b is a function of the intensity of the stimulus, the strength of the reward, and the time tc . CONDITIONING 85 DATA BY HGULLIKSEN ■ 00 200 300 NUMBER OF TRIALS fl — Figure 3. — Comparison of theory with experiment: simple learning. Curves, theoretical from equation (10); points experimental (Gulliksen, 1934). Abscissa, number of trials; ordinate, number of errors. Similarly, one can obtain fl(Sw , R? , tw). Thus we should be able to make predictions for data as in Figure 3, but for various strengths of reward or punishment and for other variables. In this way a con- siderable amount of data could be brought into a single formulation and the prediction tested by experiment. By considering various modifications of the experimental situa- tion consistent with the restrictions imposed, we can derive other re- lations which could be checked by experiment (Landahl, 1941). If the responses R% and R2 are made identical but tc ¥^ tw , we have an analogy to a situation in which there are two paths to a goal-response requiring different times, tc and tw , to traverse. We would refer to the shorter path as correct, and thus tc < tw . For constant Sc and Sw , the coefficient b will be a function of tc or tu, only. If Sc and Sw are not too different, ec will equal eoc + b(tc)c and sw will equal Sow + b (tw)w since the final response is the same. But, as b decreases with t, b(tc) > b(tw). Thus, at least when Sc = Sw initially, the probability of the wrong response will decrease towards zero since, for small c and w , c = w , sc — e„ = 6 (tc)c — b(tw)w > zero. Thus, we could determine the number of errors as a function of the number of trials for various t c and tw . If eoc < s„w , the correct response may never be learned. Elimination of a blind alley can be accounted for since, the cor- rect path being entered last, the time between Sc and the reward is less than the time between Sw and the reward. Hence on later trials there is a tendency to turn away from the wrong stimulus. An equa- 86 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM tion for the number of errors as a function of the number of trials has been obtained on this basis and studied in relation to such para- meters as strengths of reward and punishment, length of the alley, and distance (time) from blind to goal (Landahl, 1941). One finds that generally fewer errors will be required to eliminate a blind alley if it is close to the goal. The dependence on the length of the alley of the number of errors required to eliminate the blind is found to be fairly complex. According to the strength of the reward, we find that a blind far from the goal will be eliminated with more difficulty the longer it is, while if it is near the goal it will be eliminated more readily if it is long. What we wish particularly to emphasize is that from relatively simple structure can be deduced fairly complex ac- tivity. It is possible to generalize the mechanism to include a choice from among any number N of stimuli by constructing a net similar to that of Figure 2, but with N afferents and N(N — 1) crossing inhibitory neurons (cf. chap. iii). Suppose that out of the N stimuli there is but one correct stimulus Sc . Hence, instead of considering the indi- vidual wrong responses, we may consider their average effect. Thus if sc is the net value at sc due to the correct response and if sw is the average value of all the e«/s we may write [Landahl, 1941, equation (9)] NP \og—-^- + k(ec-8w)=0 (11) N — 1 'C in place of equation (8) of chapter ix. We note that for ec Pw = (N — 1)/N as would be expected by chance, while for large ec — sw , Pw tends toward zero. In the experimental situation, let a stimulus S'» (i s= 1 , 2 , • • • , M) accompany a group of stimuli Sj (j '= 1 , 2 , • • • , N) of equal intensity, one and only one of which will elicit its response. Among the stimuli S) is a stimulus Sic , the "correct" stimulus corresponding to S'i , which when chosen results in a reward ; response to any other stimu- lus Sj when accompanied by S\ results in punishment, or at least no re- ward. The number N may be referred to as the number of possible choices, while the number M is the number of associations to be learned in the experiment. After a wrong response is made, the ex- perimenter may choose to assist (prompt) the subject in making the correct response or he may not. He may do so each time, not at all or, in general, some fraction, 1 — / , of the times. Thus / is a variable under the control of the experimenter just as are M and N . We shall assume that throughout any particular experiment M , N , and / are CONDITIONING 87 not changed. Then £c = £oc + bc + b(l-f)w, (12) since conditioning improves with each correct response as well as with a fraction (1 — /) of the wrong responses. The prompted correct responses are not counted in c so that we do not change the relation n = c + w . At each wrong choice, a quantity p is subtracted from £ow . This contributes only p/ (N — 1) to the average. Thus ** = 8a»-fi/{N-l). (13) The parameter b gives a measure of the amount of conditioning per trial. If a response to one stimulus has no effect on the condition- ing at centers corresponding to other stimuli, the n is independent of M . But, if the response to one stimulus results in the stimulation of inhibitory neurons terminating at the various other conditioning cen- ters, then b will be less when there are more items M to be learned. We may account for this by introducing as a rough approximation, the relation k where r\ and 'Q are two parameters replacing bk. Assuming eow = eoc , substituting equations (5), (12) and (13) into (11), and eliminating c by equation (7), we obtain a differential equation in zv and n . For the initial condition w = 0 for n = 0, and with b eliminated by relation (14), the solution of the differential equation is (N-l)et» N w = log . (15) Nf-f-0 (Nf- f-0) e-vm + N - Nf + f This equation gives the number w of errors as a function of the num- ber n of trials for any number M of items, for any number of N pos- sible choices, and for any fraction (1 — /) of prompting by the ex- perimenter. All this involves only two parameters C and r\ . As we have considered a highly over-simplified mechanism and introduced a number of approximations, it is not to be expected that the predic- tions of equation (15) should hold over too wide a range of values of M and N . In Figure 4 are data obtained from a single experiment for the purpose of illustrating a rather special case of the experimental pro- cedure outlined above. The experiment corresponds to the case in 88 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM n-»- Figure 4. — Comparison of theory with experiment: simple learning. Curve, theoretical from equation (15); points, experimental (Landahl, 1941). Abscissa, number of trials ; ordinate, number of errors. which M = N and / = 1 . Then setting y\ = 1.15 and C = .098 we ob- tain the three curves for which N = 4, N = 8, and N = 12 respectively. The values of v\ and C may be determined from one point of each of any two curves. The third curve is then without unknown parameters. But another family of such curves is also determined by equation (15), without any additional parameters, for the case in which prompting- follows each wrong' response, i.e., f = 0 . In fact, / can be given any value in the range 0 = / = 1 , and M and N need not be equal. From the previous discussion, we should expect b to depend upon the strength of reward. To a first approximation, we may set 6 = alP/k , where p is a measure of the strength of reward. Similar- ly, we can set ft ==■ a2/pk where p is a measure of the strength of pun- ishment. Equation (15) then determines a hyper-surface in the seven variables, w , n , N , M , f , p , and p in terms of three parameters a-! , a2 , and C • CONDITIONING 89 Furthermore, we may wish, to determine what the performance in terms of the probabiliy of a correct response would be if, after n' trials, the experiment is discontinued for some interval of time. The addition of a single parameter enables one to determine the perform- ance in terms of the various experimentally controlled variables. One can then also determine the time required for the performance to drop to some preassigned level. It is found, for example, that while in- creasing the number of trials beyond some value has little or no effect on the performance at the time, it may have a considerable beneficial effect on the performance at some subsequent time. This is essen- tially what occurs in overlearning. But these results apply only to the case of recognition-learning as we have assumed that the stimulus to which response is to be made is present at the time of choice. A generalization of the results can be made so as to include the case of recall-learning (Landahl, 1943). A number of parameters must be introduced in this case, but also two new experimental variables enter. One variable determines whether the experiment is that of recognition or recall. The other variable is the number of correct responses, as in this case c cannot be deter- mined from c = n + w due to the "equality" response, which in this case is a lack of response. Thus with a small number of parameters specified experimentally, c and w can be determined for various values of the seven other variables and the result may then be compared with experiment. XII A THEORY OF COLOR-VISION According to the Young-Helmholtz theory, any color can be matched, at least after sufficient desaturation by admixture with white light, by combining in suitable proportions lights of three given colors (cf. Peddie, 1922) . These three colors may be chosen arbitrari- ly except that no one is to be matchable by combining the other two. The fact that three primary colors are sufficient for a match of all others strongly suggests, as Helmholtz brought out, that retinal ele- ments of three distinct types are involved in the perception of color. This interpretation has not gained universal acceptance by investi- gators, two of the objections being that anatomical studies fail to differentiate three types of receptor, and that the degree of acuity with monochromatic illumination is too high to be accounted for by only one-third the total density of elements. Hence theories have been proposed to yield quantitative predictions of the discriminal pre- cision in judging color-differences in particular without postulating three distinct types of receptor (e.g. Shaxby, 1943). However, the case here is analogous to the case of discrimination of intensity-dif- ferences in general: different intensities, and also different colors, can occasion qualitatively different responses so that at some place in the neural pathway from receptor to effector the locus of the a must be capable of varying with variation of the stimulus. This state- ment is practically self-evident since obviously different final path- ways are involved in reaching the different effectors, so that the only question is where the variation occurs — centrally, along the afferent pathway, or along the efferent pathway. Wherever this may take place, the method demands explanation. But the statement is also a consequence of the well known Muller's law of specificity, as is clearly brought out by Carlson and Johnson (1941). Hence the neural mechanism which mediates any discriminal pro- cess must provide for a segregation of the neural pathways affected by different stimuli. In the case of a simple stimulus characterized by a single parameter, S , there needs to be only a one-dimensional array of synapses reached by neurons from the receptor in such a way that when S has one value the resultant a is positive at one set of the synapses, and when the value of & is changed sufficiently the resultant a is positive at a different set of the synapses. A mechanism capable of bringing this about was described in chapter ix. When the stimulus is complex and requires three parameters, Sv(i = 1 , 2 , 3), 90 A THEORY OF COLOR- VISION 91 to specify it, then a three-dimensional array of synapses is required. In speaking- of one-dimensional and three-dimensional arrays we are not referring1 to their actual spatial distribution in the cortex, since manifestly all synapses are distributed in three spatial dimen- sions. We refer only to an abstract mode of classifying all the syn- apses in question. In the case of the discriminating mechanism of chapter ix, each synapse of the discriminating center is character- ized by a certain h in such a way that given any interval bounded by the numbers h' and h" , it is possible to say unambiguously of any given synapse whether its associated h does or does not lie within this interval. In the mechanism now to be discussed, of which we say that the synapses form a three-dimensional array, each synapse is char- acterized by the set of three parameters Sx , S2 , and S3 and given any set of these intervals Si to Si", we can say unambiguously of any synapse that each Si associated with it does or does not lie upon the interval from Sh' to Si". We shall now describe a mechanism which generalizes that of chapter ix and provides the segregation of pathways required for the discrimination of colors. While it may not be the simplest one possible, it does possess the necessary qualitative properties and no other mechanism has been proposed which does. We follow Helm- holtz and assume three types of retinal receptors, each connected to all the synapses of the three-dimensional array constituting the "color-center." We utilize the three-receptor hypothesis because it is convenient, not because we are necessarily convinced that it is "true." We consider a small region of the retina only, containing one of each of the three primary receptors, and we disregard the problem of spatial localization or other attributes of the sensations which accom- pany the stimulation of these receptors, confining ourselves exclu- sively to the sensation of color. Admittedly the other attributes de- mand explanation, but we regard the problems as distinct. The spatial arrangement required of the synapses at the color- center is highly arbitrary, and while we imagine a specific spatial localization of the various synapses this is for convenience of de- scription only. With this understood, we suppose: ( 1 ), each primary receptor is associated with a particular one-dimensional array, or axis, of synapses, the three axes being mutually orthogonal and all concurrent at a point O; (2), the stimulation of the i-th primary re- ceptor in the amount Si occasions the production of a throughout the color- center, the density being greatest all along its associated axis and being everywhere a function a, (Si , P) of Si and of the assumed position P; (3), as a function of position 2 = 0 . A more exact specification of the form of the functions en can be made only by a detailed comparison with experimental facts. XIII SOME ASPECTS OF STEREOPSIS In this chapter we present some purely formal considerations of the visual perception of space without providing any specific neural mechanism. Such a formal analysis is, of course, a necessary pre- liminary since evidently when we do not know in advance the struc- ture of the neural mechanism involved we must know what is re- quired of it if we are to deduce the structure. The structure of subjective space is developed gradually during the life of the individual and is the resultant of diverse sensory cues — visual, auditory, kinaesthetic, and perhaps others. The recognition of two pin-pricks simultaneously applied to different parts of the body, as distinct, involves discrimination of a certain primitive type and requires a certain minimal separation of the points; to recog- nize that one pin-prick is located at such a distance and in such a di- rection from the other involves a judgment much more advanced in form and requires a neural mechanism of much greater complication. To account for the first judgment no assumption is required beyond the distinctness of the neural pathways, and of the cortical centers ultimately affected by the two pricks. But the second judgment, while keeping the pricks distinct, assimilates them into a certain continuum and hence relates them in a definite way. Consequently the cortical centers must be connected with each other and with the motor cen- ters in some definite way, perhaps in such a way as to make possible the continuous movement of, say, an index finger from the location of one prick to that of the other. Similar remarks may be made of vision. Let us confine ourselves for the present to monocular vision. The judgment that two objects, seen simultaneously, are distinct, and the judgment as to their rela- tive positions, are quite different judgments and the first does not by any means imply the second. The second judgment may be somehow associated with the ocular rotations that would be necessary for the fixating of first the one point and then the other (cf. Douglas, 1932), in which case if the visual space has been integrated into a unified space of perception as a whole, there may be associations of some sort with movements of the body or of a member from the one point to the other. In either case, however, that of the pin-pricks or that of the visual objects, we say nothing about whether the motor and kin- aesthetic connections are a part of the native endowment of the or- 94 SOME ASPECTS OF STEREOPSIS 95 ganism and formed, possibly before, possibly after birth, indepen- dently of any experience, or whether they are somehow due to con- ditioning. Whatever may be the nature of a given perception, therefore, the perception of a disjunction implies a neural disjunction of some sort, whereas the localization of the disjoined elements within a field of relations implies the presence of interconnections of some sort be- tween nervous elements involved in the perception of these elements, whether these interconnections were previously functional, or only became functional as a result of the perception itself. And while there may be no unique structure of nervous interconnections capable of yielding the system of relations as perceived — the solution of the in- verse problem of neural nets is not necessarily unique — nevertheless there are limitations, and empirical anatomical data may serve, in time, to complete the characterization. The problem we wish to consider is the following. By whatever means it is acquired, the normal human adult does possess a percep- tual space within which he locates the objects of his perception. With objects some distance away from him, visual cues are the chief, and frequently they are the only, means available to him for the localiza- tion of these objects within this space. In these cases, where the visual cues are the only ones, how is the localization effected? That is, what kinds of cues can be provided by the eyes alone, which, to- gether with past associations (but not the memory of a previous lo- calization), may serve the subject in localizing the object within his perceptual space If the eyes themselves provide several, more or less independent, cues for localizing the same object— and they certainly do— then the final localization must come as a kind of resultant of all of these. Under normal conditions these cues would doubtless act in harmony and reinforce one another, thus providing a fairly ac- curate judgment. But under abnormal conditions due to pathology or instrumentation, the normal harmony would be disrupted, making the perceived relations a bizarre distortion of the true ones. In fact, it is by considering the nature and occurrence of these distortions when the cues are in disharmony that we might hope to get our best information as to their separate modes of operation. In the system of spatial relations among disjoint elements, per- haps the simplest and most primitive, beyond, of course, the mere fact of separation, is the amount of separation. This is easily under- stood in terms of kinaesthesis. Greater separation requires more movement for crossing it, a more intense kinaesthetic sensation, and in these terms our discussion of the discrimination of intensities may find an application here. With reference to visually perceived extent 96 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM a suggestion has been made already in this direction (Householder, 1940; cf. chap ix). Perhaps the most obvious cue for judging distance is the "ap- parent size" of the object when the actual size is known or inferred from previous experience. The "apparent size" is by definition the solid angle subtended, but is not necessarily the size the object ap- pears to have. Thus a distant man appears small, but when he is fairly close the size he appears to have stays fairly constant while his distance, and hence his "apparent size," varies over quite a wide range. The "apparent size" is proportional to the size of the retinal image and provides a distance cue. An object seen through a spyglass appears flattened and a pos- sible explanation can be found from considerations of apparent size. Thus consider a cube with one edge nearly, but not quite, sagittal in direction, and suppose it is viewed through a spyglass magnifying in the ratio M = 1 + u . That is, let the retinal image of the cube as seen through the spyglass be M times that formed by the cube seen at the same distance by the naked eye. If the actual distance of the front face is d , and if the edge of the cube has length s , then the back face has the actual distance d + s , but due to the magnification, front and back faces appear to be only 1/M times as far. They ap- pear, therefore, to have the distances d/M and (d + s)/M . But this leaves for the apparent depth of the cube only the distance s/M . While it is well known that qualitatively the effect is present as described, no quantitative data are at hand, and the theory here sug- gested might fail to meet the more exacting requirements of a quan- titative test. In the first place, it is assumed that, in the absence of other cues, the perceived distance would be exactly 1/M times the actual distance. This might not be the case. The percieved distance might be, say 1/li times the true distance, where /u < M . But if so, we should expect the judged depth to be 1/// times the actual depth and the judged size to be M/li times the actual size. Moreover one could perform an experiment in which a truncated pyramid is pre- sented instead of a cube, the dimensions being such that the subject would be expected to perceive it as a cube. For this, when the dis- tance is judged to be 1/// times the actual distance d and the magnifi- cation is M times, the retinal image is the same as would be produced by a cube seen with the naked eye at a distance of dj ii . Now if a cube whose edges are Ms/tu is placed so that the nearest face is at a distance of d/ n , then the visual angles subtended by an edge of the front face and an edge of the back face are Ms/d- and Ms/ (d + Ms). These must be the visual angles of the faces of the truncated pyra- mid placed at a distance d and magnified M times. Hence, if the SOME ASPECTS OF STEREOPS1S 97 depth of the frustum, as well as each of the edges nearest the obser- ver, is s , the edges of the other face must be s(s + d)/(Ms + d) =s[l -us/d] , if Ms is small by comparison with d . Note that this result is inde- pendent of // . While no mechanism is suggested here for this dependence of perceived distance upon apparent size, we note that a converse mech- anism has been suggested by Landahl (1939a) to account for the constancy of perceived size with the concomitant (mutually inverse) variation of apparent size and actual distance. The factor of accommodation (Stanton, 1942) is certainly not sensitive as a distance-cue, but there is evidence that it does operate (Grant, 1942). Distant objects are seen with the relaxed eye (if em- metropic), whereas to focus clearly on nearby objects requires an effort of accommodation. Convergence, a binocular cue, is more certain (Householder, 1940c). Convergence upon near objects also require an effort, although it is known that the visual axes of relaxed eyes are not parallel, but diverge, so that some effort is required also for bi- nocular vision at a distance. The cues of both accommodation and convergence are muscular, and neither is sensitive enough to provide the fine discrimination known to be possible in binocular stereopsis. In fact, binocular stereopsis can be achieved by means of a stereo- scope when the visual axes are parallel and accommodation is relaxed. Thus other cues must be sought. In normal binocular vision whereas there are two retinal images, one in each eye, of the object fixated, there is but one cortical image — the individual sees but one object. On the other hand if the atten- tion, but not the fixation, is shifted to an object enough nearer or far- ther than the point of fixation, then two images of this single object are seen. Objects somewhere in between the point of fixation and the object seen doubly may be seen singly but, if so, they can be recog- nized as nearer to or farther from the observer, as the case may be, than the point of fixation. Let CL and CR represent the centers of rotation of the left and right eyes, respectively. The separation between nodal point and center of rotation is so slight that for present purposes we may regard CL and CR as being also the nodal points. Let P represent the fixation- point. Then PCL and PCR are the two visual axes; let these cut the retinas at PL and PR . The two retinal images of P are therefore lo- cated at PL and PR , and when so located these images fuse and the point P is seen singly. Let P be any point on the line CLP be- tween CL and P . Then P is also imaged at PL on the left retina, but 98 MATHEMATICAL BIOPHYSICS OF THE CENTRAL NERVOUS SYSTEM on the right retina the image is at some point PR in the temporal direction from PR . If an object is located at P', and if P' is not too far removed from P , there may be fusion even of the images on PL and P'R , though because of the more temporal location of P'R with respect to PR , P' is judged nearer the observer than P . The situation when P' lies beyond P is similar except that P'R is then medial to PR . But if P' is moved much farther or much closer, fusion is no longer possible, and two images result. We shall say that PL and PR are corresponding points on the two retinas, whereas Ph and any other point P'R different from PR are disparate. While holding the fixation at P , any other point Q within the binocular visual field will form an image on a point QL of the left retina, QR on the right, and the images may fuse, but only if Q is not too near or too far away. We suppose that associated with each point QL of the left retina there is a unique point QR of the right retina which corresponds to it, while QL and any other point Q'R are dis- parate. If Q has as its images the point QL and a disparate point Q'R not too far removed from QR fusion may still occur, but fusion is in some sense optimal when the images fall on corresponding points. The extent of the disparity, or the absence of it, then provides a cue for the localization of the point Q in the subject's visual space. It is suggested that the degree of convergence, and possibly also the degree of accommodation serve as cues for the localization of the general region about P with respect to the observer, while the binocular dis- parity gives depth to this region and makes possible the localization with respect to P of the objects in its immediate neighborhood. Thus pictures seen in a stereoscope are seen as vaguely somewhere far off, although the positions of the various details with respect to one an- other are very definite. The simplest neurological picture to correspond to this seems to be the following. Suppose there is a binocular representation in the visual cortex which is three-dimensional in character. Simultaneous stimulation of corresponding points QL and QR on the two retinas leads to maximal excitation a at an associated point of the binocular cortex. Simultaneous stimulation of slightly disparate points QL and Q'R leads to excitation