SYMPOSIUM ON INFORMATION THEORY IN BIOLOGY y r, o SYMPOSIUM ON INFORMATION THEORY IN BIOLOGY Gatlinbiirg, Tennessee, October 29-31 , 1956 Edited by HUBERT P. YOCKEY Oak Ridge National Laboratory With the assistance of ROBERT L. PLATZMAN HENRY QUASTLER Purdue University Brookhaven National Laboratory SYMPOSIUM PUBLICATIONS DIVISION PERGAMON PRESS NEW YORK LONDON • PARIS • LOS ANGELES PERGAMON PRESS INC. 722 East 55th Street, New York 22, N. Y. P.O. Box 47715, Los Angeles, California PERGAMON PRESS LTD. 4 & 5 Fitzroy Square, London W.L PERGAMON PRESS, S.A.R.L. 24 Rue des Ecoles, Paris, V^ First published J 958 Library of Congress Card No. 58-9687 Printed in Northern Ireland at The Universities Press, Belfast CONTENTS PAGE Foreword ix A. M. Weinberg Preface xi PART I. INTRODUCTION A Primer on Information Theory 3 Henry Quastler Some Introductory Ideas Concerning the Application of Information Theory in Biology . 50 Hubert P. Yockey PART II. STORAGE AND TRANSFER OF INFORMATION Editorial Introduction 61 The Cryptographic Approach to the Problem of Protein Synthesis 63 George Gamow and Martynas Ycas The Protein Text ^ 70 Martynas Ycas Discussion 101 Protein Structure and Information Content 103 L. G. Augenstine Discussion 1 23 Specific Mechanisms of Protein Synthesis and Information Transfer in the Developing Chick Embryo 124 H. R. Mahler, H. Walter, A. Bulbenko and D. W. Allmann Discussion 135 The Mechanism of Action of Methyl Xanthines in Mutagenesis 136 Arthur L. Koch Evidence for a Negative Feedback System Controlling Liver Regeneration 148 Andre D. Glinos Fluctuations in Neural Thresholds 153 Lawrence S, Frishkopf and Walter A. Rosenblith PART III. DETERMINATION OF INFORMATION MEASURES Editorial Introduction 169 Chemistry and Biochemistry at Low Temperatures and Discrimination of States and Reactivities 171 Simon Freed Discussion 1 80 75193 vi Contents Information Content of Tracer Data With Respect to Steady-state Systems 181 MoNES Berman and Robert L. Schoenfeld The Domain of Information Theory in Biology 187 Henry Quastler Discussion 196 Some Membrane Phenomena from the Point of View of Information Theory 197 Herman Branson Efficiency of Information Transmission by Biochemical Co-factors 204 Peter D. Klein Discussion 209 Antigenic Specificity 211 Bernard N. Jaroslovv and Henry Quastler Information Content and Biotopology of the Cell in Terms of Cell Organelles 218 Charles F. Ehret Quantification of Performance in a Logical Task With Uncertainty 230 A. Rapoport PART IV. DESTRUCTION OF INFORMATION BY IONIZING RADIATION Editorial Introduction 239 Electron Spin Resonance in the Study of Radiation Damage 241 Walter Gordy A Physical Mechanism for the Inactivation of Proteins by Ionizing Radiation 262 Robert Platzman and James Franck Information and Inactivation of Biological Material 276 Harold J. Morowitz Discussion 28 1 The Absence of Radiation-Induced Disulfide Interchanges 283 Arthur L, Koch A Proposed Mechanism of Protein Inactivation 287 L. G. Augenstine Discussion 29 1 PART V. AGING AND RADIATION DAMAGE Editorial Introduction 293 A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 297 Hubert P. Yockey Entropic Contributions to Mortality and Aging 317 George A. Sacher Contents vii A Quantitative Description of Latent Injury from Ionizing Radiation 331 H. A. Blair Some Notes on Aging 341 Hardin B. Jones Cancer as a Special Case of a General Degenerative Process 347 Harry Auerbach Discussion 351 Free Radicals as a Possible Cause of Mutations and Cancer 353 Walter Gordy PART VI. INFORMATION NETWORKS A Probabilistic Model for Morphogenesis 359 Murray Eden Functional Geometry and the Determination of Pattern in Mosaic Receptors 371 John R. Platt PART VII. THE STATUS OF INFORMATION THEORY IN BIOLOGY The Status of Information Theory in Biology: A Round Table Discussion 399 Edited by Henry Quastler Author Index 403 Subject Index 41 1 FOREWORD Alvin M. Weinberg Director, Oak Ridge National Laboratory The reader of this book may wonder why it is that an institution such as the Oak Ridge National Laboratory, which is primarily interested in the control and release of nuclear energy, should also be interested in sponsoring a meeting on Information Theory in Health Physics and Radiobiology. The answer rests in the fact that among the activities that are pursued at this Laboratory there are two which bear very directly on general problems of growth and of the impairment of growth by radiation and allied agents. Broad programs in fundamental research in the basic physical mechanisms and in the basic biological manifestations of radiation damage have been established in the Health Physics Division and in the Biology Division. In the Biology Division there is a great deal of experimental work being done on protein synthesis, on the mechanism of action of the nucleic acids, and on problems of the characterization of the nucleic acids. In the Health Physics Division there is a lively interest in the problems of dosimetry and the basic mechanisms of the interaction of radiation and matter. It is in establishing a tie-up between the physical and biological aspects of radiation damage that information theory may play an important role. We hope that this conference will help to assess the value of information theory to phenomena involved in the interaction of radiation and living matter. (/) ; in our example: 2p{i) • z, = 1/2 + 3/8 + 3/8 + 3/8 + 4/16 + 5/32 + 5/32 = 70/32 = 2.19 i From p{i) = (1/2)^ we get : logg p(/) = z, • logg ( 1 /2) and, because: logg (1/2) = —1 we have: z^ = —loga /?(/)• We get (for/?(/)'s which are integral powers of 1/2!) the following result: Average number of binary symbols per event = —^p(i) logg /)(/). i We will check this result for the case of equiprobable categories. For r categories, the probabihty of every one will be 1/r; so: -lp(i) log2/'(0 = -r--- log2 - = log2 r i r r This is the expression previously obtained for equiprobable categories. Any Probabilities — What if probabihties are not limited to the values 1/2, 1/4, 1/8, etc. ? In this case, it will — in general — not be possible to make divisions into exactly equiprobable groups. We would suspect that in this case the coding will be less than optimally efficient; accordingly, the average length of a code word will be somewhat higher than —^p(i) logg /?(/). The approxi- mation is usually not bad. This is illustrated in the following example which shows the construction of a binary code for the letters of the English alphabet, taking into account their relative frequencies. As expected, it turns out that A Primer on Information Theory 15 each category, /, is represented by a code word of approximately —\og2pO) digits; accordingly, its contribution to the weighted average is not far from the ideal value of —p{i) log, /?(/), and the mean code length is only very slightly greater than the limiting value of —]£/?(/) logg /?(/)• Table I. Fano Code for English Letters 1 2 3 4 5 6 7 No. of digits in code word Contribution / pU) Code -10g2/'(/) to weighted -/>(/) X logipO) average 2x4 2x5 E .132 HI 3 2.92 .393 .384139 T .105 110 3 3.25 .315 .341411 A .086 101 3 3.54 .258 .304398 0 .080 1001 4 3.64 1 .320 .291508 N .071 1000 4 3.82 .284 .270938 R .068 0111 4 3.88 .272 .263725 I .063 0110 4 3.99 .252 .251275 S .061 0101 4 4.04 .244 .246137 H .053 0100 4 4.24 .212 .224606 D .038 00111 5 4.72 .190 .179278 L .034 00110 5 4.88 .170 .165862 F .029 00101 5 5.11 .145 .148126 C .028 00100 5 5.16 .140 .144436 M .025 0001 1 1 6 5.32 1 .150 .133048 U .020 000110 6 5.64 .120 .112877 G .020 000101 6 5.64 .120 .112877 Y .020 000100 6 5.64 .120 .112877 P .020 000011 6 5.64 .120 .112877 W .015 000010 6 6.06 .090 .090883 B .014 000001 6 6.16 .084 .086218 V .009 0000001 7 6.80 .063 .061162 K .004 00000001 8 7.97 .032 .031863 X .002 0000000011 10 8.97 .020 .017931 J .001 0000000010 10 9.97 .010 .009965 Q .001 0000000001 10 9.97 .010 .009965 z .001 1.000 0000000000 10 9.97 .010 .009965 4.144 4.118347 We have already met a situation where a binary code was less than optimally efficient (in the sense of minimum length of code words); that was the case of r equiprobable categories, when r was not an integral power of 2. In this 16 Henry Quastler instance, it was possible to approximate optimal efficiency by symbolizing groups of events instead of single events. The same principle works in the case of probabilities which are not integral powers of (1/2). We will illustrate the method in the case of a situation with two alternatives. Example: Let there be two categories of events, 'A' and 'B', with associated probabilities, p{K) and /?(B) : /'(A) = .7 /KB) = .3 The limiting value of symbols per event is: -IKO log2 AO = -(0.7 logo 0.7 + 0.3 log2 0.3) = 0.881291 . . . i If this situation is to be represented on the basis of single events, then one needs one binary digit per event. Event Probability Representation A 0.7 1 B 0.3 0 Average number: 1.0 symbol per event; excess 12 per cent. The following two-event clusters are possible: AA, AB, BA, BB. If the two events are independent, then the probability that both occur is the product of their individual probabilities : p(AA) =7;(A) -piA), p(BA) =/7(B) - p{A\ etc. Setting up a Fano code, we get: Event Probability Representation AA .49 AB .21 BA .21 BB .09 1 0 1 0 0 1 0 0 0 Average 1.81, or 0.905 symbols per event; excess 3 per cent. If we can encode groups of three real events, then we get still closer to optimum economy : Event Probability Representation AAA .343 AAB .147 ABA .147 BAA .147 ABB .063 BAB .063 BBA .063 BBB .027 1 1 1 0 0 1 1 0 1 0 0 0 10 0 0 11 0 0 0 0 0 0 0 1 Average: 2.686, or 0.895 digit per event; excess 1^ per cent. A Primer on Information Theory 17 Even with more pronounced unbalance of frequencies, tiie minimum value of binary digits per word is soon approximated. For/)(A) —■ .89 and /;(B) = .11, the limiting value is .50. In single-event-code, one needs one digit per event; for two-event-sequences, .66 digits; for three-event-sequences, .55; and for four-event-sequences, .52. We have begun our discussion of binary representation with the case of 2, 4, 8, 16, ... , equiprobable categories. We then generalized to cases with any number of categories, and proceeded from the representation of single events to clusters of events. Next, we introduced unequal probabilities, of value 1/2, 1/4, 1/8, ... . Finally, we dropped all restrictions. We can now state, with full generality: If a real situation is categorized into r categories, with associated proba- bilities p(i), (where / = 1, 2, . . . , r), then it is possible to represent each r event with an average of no more than — 2 p(i) log2 pii) binary symbols. i = l Representation Theorem — In general, the closer we v/ant to approximate the minimum bulk of representation, the larger the groups of sequences which must be encoded. This entails the following penalties : 1. There will be a delay in waiting for a whole group of events to occur or to be registered, and 2. The encoding and decoding procedures, and the code book itself, will become the more elaborate the larger the groups coded. It is obvious that the code which is most economical in terms of bulk of representation is not necessarily optimum in over-all performance. There will be cases where it might be worthwhile to sacrifice economy in word length for ease in decoding. If the reader will work through exercise 4, then he surely will appreciate this possibility. Whether or not minimum bulk of coding is favorable, in a given case, cannot be derived from informational analysis. What information theory does is to establish a limiting value of the number of symbols, of a given kind, which are needed to represent the information in a given factual situation; in some cases, like those here discussed, information theory will also show how such coding economy can be achieved; but it can never prescribe that this is what should be done. It would be quite legitimate to inquire, at this point, why we have gone to so much trouble to find out how to achieve binary representation with minimum bulk ? Is not the result of doubtful value, in view of the fact that a tolerable approximation to minimum bulk can usually be achieved with the simplest means, and that a close approximation often entails prohibitive costs in encoding and decoding? The answer is this: by establishing the minimum length of code words in standard binary representation, we have implicitly established a general condition of representability : If an event can be represented by (on the average) n binary digits, then it can symbolically represent, or be represented by, any other event that can also be coded into n binary digits. This can be immediately generalized to groups of events: Let 5"^. and Sy be the number of real and symbolic events in a group, and n^. and «„ the average 18 Henry Quastler binary representation per event and per symbol. Then, the general condition of representability can be stated as follows : ^y ' f^u ^ ^X ' ^X EXERCISES 1. A weakness of the Paul Revere code is that there is no positive signal for "peace and quiet". Hence, the colonists could not be sure whether the absence of a warning signal meant "peace and quiet" or a disturbance in the communication system. Show how two lights could be used to indicate the four situations by positive signals. 2. Any integer can be written as a sum of powers of 2(1,2,4,8,16, • • •)• For instance: 27 = 16 + 8 + 2 + 1 = 2* + 2=' + 21 + 2» In binary notation, one indicates the power by position, and writes a '1' in appropriate position if this power does enter the sum, a '0' if it does not. Thus, '27' becomes 11011. (a) Write the following numbers in binary notation: 0,1,2,3,4,5,6,7,8,9,10,12,16,1955. (b) Write the following binary numbers in decimal notation: 1001, 1011, 10010011, 100000. Any proper fraction can be written as a sum of powers of 1/2, (1/2, (1/2)^ = 1/4, (1/2)^ = 1/8, etc.). For instance: .75 = 1/2 + 1/4, or, in binary notation, .11. (c) Translate into decimal notation: .001, .1001001 3. (a) encode the message 'ABCDE' in code (a) of the five-word codes described earlier, (b) decode the message: '000001011011' in code (b). 4. This assignment is coded in the Fano code for English letters given earlier. 001001001 10000101 1 1001 1 10001 10001001 10101001001001 1000001010001 1001010 1 101001 10000010101 1111111 1001001001001 1 1 1 1 10001 10010101 101000000101001 0101 100000001 1 1 100000101 10100010101 1 1000100001 1 101 1000010101 101 1001010 0101 100101 1 1 1 1 1 101001000100001 101111 101 101 110111 101 1000001 1 10010010010 001 1 100001 110101111111 1001001 1 100001 1111011 100101 100101 1 10001 1 1 101 1000 001001 1 1 100100101 1 10010001 100101001001001001 1 1 1 1 100001001 101 1001001 100 100101 1 10100100101 1 1001001 1 1 10 100000 1 10010000001 1 1 10000010001001 1 1 1000 001001001001 1 101 101000000101 101 1000001 1 1001 1 1 1 1 1001001001001 1 101 101000 000 101 1010001 1 1 1 1 101010101 101000101 1 1 1001 1001 1000000001 1 1 1 1 10010001 10 010110011000 111 (This assignment is very tedious but it is good practice.) 5. Given a real situation with three categories and probabilities p(A) = .8, p{B) = .15, p(C) = .05. Construct a binary code which comes within 10 per cent of the minimum bulk. 6. A protein is thought to be a linear arrangement of amino acids of which there are (about) twenty kinds in each cell. The specificity of a protein depends mostly on the sequence of amino acids, i.e. a protein can be considered as a 'message' written in a twenty-letter alphabet. It is known that, in the living cell, protein specificity is determined by nucleic acids. These are linear arrangements of nucleotides, of which there are four different kinds. Question: what is the minimum number of nucleotides needed, on average, to specify each amino acid? Assume all amino acids to be equiprobable. III. THE MEASURE OF INFORMATION OR UNCERTAINTY It seems reasonable to equate the amount of information acquired, as a result of an event, to the amount of uncertainty which its occurrence has A Primer on Information Theory 19 abolished*. The prior uncertainty does not depend on the event that has actually happened, but, rather, on the whole set of events which could have happened at this particular occasion. For instance, if one wishes to compute how much information is acquired, on the average, by a glance at the speedo- meter, one proceeds to estimate how uncertain a motorist is before he glances. The amount of this uncertainty must depend on the number of needle positions which the motorist thinks he can distinguish. Suppose his speedometer scale reaches from zero to one hundred and he can read the position to the nearest mile per hour; then, he will be able to distinguish 101 positions, and the amount of his uncertainty will be somehow related to this number. However, it wouldn't be realistic to relate his uncertainty only to this number, 101. Because, suppose his speedometer scale ranges up to 150 instead of 100 miles per hour; yet, when he is driving along the highway at a moderate speed, this extra portion of scale does not contribute in any way to his uncertainty; he will be quite sure that his needle will not be in this interval. In fact, he will expect to find his needle somewhere within a range of about 10 m.p.h., and he will be almost certain to find it within a somewhat larger range of, say, 20 m.p.h. Thus, to describe his uncertainty realistically, we must not only state every possible result of his reading, but will have to qualify each by a statement of expectation or probability. The Amount oj Uncertainty As before, we turn to a binary situation to obtain a simple perspective of the problem. Suppose somebody has made a record of 100 tosses of a coin; he has registered only whether the coin fell 'head up' or 'tail up', but neglected all other features such as on what spot the coin came down, which direction the head faced, etc. What is the average amount of information in the record of any one toss? In other words, what is the amount of uncertainty before the record is seen ? The uncertainty must be a function of 'two', the number of alternatives; it must be modified by their relative frequencies. If it is known that the record is that of a coin so thoroughly biassed that 'head' always turns up, then there will be no uncertainty at all; if the coin is moderately biassed, then the outcome of a toss will be uncertain but not qui.te as much as with an unbiassed coin. If we don't know the bias of a particular coin, then we do not know exactly how uncertain we should feel about the outcome of a toss. If we know that the record contains 60 'heads' and 40 'tails', then a record of 'head' will show up with a probability of .60, a record of 'tail' with a probability .40. The uncertainty can be described by a statement of these probabilities: Probability of head up 0.60 Probability of tail up 0.40 In the same way we can describe any number of binary uncertainties with a 60-40 choice between any class 'A' and its complement 'non-A' — where 'A' and 'non-A' may be males and females, hits and misses, friends and foes. * At some time there was some discussion whether uncertainty and information should be given opposite signs. Present usage prescribes the same sign for both. 20 Henry Quastler These uncertainties differ in any number of respects from each other. They win be of interest in very different situations; the kind of infomiation needed to produce certainty is not the same; neither is the usefulness of this information, and so on. However, there is something in common between all uncertainties which can be characterized by the probabihties: Probabihty of 'A' 60 Probabihty of 'non-A' ... 1 — .60 One aspect of this 'something-in-common' is that an arrangement of any 60 A's and 40 non-A's can be coded to represent any other 60 A's and 40 non-A's —heads or tails, males or females, hits or misses, friends or foes. Once such representation has been established, then the uncertainty concerning one event will be abolished by information concerning the other. We have previously equated the amount of information with the amount of uncertainty it removes. Accordingly, it can be said that the amounts of uncertainty and information must be equal in all situations characterized by a binary alternative with probabilities .60 and .40. The foregoing consideration exposes the fundamental features of the measure of information : (1) Information is a measurable abstract quantity; its value does not depend on what the information is about, just as length, or weight, or tempera- ture have values which do not depend on the nature of the thing which is long, heavy, or hot ; (2) Information is related to the ensemble of possible outcomes of an event; its value depends on the probabihties associated with these outcomes, but not on their causes, and not on their consequences. What remains is the development of a measure which comphes with this concept of 'amount of information'; this is merely a technical problem. An obvious generalization states that whenever two events have the same number of possible outcomes, and identical sets of probabihties are associated with the two ensembles of possible outcomes, then these two events have identical information contents. However, we wish to be able to compare events with quite different probability sets; for instance, we wish to be able to say which uncertainty is greater, that associated with a situation with three equiprobable alternatives, or that where there are four possibilities with probabilities .8, .1, .05 and .05. To answer such questions, we have to derive a measure which is a single number, whatever the number of possible categories and their associated probabihties. Such a measure is readily derived from the equivalence of uncertainty with the information which removes it. We may represent the information content of an uncertainty-removing piece of intelligence in any manner we wish. We stipulate that this information should be represented in a standard fashion, namely, by using a binary alphabet. In addition we stipulate that the binary representation be coded in such a manner that the expected number of symbols is minimized. We thus obtain a unique number; namely, the minimum average number of binary symbols needed to abolish the uncertainty associated with a given situation. This number will be called the amount of uncertainty or information of this situation. A Primer on Information Theory 21 The function here needed has already been derived as the condition of representabiHty. If two situations can be made to represent each other, then information on one can aboHsh uncertainty concerning the other. Thus, mutual representabiHty implies equal information content, and representation in the standard binary system yields a general measure of information content. This measure is the 'amount of selective information' as defined by Shannon and Wiener (4, 5). It is expressed as follows: Let X be a classification with categories i and associated probabilities p{i); then the information content oj x is designated H(x) and given by*: H(x)^ -2 p(i) logo p(i) i The units of this function are the binary digits needed for representation of a given event, and are called bits. It must be remembered that the 'bit' is a technical unit of amount of information and not a small piece of information. A single chunk of information may contain many bits or a fraction of a bit. Some Properties of the Shannon- Wiener Information Function The Shannon-Wiener information function has been derived (admittedly, in a loose fashion) from a consideration of standard representation of information. We will now consider a number of its properties and see that they correspond losely to the behavior which one would intuitively expect from a good measure of information. (1) Independence — Let / be one of the possible categories of an event x, p{i) the associated probability, and F{i) the contribution of the /th category to the uncertainty. It is desirable that F{i) be a function of and only of p{i). The function / ^ F{i)^ -pii)\og^p(i) fulfills this requirement. / (2) Continuity — A small change of /;(/) should result in a small change in F(i); in other words, F(i) should be a continuous function of p(i). The function p{i) log2 p(i) is continuous. /(3) Additivity — It is desirable that the total information derived from two dependent sources should be the sum of the individual information; in other * The information function looks (except for a scale factor) like Boltzmann's entropy- function; this is not a mere coincidence. The physical entropy is the amount of uncertainty associated with a state of a system, provided all states which are physically distinguishable are considered as different, that is, if the categorization is taken with the finest grain possible. In most situations dealt with in information theory, large numbers of states which are physically distinguishable are lumped into equivalent classes. The category "one light on the steeple" is a good example; an enormous number of physically distinct states are compatible with this definition, but they are all lumped into one class. The distinctions upon which categorizations are based are usually a very small percentage of the distinctions one could make. Thus, physical entropy is an upper bound of the information functions which can be associated with a given situation, but it is a very high upper bound, usually very far from the actual value. For this reason, I prefer not to use the word 'entropy' as synonymous with 'information'. A very thorough discussion of the relation between information and entropy has been given by Brillouin (9). 22 Henry Quastler words, the uncertainty concerning independent events should be the sum of the individual uncertainties. Let y be an event with categories j and associated probabilities p{j). Let p{i,j) be the probability of the event pair that x falls into category / and v into category y. Then, the function Hix,}') = -lp{i,j)\og2p(i,j) will measure the uncertainty associated with the event pair. If X and y are independent events, then p{Uj)^p{i)-p{]) As a matter of fact, this relation is often used to define independence. In this case, we have H{x, j) = - 2 p{i.j) logo p{i) ■ pij) i.j = -lp(hj)^og^pii) - lp('J)\oz.2p(j) It is known that J.piUj)=p{i) j IpiUj)=p(j) Substituting these expressions, we obtain Hix, >0 = - 2 Pii) log2 /XO - 2 p(j) loga p(j) i . = H(x) + H(y). ^ H^^^) ^ H^'^^^'i^f^^ Thus, the Shannon-Wiener function fulfills the postulate of additivity. (4) Natural Scale— X\yQ prototype of uncertainty is that associated with a 50-50 choice. So, the unit of uncertainty should be the uncertainty associated with this situation. In this case, both/s have the value 1/2, and Hix) = -(1/2 log2 1/2 + 1/2 log2 1/2) - 1 Thus, the Shannon-Wiener function is seen to have an appropriate scale factor. We have derived the infonnation function from the postulate of eflScient binary representation, and have found that the function so defined has the desirable properties of independence, continuity, additivity, and natural scale. We could have started differently, setting up these four properties 2i^ postulates. It can be shown that these four postulates (or other sets of four similar postu- lates) define uniquely the Shannon-Wiener function. Working it this way, we would have derived the fact that the function so defined has the desirable property of efficient binary representation. The function F{p) is plotted against/; in Fig. 1. The graph shows a curve which originates and terminates at F = 0, and has a flat top with a maximum A Primer on Information Theory 23 of F= 0.53 for/) = 0.37. Inspection of the graph reveals some more important properties of the function F{p) : (5) nO) = 0: When a particular class of events is certain not to occur {p = 0), then it does not contribute to the measure of uncertainty. (6) F(1) = 0: F(p) = - p logp p F(p) 0.3 Fig. 1. Graph of F{p) as a function of/? When a particular class of events is certain to occur {p = 1), i.e. excludes all other classes, then there is no uncertainty about the outcome. (7) Effect of Averaging: F > i[F(p,) + F(p^)] The function of the average is greater than or at least as large as the average of the function. When the probabilities associated with two disjoint categories are averaged, then the uncertainty becomes larger. Figure 2 is a graphical demonstration of this effect. The extreme case of averaging occurs if all r categories in a classification are considered equiprobable. Then, Pi') = 7 I, 1 1 11 max. of H(.x) = — ^ - log., - = — /• • - lo?., - ,-1 /• ^- r r - r max. of H(.x) = log, r 24 Henry Quastler In particular in a binary classification, r = 2 max. of H{x) = 1 Thus, the maximum uncertainty associated with two alternatives is one bit; it occurs if both alternatives are equally probable (this is the case of the unbiassed coin!). (8) Ejfect of Pooling: F(pi + P2X F(Pi) + np2) The function of the sum is smaller than the sum of the functions. That is, pooling of two classes in one equivalence class reduces uncertainty (exactly P|+ Po F(P|1 + F(P2J Fig. 2. Graphical demonstration of the effect of averaging by that uncertainty which is associated with the distinction between the two pooled classes). Extreme pooling results in a single category with probability 1 ; this means uncertainty 0. Figure 3 demonstrates the effect of poohng. The function F(p) = —p logg p has been tabulated. The reader is advised to use Fig. 1 to obtain approximate values for use in working the exercises below. For more precise values, one of the existing tables may be consulted (10, 11). EXERCISES 7. Compute the uncertainty associated with: p(A) = .60 /•(non-A) = .40 8. Compute H(x) for two alternatives, and plot the value against /7(A). 9. Answer the question posed previously: which uncertainty is greater, that associated with a situation (x) with three equiprobable alternatives, or that (y) where there are 4 possibili- ties with probabilities .8, .1, .05 and .05. A Primer on Information Theory 25 10. Estimate the uncertainty of a motorist like the one described at the beginning of this section. 11. Certain languages have considerably fewer letters than English (that is, about 18 to 20), yet the information content per letter is nearly the same. How is this possible? 12. A situation has an unlimited number of alternatives, with probabilities of 1/2, 1/4, 1/8, 1/16, etc. in geometric progression. What is the measure of uncertainty? F(P|)+ Flp^) FlPl + Pg) fj P,+ Pg P2 p,+ P2 2 Fig. 3. Graphical demonstration of the effect of pooling The function of the sum is on the intersection between the curve and the ordinate over the sum; the sum of the functions is on the intersection of the same ordinate with a straight line through the origin and the midpoint of the straight line which connects the intersections of the curve with the ordinates over pi and P2, hence: F(p, + P2) < F(p,) + F(p,) IV. INFORMATION MEASUREMENTS PERTAINING TO TWO RELATED VARIABLES In the two preceding sections we have discussed how to represent information, and how to measure amounts of information. Both procedures become impor- tant if information is to be manipulated. The manipulation most commonly used is communication. 26 Henry Quastler In infonnation theory, we use the word 'communication' in a wider sense than usual — just as the word 'information' is used in a wider sense than usual. We understand by 'communication' any relation between variables, accomplished by any means whatsoever, conscious or otherwise, provided that it results in a mutual reduction of uncertainty. For instance: if one watches one of two tennis players, without looking at the other, he derives a considerable amount of information about the unseen player's action. Thus, the seen player transmits information about the unseen player — although in this case, the transmission of information is incidental and not normally utilized, as one ordinarily looks at both players. An Example of Two Related Variables The following example is purposely selected to represent an instance of unintentional communication. The table below is based on Pearson and Lee's measurements of heights on 1376 father-daughter pairs. To simplify the analysis, we have grouped the data in coarse intervals of 3 in. each, and converted all frequencies into percentages. Table II. Heights of Fathers and Daughters; Probabilities and Information Measures Joint probabilities of heights, pii,)) (Pearson and Lee's data, 1376 father-daughter pairs) jt 59.5 62.5 65.5 68.5 71.5 74.5 pU) -p\og2P 1% = 53.5 .001 — . — — — .001 .01 56.5 .001 .007 .006 .001 — — .015 .09 59.5 .005 .022 .060 .027 .005 — .119 .37 A-t 62.5 .004 .042 .156 .152 .039 .001 .394 .53 65.5 — .009 .075 .175 .095 .010 .364 .53 68.5 — .001 .011 .035 .039 .010 .096 .32 71.5 — — • — .003 .006 .002 .011 .07 Pij) .010 .082 .308 .393 .184 .023 1.000 1.92 -plog^p .07 .30 .52 .53 .45 .13 * height of fathers, in 3 in. intervals t height of daughters, in 3 in. intervals + center of intervals 2.00 Information Functions: H{x) = -S/'(/)log2/j(/) = 1.92 bits i my) = -i:p(j)\og,p(j) = 2.00 bits H(x) + H(y) = 3.92 bits H{x,y) = -i:piij) log, p(i,j) = 3.70 bits ij Tix.y) = H(x) + H(y) - H{x,y) = 0.22 bits A Primer on Information Theory 27 From the marginal sums, the uncertainties concerning the height of daughters, H{x), and of fathers, H(y), are computed as described in the preceding section. The uncertainty concerning both heights in a father-daughter pair is computed in similar fashion from the joint probabilities, p(i,j). This function is properly called the Joint uncertainty, or uncertainty of the two-part system ; its symbol is H(x,y). It is compared to the sum of the two individual uncertainties. If the two heights were completely independent of each other, then the joint uncertainty should be equal to the sum of the individual uncertainties. In our case, it is smaller by 0.22 bits. The deficit is a measure of the internal constraints in the system, which lead to an association between heights of fathers and daughters. The function is designated by the symbol T(x;y). Its defining equation is : j,^^ .^^ _ ^^^^ _^ ^^^^ _ j^^^.^^ This information function is germane to other statistics which measure the relatedness of two variables, such as the coefficients of correlation and of contingency. The T-measure is of very general applicability; the values of the variables do not have to be quantitative, not even ordered — they must only be distinguishable. For instance, one can compute a T-measure for a relation between color and shape. The two functions, H and T, differ in the way in which they are affected by change of scale. Let us consider what would have happened if he had chosen one-inch intervals instead of three-inch intervals. It could be the case that only one one-inch interval out of any group of three is occupied at all. Then, the information that a certain height falls into a given three-inch interval would automatically locate it in some one-inch interval; hence, the uncertainty is not increased by the subdivision of intervals. However, this is an extremely unlikely situation. It is much more likely that the three one-inch intervals are populated with approximately equal frequencies. In this case, additional information of logg 3 = 1.58 bits is needed to specify the proper one-inch interval. Then, the uncertainty concerning the height of fathers with regard to a one-inch scale will be 2.00 + 1.58 = 3.58 bits, and the uncertainty concerning the height of daughters 1.92 + 1.58 = 3.50 bits. The joint uncertainty will be increased by a factor of logg 9 = 3.17, because each cell in the table will be replaced by nine cells as one goes from three-inch intervals to one-inch intervals. If one uses a still finer grain, going from inches to millimetres, then the individual uncertainties can be increased by another 4.7 bits, the joint uncertainty by 9.3 bits. This is quite the expected behavior. The more categories are recognized, the greater the uncertainty of classification. The uncertainty can become infinite for a continuous function. However, it will always remain finite for any set of real observations. T, on the other hand, depends very little on the scale interval used. With very coarse grouping, T tends to be less. In the extreme cases, where all heights are pooled into one single class, all individual and joint uncertainties vanish, and with them their differences. In the other extreme case, where measurements are taken and registered to so many digits that no two results are alike, we must get //(x) = //(;,') = //(x,v) = r(x;v) = loga 1376. But, between these un- reasonable extremes, the measure of constraints is characteristic of the system and not of the scale which is used in measuring it. 28 Henry Quastler Two-part Systems in General We proceed to a general treatment of a two-part system x, y. Let / and 7 be the categories of x and v, respectively, and p{i) and p{j) the associated probabilities. Further, let p{i,j) be the probability of the joint occurrence [{x = i) and (y =;)]. Then: H(x)^ -2 p{i) 10^2 p{i) i H(y) = -Ipij)^og,p(j) H(x, y) = - 2 p(i, j) logs p{i, j) ij We introduce the conditional probabilities, Piij) Prob { V = y if X = /} />,.(O....Prob{x = /ifj=y} When X = i then y must have some value j with certainty (or probability 1.0), that is IPiiJ) = 1 j Equally, Ip^iO = 1 i Furthermore, the probability of the joint occurrence [x ~ i and y = j] can be factored into the product of the probability that x equals /, times the conditional probability that y = j ii x — i; equally, it can be factored into the product of pij) times Pj{i). So : P(i,j)=pii)-Pi{j) ^Pij)-Pj(0 The conditional probabilities yield naturally conditional uncertainties. For instance, the uncertainty of j, if it is known that x = i, will be Hiiy) = -IPiij) loga/^XO 3 The average uncertainty of j, under the condition that x is known, is designated by H/y). It is obtained as the weighted average of the //Xv)'s- i Substituting the value of H^{y), we get tJxiy) = -Ipii) 1 Piij) ^og2Pi(j) I j and remembering that Pii}) - -jay we get A Primer on Information Theory 29 Expanding the logarithm gives HAy) = -IpUj') iog2/X^y) + Ip(ij') Iog2/X0- ij io Noting that lpiJJ)=p(i) 3 we get H/y) = -IpiiJ) loga /?(/,;■) + lp(i) loga /?('■). ij i We have seen that the first term on the right side is H(x, y) and the second -H{x). So: H,(y) = H(x, y) - H{x) and H(x, y) = H(x) + H^y) A parallel development shows that H(x, y) = H(y) + H,{x) This relation is quite obvious if put into words: the joint uncertainty con- cerning two variables is equal to the sum of the uncertainty concerning either one variable plus the conditional uncertainty concerning the second variable if the first one is given. H( K) W/////////////////^^^^^ - > ' f II ^ ^ "• lly (X) *■ 1 "(x;y) — U f V ^ K - H^(y) y//////////////^^^^^^ .^ H (> 1 y) V » \ ^ Fig. 4. The relation between information functions shown graphically The difference in uncertainty concerning )', depending on whether or not x is known, H{y) - Hly\ is the gfl/rt in certainty about y derived from observing x. Substituting for ^^rCj')' weget: H{y) - Ely) = H{y) + H{x) - H{x, y) The expression on the right side is the defining equation for T{x\y): H(y) + H{x) - H{x, y) - T{x; y). It follows from this derivation that Tis a symmetrical function: r(x; jO = rO-; x) = H{x) - H,{x) - H{y) - HJy) and it becomes clear why Tis a measure of the mutual reduction of uncertainty. The relations between the six information functions, H(x), H{v), H(x, v), H^(y), Hy(x) and T(x;y), can be demonstrated graphically as in Fig. 4. 30 Henry Quastler In normal code representation, i.e. reduced to efficient binary operations, the information functions have the following meaning: H(x) . . . .number of operations which specify x Hy{x) . . . .no. of operations which specify x if v is given T{x; v) . . . .no. of operations which apply to the specification of both x and v H(x, y)- ■ ■ .no. of operations which specify the whole system. Inspection of the graph shows that: H(x) > H,ix) H(y) ^ H,(y), that is, the conditional uncertainty cannot be greater than the unconditional uncertainty.* Communication Systems When a system not only transmits information but exists primarily for that purpose, then it is called a communication system. No class of two-oart systems has received as much attention as that of the communication system. In a simple communication system, tlie two parts are called the source and the destination of information. The distinction between source and destination must be based on external grounds; the informational relations between the two are perfectly symmetrical. The relevant states of the source are called the inputs, or signals sent, and the relevant states of the destination are the outputs, or signals received. A single state is called a symbol, and a higher unit composed of several symbols, a message. The conditional probabilities for each pair of signals sent and received form a matrix called the channel. Note that the word 'channel' is again used in a sense wider than customary. A 'channel' may but does not have to be a means of physically conveying information. For instance, if two variables x and y do not affect each other but are both affected by a third variable r, then knowledge of the state of x is likely to reduce the uncertainty concerning the state of y, and vice versa; hence, information is transmitted between the two variables, and they are connected by a 'channel' in the sense of information theory — although they do not communicate with each other directly. * However, this is true only for an average conditional uncertainty, and does not apply to every particular condition. The following example will help to fix the ideas: Consider a diagnostic test for a certain disease; suppose the nature of the test and the occurrence of the disease are such that in 98 per cent of the patients the test is negative ; that of the positive tests, 50 per cent are spurious ; and that virtually every case of the disease will give a positive test. Then, if the test is not performed at all, the diagnostician's uncertainty as to the presence of the disease in any given patient, is -(.99 log2 0.99 + .01 log2 0.01) = .081 bits/patient. If the test was negative then the uncertainty is zero. But, if the test is positive, the chances are equal that it is or is not spurious; hence, the uncertainty is I.O bit, and the diagnostician is more in doubt than he was before. However, the average uncertainty, conditional upon his performing the test, is reduced to .98 X 0 + .02 X 1.0 = 0.020 bits/patient. A Primer on Information Theory 31 The information functions in a communication system are designated as follows: H(x) H{y) Tix;y) .uncertainty of source .uncertainty of destination . ambiguity .equivocation .information transmitted, or communicated Amounts of information transmitted must be referred to some unit of action. In particular, it is customary to compute transmissions per symbol or per unit time. A channel which associates one and only one output with each input, and no output with more than one input, is called a noise-free channel or transducer; in this case, H{x) = H(y) = H(x,y) = T(x;y); HJ,y) = H,{x) = 0. We can think of a noise-free channel as a means by which information at the source is represented at the destination. Physically, this involves two acts of representation: first, states of the channel are selected so as to represent the inputs, according to some agreed-upon code; this is called encoding. Next, the states of the channel are translated into meaningful states at the destination ; this is called decoding. All we have stated about representation, representability and amounts of information could now be restated in terms of encoding and decoding operations. In this sense, the relation which we introduced as the 'condition of representability' is also known as the Theorem of the Noise-free Channel; and all the examples and exercises of representing information could be re-interpreted as coding operations. Noise — Few real channels are noise-free; in general, more than one output can follow a particular input. For instance, the 'channel' which links a daughter's height to her father's is far from noise-free; the following table gives the conditional probabilities: Table III. Data of Table II in Form of a Communication Channel Conditional probabi ities, p (0 / = 53.5 56.5 59.5 62.5 65.5 68.5 71.5 HAx) j = 59.5 .10 .50 .40 _ 1.36 62.5 .01 .09 .27 .51 .11 .01 — 1.80 65.5 — .02 .19 .51 .24 .04 — 1.74 68.5 — — .07 .39 .45 .09 .01 1.70 71.5 — . .03 .21 .52 .21 .03 1.74 74.5 — — — .04 .45 .43 .09 1.55 32 Henry Quastler The last column, Hj{x), is the uncertainty concerning the height of the daughter if the height of the father is known ; it is not too surprising to find this uncertainty smallest in the extreme cases, and always smaller than the unconditional uncertainty of 1.92 bits. The father's height 'communicates' some information about the daughter's height; the amount communicated is 0.22 bits. It is not more than that for a number of reasons. Some of the deficit in information about the daughter's height is undoubtedly due to ignorance, and could be reduced by taking proper account of various concomitant factors. Some of the uncertainty may be irreducible, due to a truly random process — possibly the selection of the particu- lar chromosomes which go into determining the daughter's height. In the strict sense, the term 'noise' is reserved for the effects of random disturbances, and not to the eff"ects of ignorance. However, the problem of the final distinction between uncertainty due to randomness and uncertainty due to ignorance is an extremely delicate one; the practical information analyst will usually be satisfied to treat any uncertainty as due to noise, which results in the greatest reduction of certainty. This interpretation will be subject to revision in the light of additional knowledge. The two-part system 'father's height-daughter's height' is not a communica- tion system, and this is one reason why so little information is transmitted. Suppose the numbers which define the 'father's heights' categories were not observed in a given population but could be chosen arbitrarily; for instance, they might be input voltages applied to a system. Accordingly, the 'daughters' heights' might be output voltages, and the table of conditional probabilities becomes a statement of the transfer function of the system. It is obvious that this system can be made to transmit more than 0.22 bits per symbol. For instance, using onlyy = 59.5 andy = 74.5, with equal frequencies, one would transmit about .90 bits per signal. In general: for each channel, Piij), there exists a set of input probabilities, p(i), which maximizes the transmission rate. The rate so obtained is called the channel capacity. Even with best utihzation of the possibilities of a channel, it can do no more than transmit all the input information, and in general it will not transmit quite all of it. This leads to an important generalization : Manipulation of information cannot increase its amount; it can at best preserve it, and it is likely to reduce it. This important statement will be clarified by the discussion of an apparent exception. Suppose A wishes to send a message to B over the channel C; conditions being very good, B picks up not only almost perfectly the message sent by A but acquires, in the course of doing so, considerable amount of information about conditions in the channel. His total information received might be more than that contained in A's message; still, he has lost some of the information contained in the message. In general: as a result of manipulating information, there can be more output information than there was input information — but the contribution of the input information to the total cannot be more than the amount of input information. Error Detection and Correction A codebook states which output should be associated with any given input. A noise-free channel fulfills these requirements perfectly. In a noisy channel A Primer on Information Theory 33 Other outputs than the required ones appear; in other words, a noisy channel produces errors. Errors lead to loss of information, and a reduction in the rate of transmission; in a noisy channel, Tix;y) 0. This loss is unavoidable. However, it is at least possible to spot and correct the errors which have occurred. It is one of the main endeavours of information theory to devise methods to do this efficiently. An error in a message can never be found unless the message contains some extra information which can be used for this purpose. For instance, if the message consists of a string of four digits chosen without any constraint: 5 3 8 7, one has absolutely no possibility of knowing whether or not it contains any errors. If it has been agreed upon that the message will be repeated, then one can detect errors : 5 4 8 7 5 3 7 7, and if the message is repeated several times, these errors can be detected and corrected, with arbitrary certainty if the number of replications can be made sufficiently large: 5 3 8 7 5 3 7 7 5 3 8 7 5 4 8 7 5 3 8 1. In the second case, the possibility of error detection was bought at the price of making two digits do the work of one; the message is said to be 50 per cent redundant. In the last case, the price of error correction is the use of five digits to transmit a single one, or a redundancy of 80 per cent. Introducing redundant information in the fonn of a simple replication is straight-forward and eiTective, but not very economical. Error detection could be achieved more efficiently by simply adding the sum of the digits to the message: be achieved more efficiently by simply adding the sum of the digits to the message : 5 3 8 7 2 3. Here, the redundant information is only one-third of the total. In fact, giving only the last digit of the sum as 'signature' is almost as effective, and requires only 1 digit in 5, or 20 per cent redundant infonnation. The signature check illustrates a general principle: a given amount of redundant infonnation in a 34 Henry Quastler message can be used for error checking the more effectively the more evenly it is related to all parts of the message. It is always possible to achieve reliability, in the presence of noise, by the use of redundant information; in fact, one can approach perfect reliability arbitrarily closely if one is willing to provide enough redundant information. The amount of redundant information needed, for a given noise level and a given desired reliability, will depend on the efficiency of coding. The ideal relation between noise level and redundant information needed is formulated in Shannon's fundamental Theorem of the Noisy Channel. This theorem can be stated as follows: if a certain amount of information is to be transmitted with perfect reliability in the presence of noise, then it is necessary to provide at least as much redundant information as the amount of equivocation introduced by the noise ; furthermore, this amount will be sufficient if the coding is maximally efficient. There exist several proofs of this theorem; none of them is easy to follow, and all are existence proofs — that is, they prove that an error-checking code exists which will fulfill the requirements, but they do not say how to construct it. In fact, perfectly efficient error-checking codes seem to be realizable only in a few special cases; however, close approximations to ideal efficiency are easily obtained if it is permissible to use message blocks of great length (12). The economics of error-checking are dominated by three factors: (I) the frequency and costliness of errors (II) the cost of adding redundant information (III) the availability and costliness of checking procedure (encoding and decoding). The work of Shannon and his followers has dealt with one particular situation : encoding and decoding procedures are supposed to be reliable and gratis, the error frequency is to be reduced to almost zero, and redundant information is supposed to be used as sparingly as possible. As long as the theory is not completed even for this case, one cannot expect to develop a more general theory. Some qualitative notions of what it will entail can be gathered from a considera- tion of a much-used, and presumably well developed communication system, namely, printed language. Symbols are gathered into various checking units (words, sentences, paragraphs, chapters) ; on each level, there operate constraints which will help to locate and correct errors. For instance, this sentence will be read corretly even though one letter has been onitted and one word misspelled. It 3eems that the redundancy per letter, in a coherent English text, is about 60 per cent. Paragraphs are constructed in such a way that the sense can be grasped even if whole words or even sentences are missing or perturbed, and the essence of a whole chapter is, in general, understandable even if a whole paragraph should be left out. Actual Communications System So far we have dealt with two-part systems in a purely abstract way. 'Sources' and 'destinations' are defined simply by the states which they can assume. 'Channels' are tables of conditional probabilities; in the simplest case, the channel is a kind of telephone book which associates every input to some A Primer on Information Theory 35 particular output. If the association is not unequivocal, then the channel is said to be noisy. 'Noise' is defined as a random perturbation of the input- output link. Those are nice, clean concepts, not to be confused with realities. The 'channel' exists on paper only, and is not the same as the mechanism which links two parts of a system. The infonnational relation between heights of fathers and daughters does not reveal the nature of the mechanisms involved; whether fathers affect their daughters' heights by means of their genes, or of the food they provide, or of the mother they select for them, cannot be decided on grounds of informational relations. Indeed, I believe that Buddhist tradition would explain the correlation on the grounds that daughters select their fathers; as far as information theory is concerned, this is perfectly acceptable. The scheme shown in Fig. 5 is a somewhat closer approximation to reality: NOISE SOURCE MESSAGE. • ENCODER TRANSMITTER SIGNALS CHANNEL SIGNALS DESTINATION [J^^SSAGE ^ qe-qqqer l.^ RECEIVER (— ' Fig. 5. A diagrammatic representation of a communication system It is customary to treat all links but the channel as noise-free. If need be, one can introduce noise into the other links of the model by some straight-forward adaptations. If signals and channels are physical entities, then it is relevant to investigate their physical capacity of carrying information. Suppose the nature of a unit of action and the physical constraints are such that the channel can assume any one of m states during one unit of action; then, these states can be made to represent log., m bits of information. It is the function of the encoder-trans- mitter system to match the diversity of messages generated by the source to the diversity of states which can be assumed by the channel; those, in turn, are matched to the diversity of messages intelligible at the destination by the receiver-decoder system. As long as the demands on the channel are light, the matching process is not much of a problem. However, it may become very difficult if the channel is to be driven at capacity, and if the various states of the channel are not of equal value; some may be more subject to noise effects than others, some may need more time than others, some may necessitate more effort than others. In general, one will tend to favor the safest, shortest, and easiest states. However, this must not go too far; if one goes to the extreme of using the very 'best' state, then the channel does not transmit any information at all. To find optimum compromises between informational needs of source and destination and physical capacities of the channel, between amount of information used to carry messages and amount of information needed for noise reduction, is one of the fundamental problems of the theory of information and communication. 36 Henry Quastler EXERCISES 13. The following table gives the number of times the four possible combinations of two flower colors with two pollen shapes were found : Pollen shape Flower color : Purple Red Long 296 27 Round I 19 85 Is there information transmission between these two characters ? 14. Define the following functions, and derive their values (in terms of //-functions) J{x,y; z) T(x; y, z) T(x; y; z) 15. Ad agnostic test gives the following results: true negatives . . false negatives . . true positives false positives . . 85% 5% 3% 7% What is the informational value of the test? What is the maximum informational value that any test could give in this situation? 16. A teletype machine sends 2.3 groups of five binary symbols per second. What is the maximum possible rate of information transmission ? 17. Same machine as in Exercise (16). All code groups are equiprobable. Error probabili- ties are as follows: symbols nos. 1 and 4 are always received correctly, nos. 2 and 3 are wrong 1 1 per cent of the time, no. 5 is wrong 1 per cent of the time. All errors are equiprobable. Compute equivocation and amount of information transmitted. 18. You are to send 2-bit messages through a channel which has the property that one in five binary symbols is bound to be in error. Construct four sequences of five binary messages which will allow the reconstruction of the original message. What is the efficiency of the code? V. ORGANIZATION Systems, Structures, Pattern A system is an organized whole made up of interrelated parts. Organization is based upon the interrelations between parts. The parts may be strongly or weakly coupled; their effect on each other may be quantitative or qualitative. (Z> Kr) Fig. 6. A simple communication network If two parts are coupled in any fashion, then knowledge of the state of one must imply some information about the state of the other. Accordingly, any interrela- tion can be technically represented as a channel. So, two components of a system can be symbolically represented by a simple communication network of two parts, referred to as two nodes and one channel: A Primer on Information Theory 37 Let H(x) be the amount of infonnation needed to know what state x is in. If y is known, some of this information becomes unnecessary, or redundant. This amount, T{x;y), is an index of the degree of coherence, constraint, integra- tion, or organization which prevails in the system. Consider the pair of words 'green valley'. These two words form a small system — a whole made up of interrelated parts. The whole has a meaning which neither part alone has. The price for this feature is elimination of many other possible connotations of 'green' and 'valley'. As a result, the information content of the word combination is smaller than the combined information contents of the two words. The difference must show up as redundant informa- tion. The presence of redundancy implies that each word contains some information about the other. This is best demonstrated by successful error checking. The errors 'preen 'for 'green', and 'volley' for 'valley' would not be found in isolated words, but can be spotted in the pair. System Analysis — There seem to be three general viewpoints under which relations within a system are assessed: (a) the amount of information trans- mitted— on the technical, semantic and pragmatic level ; (b) the degree of control or cause-effect relations, dominance; and (c) the utility, or value, of the relation to one or both of the related parts. Information theory deals only with the first viewpoint. It does not concern cause-effect relations, or what causes the information to flow, and it is not concerned either with the utility of the flow of information. Informational analysis of a system will be of interest if and only if the informational challenge is serious, that is, when a system has to process informa- tion at a rate which crowds its capabilities. The informational challenge is the result of: (1) The diversity which is characteristic of the tasks; this can be expressed as ///task. A system which is faced with the same task all the time or most of the time may be working very hard but the difficulty is not an informational one. (2) The precision which is required ; this can be expressed as the ratio TIN. That is, the diversity of tasks is informationally challenging only insofar as it is expressed in a diversity of responses. A system with a small response repertoire may be working very hard, but not in the informational domain. (3) The time which is allotted for the fulfillment of each task. A system with very modest informational equipment can solve many tasks if given ample time. For instance, the extremely simple logical machine devised by Turing (13) will solve any solvable problem if given very much time. The time rate of informational challenge of the system is the product H T tasks _, . . X 7> X — TT — = Tlumt time. task H unit time The infoiTnational output of the system will be measured in //-measures but the effective output, or informational performance, in terms of T-measures, as T per task or T per unit time. The limits of the informational performance of a system can be found by systematically varying the informational challenge and observing the resulting performance. In such studies it is important to 38 Henry Quastler make sure that the system's performance is hmited informationally, and not by difficulties of sensing inputs or generating outputs. It is possible to vary the informational challenge in a number of modes; e.g. one can vary the number of sources of information, or the amount of information per source. Challenging in various modes reveals whether or not there exist several modes of limitation. It seems that the informational perfor- mance wliich a system can produce in single tasks may be limited by the follow- ing factors, singly or in conjunction : (1) the amount of information which can be processed effectively in a single task, (2) the number of independent information-carrying components which can be involved in a single act of infomiation-processing, (3) the informational contribution from each independent component, (4) all information-carrying components must be assembled within a certain length of time ; (5) in addition, there seem to be two general limitations on time rates: there is a minimum time for each act of information processing, and (6) the over-all rate of information-processing is limited (only this last limitation has the character of a channel capacity). This list of limitations is based on psychological experiments (14) but is believed to apply to all types of systems. Multi-part Systems — The informational system analysis is not restricted to two-part systems. A system of three components can be represented as a three-node network with a connecting channel: Fig. 7. A simple three-node network Again, it is merely a matter of convenience which node, or set of nodes, one treats as the input, or independent variate. The treatment can be extended to any number of components. Thus, a nine-node network is equivalent to one man receiving infomiation from eight sources, or feeding information into eight sinks; or, to four men watching two sources, communicating with each other, and feeding information into three sinks; to a sentence of nine words; to a decision based upon eight factors. The more parts there are to a system, the more difficult becomes the infor- mational analysis (15, 16). This is territory that has been but recently opened, and we are still largely concerned with the formulation and highly tentative application of concepts. It will be helpful to consider a parallel effort, namely, the study of organization by game theory (17). One result of this study is that each time a new player is added, the organization (the 'game') acquires a new qualitative feature. One-person games deal with problems of maximum; the addition of a second person introduces competition; of a third person, coalition; of a fourth person, an asymmetric role of one player in relation to the group of the other three, von Neumann (17) points out that it is at this junction that the most remarkable problems begin to appear; also at this junction, A Primer on Information Theory 39 there occurs a change from a rigorous and complete exposition to a heuristic and incomplete one. The situation is similar in the study of organization by information theory. Each time a new part is added to a system, a qualitatively new information function appears. As long as one deals with a single variable, the problem is one of efficient use of existing variations. A two-part system introduces relations between parts; a three-part system, relations between relations; a four-part system, relations between a part and a complex of relations. Unitization — It is an empirical fact that when a system is complex enough to require very many components, the phenomenon of unitization occurs. That is, some components get organized in such a way that they interact strongly among each other, and act as a unit with respect to the remainder of the system and the external world. Unitization seems to be a necessary evil; it might be an important key for the study of complex organization and complex mental activities. The phenomenon has never been really explained; it is possible that a quantitative treatment will be made possible through the use of infor- mation theory (18). Unitization is always coupled with the phenomenon of limited span. Any real part has a limited information content. In any single act of communication, the capacity for non-redundant transmission of a part is limited by its own infomiation content. This amount must somehow be partitioned into inter- action with the external world, and interaction with the other members of the unit. If each of these interactions is to be of significant size, then only a limited number is possible. The interaction of a unit with the outside may be only a fraction of the information traffic within the unit. Hence, several units can be organized into a secondary structure of greater versatility, and this process can be repeated on successive levels of organization. There appears, thus, a possibility that information theory can be helpful in formulating both the causes and the effects of unitization, and in establishing rational interpretations of the size of the units. This would be a very important contribution to any theory of organization. Conclusion — We have proceeded from simple processes of representation to discussions of communication and, finally, organization. It was attempted to treat in a heuristic and perspicuous manner the basic principles of Information Theory: there exists a generalized concept of 'information' which includes communication and organization and is so general that every real event or structure has its informational aspects; this general concept is related to a measurable quantity; the operation of taking a measurement of this quantity is done by means of symbolization in a standard language. The functions as defined obey two fundamental theorems: the Representation Theorem, and the Theorem of the Noisy Channel. Both theorems impose a limit on the amount of information which can be effectively processed in a given situation; both also state that it is possible to reach this limit. APPENDIX I THE EVALUATION OF INFORMATION CONTENT The examples and exercises should have familiarized the reader with the techniques of taking information measurements. However, the investigator 40 Henry Quastler who wishes to use this knowledge in his field is bound to run into some diffi- culties. A typical difficulty is that a natural situation does not present itself neatly classified with a complete set of categories and probability measures. It often takes considerable ingenuity to supplement the missing components of the picture. Wherever ingenuity must be used, the result will not be unequi- vocal. Hence it becomes important to estimate not individual information measures but rather whole ranges compatible with reasonable assumptions. The Relativity of Information Measures 'Information content' is a measurable quantity, just as length; and, just as length, it is a function and not a property of a particular set of events. The theory of relativity asserts that the measured length of an object depends on certain relations between the object and the measuring system. However, under everyday conditions these relations will not produce any significant effect and, most of the time, lengths behave as if they were properties of objects. The infomiation content of an event depends on the manner in which this event is related to the frame of reference of the evaluating system. Unlike with length, these relations are not fixed under everyday conditions. Therefore, information content behaves only rarely as if it were a property of an event. The amount of information, H{x), associated with an event, x, is defined as the expectation of the logarithm of the probability that x will fall into some category, /. Thus, the measure of information depends on three decisions: (1) the choice of a unit event, (2) the establishment of categories, (3) the selection of a set of probabihty measures. In general, each of these decisions involves a degree of arbitrariness. Accor- dingly, a considerable range of information measures will be compatible with a given real situation. The question of an appropriate selection of a unit event cannot be solved by mechanical application of hard and fast rules. There is a lower limit to the size of elements, imposed by limits of observability. In general, selection of these lower limits will force one to take cognizance of a tremendous amount of detail, most of which is bound to be irrelevant. Thus, one will try to select a unit event broad enough that all irrelevant details are submerged in its internal structure, yet narrow enough so that no relevant relations get lost within the unit event. In practice, one has to make a guess, subject to revision by later experience. This difficulty occurs with all kinds of analyses, and is not specific to informational analysis. The situation is quite similar with respect to categories. There, too, exists a bound, imposed by the capabiHties of discrimination. In general a large number of discriminations can be made which are irrelevant to the problem at hand. For instance, if one deals with the semantic content of a printed message, it will be quite irrelevant to categorize by shapes of letters, quality of paper, type of printing ink, etc. The decision is not always so easy. For instance, in categorizing the atoms found in living matter it will, by and large, not be necessary to distinguish between isotopes; in the overwhelming majority of occasions, differences between isotopes will have no effect. Occasionally, of A Primer on Information Theory 41 course, a particular isotope located in a sensitive spot and decaying at a critical moment can have very large effects. In a case like this, the selection of a set of categories becomes a matter of compromise. The probabilities, finally, are never actually known. We have to estimate them, on more or less sound bases. In many situations where generalized information theory is used, the bases for estimating probabilities are rather uncertain. Therefore, it becomes important to assess the dependence of information functions on fluctuations of probabilities. The contingent nature of information measures has not always been obvious. All early applications of infomiation theory dealt with telecommunication systems. In all of these, all informational characteristics are perfectly well defined. In Morse code, all we have to know is whether a particular information- carrying element is a blackness or a whiteness, and whether it is long or short. In pulse code modulation, the only thing that counts is presence or absence of a pulse within a stated interval of time. In pulse amplitude modulation, all information is vested into the amplitude of pulses. In all these cases, there is no question about the infomiational characteristics of the process under consideration. The situation is radically different in the larger domain of applied infor- mation theory. For instance, take the case of two people transmitting information to each other by talking. The information-carrying element is a clause; to simplify our analysis, let us consider just words (remembering that the infor- mation content of a clause cannot be greater than that of its constituent words). Now, each person culls his words from a reservoir which is known to be large, but its actual size is not exactly known. The information content of a single word depends on the probability of its use, and these probabilities are not exactly known either. Furthermore, they will hardly be the same for both persons involved in a conversation. Also, each word can have several meanings, one of which may be more or less determined by the context. The relations between words, meanings, and context, again, are not the same for any two people. This is not all. Information is conveyed not only by the choice of words but also by inflection of voice, loudness, timing, and accompanying gestures. In such a situation we have obviously no hope ever to obtain a precise, unequivocal, and incontestable measure of information content. We are, thus, confronted with two alternatives. These are: not to use infor- mation theory, or to try to devise ways of producing usable approximate estimates. Obviously, our choice is the latter alternative (19). Approximation MetJiods It appears that the approximation methods to estimate infonnation functions are based on the following rules: 1. Averaging increases uncertainty; 2. Pooling decreases uncertainty; 3. Disregarding constraints increases uncertainty; 4. Rare events have small effects on uncertainty measures; 5. Smafl variations in probability have small effects on uncertainty measures; 6. In systems, information functions can be estimated in different ways, and care should be taken to select the most appropriate one; 42 * Henry Quastler 7. If it is not possible to measure the actual infonnation functions desired, then one can try to substitute closely related measurable quantities. In the following paragraphs, these rules will be amplified and illustrated. 1. Averaging Increases Uncertainty — The fact was demonstrated in Section III. It suggests a simple bracketing procedure: obtain a lower and upper bound of uncertainty by using probabilities which are certainly more and less unbalanced than they actually are. In particular, if the number of categories is known but their respective probabilities are not, then one can follow Laplace's procedure and set all probabilities equal which maximizes uncertainty. 2. Pooling Decreases Uncertainty — This, too, has been proven in the third section. It is equally of value in bracketing procedures: using only categories actually discriminated puts a lower bound on uncertainty; assuming more categories than could be of interest establishes an upper bound. 3. Disregarding Constraints Increases Uncertainty — Let x and y be different events, where y may differ from x only in time or place of occurrence or in any other respects. If H(x) is the uncertainty of x, and Hy(x) the uncertainty of .Y if y is known, then: H^x) < H(x). That is, knowing some other event, y, cannot increase the average uncertainty concerning x; it will leave it unchanged if there is no association between x and y; it will reduce it if constraints exist which are manifested in a statistical association between x and y. Rule 3 can be used for a bracketing procedure. Disregarding constraints yields an overestimate of H(x) ; introducing constraints known to be too strong, an underestimate. Constraints have to be very marked to cause large changes in H(x). For instance, the large inequalities of letter frequency in English texts reduce H from a possible maximum of 4.7 bits per letter to 4.1 bits; the strong constraints between successive letters and words result in an additional reduction to 1.5-2.0 bits per letter. Formally, rule 3 is a special case of rule 1. 4. Small Effects of Rare Events — The information functional is a sum of terms of the form (—p log/»). This function rises steeply between zero and .10, hence, small probabilities contribute little to the total sum. For instance, ten equiprobable alternatives correspond to an H of 3.32. If one of these alternatives is replaced by ten separate sub-categories, each of probabihty .01, then the resulting H is 3.65. If instead of ten, one introduces 100 equi- probable sub categories, each with probability .001, the resulting H is 3.99, or equivalent to sixteen equiprobable categories. A good example turned up in a study by A. A. Blank. He calculated the information content of single Enghsh words. For particular reasons, the sample was restricted to four letter words. Thorndyke's list contains 1550 such words. H, based on the observed frequency of these words, is 8.13 bits per word. Of these words, 119 occur with the greatest frequencies. Computing H on the basis of these words alone gives a value of 6.34 bits per word. Thus, taking into consideration only about one tenth of all categories already yields about four-fifths of the final information function. A Primer on Information Theory " 43 This means that information functions can be estimated successfully as soon as the more common occurrences are categorized. The remaining infrequent occurrences will not contribute very much, and that contribution can be easily bracketed between values based on numbers of categories which are certainly too small and too large. 5. Small Effects of Small Variations in Probability — The curve of the function F(p) =^ —p log p has a flat top. Small changes in probability in this region have small effects. Consider the simplest case, of two categories. If their probabilities are equal, then //= 1. If the ratio of the probabilities is 1:2, then 7/= .92. If the ratio is 1 :3, a very considerable deviation from equality, H is still .81. For a larger number of categories, the insensitivity of H against probability distortion is still mOre pronounced. If one replaces equiprobable alternatives by probabilities staggered arithmetically or geometrically, stipulating only that the span between the extreme value should be not more than one order of magnitude, then the resulting changes in //are quite small. This implies that the assumption of equiprobability, which gives an upper bound as stated in rule 1, will not go very far from the true value unless proba- bilities are radically unbalanced. The stretch bracketed between an upper bound based on equiprobability, and a lower bound based on a distortion undoubtedly stronger than the real one, will not be very large. 6. Alternative Ways of Estimating Information Functions — In systems with several nodes, the compound infonnation functions can always be esti- mated in several ways. For instance, in a two-node communication system, the quantity which is the function of greatest interest, the amount of information transmitted, T(x;y), can be computed in three alternative ways: as differences between input uncertainty and equivocation, as difference between output uncertainty and ambiguity, or as difference between the sum of uncertainties of input and output and the uncertainty of their union. It usually is worthwhile to inspect the data very carefully to estabhsh which of the set of functions can be most easily and most accurately computed. In many cases, the quantities most readily computed are not those which result directly from the plan of obser- vation or experimentation. For instance, in most experiments it would be natural to measure output uncertainty and ambiguity, but it is easier to measure input uncertainty and equivocation. 7. Substitution of Related Quantities — In many cases where it is not practical to compute the proper information measures, one can compute information measures associated with related quantities. Take the case of estimating the amount of information v/liich an individual can transmit after a single glance at a display. This quantity is very difficult to determine; but, it is fairly easy to determine the amount of information which can be elicited from an individual by a short interrogation procedure after he has had a glance at the display. This function is not quite the one we want, but presumably closely related to it. Another example: in the case of mental arithmetic, we have no way of estimating the actual amount of information processed, but we can readily estimate the amount of information which must be processed if computations are done in the way in which the subject claims he computes. In cases of this kind one will use the measurable quantity instead of the desired one. Of 44 Henry Quastler course, results so obtained have to be used with a certain amount of restraint. Example: Rate of Information Transmission in Conversation — ^The working of the approximation methods can be shown by two examples. The first example is that which we used to illustrate the need for approximation methods; namely, that of estimating the amount of information in conversation. We consider first the infomiation carried in words. To establish an upper bound, we ask how much information must be transmitted so that the receiver can recognize every single word spoken. This upper bound, in bits per second, is the product of the rate of words per second times bits per word. A rate of 2.1 words per second is typical for lively discussions. The number of bits per word in English context has been estimated as 6.5 bits (±25 per cent). This yields 11 to 17 bits per second. Words are not the only method of communication between two persons conversing face to face. It can be shown, however, that all other means of communication add little to the total transmission rate. We will now try to establish a lower bound. Of course, no general lower bound exists; it is easy to find examples where infomiation is transmitted at the rate of 1 millibit per second, or less. What we want is an 'upper lower bound' a lower bound of the amount of information transmitted between people who try to communicate at some speed, and under reasonably favorable conditions. Such a bound is obtained by analysis of pragmatic communication. We look at situations where the verbal messages elicit or control actions. We make an informational analysis of the relations between actions and verbal messages. This will yield an amount of information demonstrably transmitted, and it certainly represents a lower bound to the amount of information com- municated. At this time, we have a single case where pragmatic communication has been evaluated accurately in informational terms. Felton, Fritz and Grier (20) measured the amount of pragmatic communication between an airplane pilot coming in for a landing and the control tower operator. They found an average rate of 2 bits per second, computed in terms of actual effects of the messages. Both pilot and control tower operator have all interest to communicate as fast as they can. On the other hand, they do so in the presence of a very high level of noise which reduces verbal communication to probably about one third of its optimum rate. We conclude, thus, that information transmitted through verbal communi- cation is certainly not less than 2 bits per second nor more than 17 bits per second, and very likely within the range between 6 and 12 bits per second. This estimate is rough but not at all unrealistic. Example: Information Content per Printed Letter— A very elegant way of computing an information measure under unfavorable conditions was used by Shannon in his analysis of the 'entropy' of printed English (21). The information content of a single letter is easily determined as a function of relative letter frequencies. However, constraints between neighboring letters lead to a reduction of information content, and in order to estimate this reduction exactly one would have to investigate the probability distributions for long sequences of letters. This is manifestly impossible. Shannon, therefore, proceeded to estimate a related quantity; namely, the amount of information A Primer on Information Theory 45 concerning language constraints which can be ehcited from a person familiar with printed English by a carefully planned interrogation. The subject is given a text which is truncated at some point; he is asked to guess the next letter. If he is successful, then he is told to go on; if not, he is told to try again. Records are taken of the number of times a letter is correctly identified at the first, second, third, . . . statement. In this setup, the experimenter acts as source of auxiliary infoimation, emitting sequences of the type 'wrong . . . wrong right', with an 'alphabet' of twenty-six different sequences (if repetitions are excluded, the letter must be identified after no more than twenty-five wrong guesses). The informational output of the auxiliary source depends on the relative probabilities of the various sequences. These probabilities are very unequally distributed. In a large percentage of the cases, the first statement is correct; the most frequent message from the auxiliary source is 'right'. The next highest probability is for the sequence 'wrong-right'. Messages with up to three 'wrongs' make up the vast majority of cases; the remaining categories, with from 4 to 25 'wrongs', have low probabilities. As was pointed out before, they contribute little to the estimated value of H. This means that we arrive at an estimate of the information furnished by the auxiliary source essentially as a function of two to four probabilities. The amount of information per single letter is known to be about 4.1 bits (on the basis of relative frequency of letters in English texts). This is the amount of information per letter which the subject needs to reconstruct the whole text. Of this amount of information, a certain measurable fraction is furnished by the auxiliary source. The remainder must come out of the subject's head, and is based on his knowledge of language constraints. The amount of infor- mation so elicited will not be quite as high as the information content of language constraints, but it is a closely related quantity. By the ingenious trick of effectively reducing the size of the alphabet, this quantity has been made easily measurable. APPENDIX II ANSWERS TO EXERCISES 1 . One light — peace and quiet two lights, vertically — enemy approaches by land two lights, horizontally — enemy approaches by sea two lights, diagonally — enemy approaches by land and sea (This is not the only possible solution) 2. (a) 0, 1, 10,11,100,101,110,111, 10000, 1001,1010,1100,10000, 11110100011 (b) 9, 11, 147,32 (c) .125, .6703125 3. (a) 10110100010000 (b) EDCBA 4. 'Construct a confusion-free code using five binary digits for each letter and compare the performance of this code with that of the above by encoding and decoding a message like this one'. Use part of the 32 code words made up of 5 binary digits, such as: 1 1 1 1 1 , 1 1 1 10, 1 1 101 , 11100, etc. The message will be, on average, 21 per cent longer than with the most efficient code (5 is 121 per cent of 4.14), but it is much easier to decode. Some of the unused code words can be used for punctuation, etc. The teletype works on this principle. 46 Henry Quastler 5. Limiting value : -(.8 logo .8 + .15 log, .15 + .05 logo .05) = .883 Single event code: A 1 .8 B 0 1 .3 C 0 0 .1 1.20 -0.883 1.20, excess is — = 36 per cent. 0.883 Two-event code: Event pair Prob. Code AA .64 1 .64 AB .12 0 1 1 .72 BA .12 0 1 0 AC .04 0 0 11 .32 CA .04 0 0 10 BB .0225 0 0 0 1 .09 BC .0075 0 0 0 0 1 .0375 CB .0075 0 0 0 0 0 1 .06 CC .0025 0 0 0 0 0 0 1.0000 1.8675 .934 digits per event excess = = 5J%<10% 6. Let X designate amino acids ;, and y nucleotides. nx = log; J 20 = 4.322 Sx=l tty = log ;, 4 = 2.0 1 > ' ^-^^^ - 2 161 Sy 2.0 ^-^^^ 7. P -p\og,p .60 .44 .40 ^ H(x) = .97 8. The curve looks similar to F(p), but has a flatter top and is symmetrical, with a maximum of 1.0 at p{l) = .50. 9. H(x) = log2 3 = 1.58 -p log, p y:.S .1 .26 .33 .05 .22 .05 .22 i/(j) = mx) 1.03 my) A Primer on Information Theory 47 10. A realistic description of his uncertainty might be: prob (55-64) = .95 prob (55-54) -- .02 prob (65-70) = .02 prob (any other speed) = .01 Within each range, all speeds are considered equiprobable. We will derive the answer in two steps, obtaining first the uncertainty as to the speed range: Range p —p log., p 55-64 .95 .07 50-54 .02 .11 65-70 .02 .11 any other speed .01 .07 .36 bits Next, we observe that the range from 55 to 64 miles per hour contains ten speeds (deter- mined to the nearest mile) which are equiprobable. The uncertainty measure for ten equi- probable categories has been found to be log., 10 = 3.32. This uncertainty will arise 95 times out of 100; its expected contribution to the total uncertainty is 3.32 ■ 0.95 = 3.15. The other ranges are treated equally : Range No. of sub-classes ir) log^r P • logo r 55-64 10 3.32 3.15 50-54 5 2.32 .05 65-70 5 2.32 .06 all other 81 6.35 .06 3.31 bits We thus need (on average) .36 bits to determine the range of speeds, and an additional 3.31 bits (on average) to identify the speed to the nearest mile, within the range. The total uncer- tainty is 0.36 + 3.31 = 3.67 bits. Of course, different expectations would yield different uncertainties. 1 1 . The letters occur with more nearly equal frequencies. 12. Two bits. „, . , /323 323 104 104\ ^^ ^. 13. i/(shape) = - — log, h — lo", — -- .80 bits \427 ^-427 427 "■427/ rrr , ^ /315, 315 112, 112\ ^. ,. //(color) = - — loga h — log., — = .83 bits \427 ^427 427 ^-427/ 17/1 K ^ ^96 , 296 27 , 27 19 , 19 //(color, shape) = - — log, 1 log., 1 log., — \ 427 ^- 427 427 ^' 427 427 ^" 427 + ^log„^) =1.26 bits 427 ^- 427/ r(color; shape) = .80 -I .83 - 1.26 = .39 bits 48 Henry Quastler 14. T{x, y; z) = mutual reduction of uncertainty between x and y on one hand, z on the other = H(x, y) + Hiz) - H(x, y, z) nx;y, z) = H(x) + H(y, z) = H(,x,y, z) T(x;y; z) = total constraint in a tri-variate system = H{x) + H(y) + H{z) - H{x,y, z) 15. Test Actual pos neg pos 3 7 10 neg 5 85 90 8 92 H{y) = .40 H(x) = .47 H(x,y) nx;y) .84 .03 The informational value of the test is .03 bits. Its maximum possible infonnational value equals the amount of uncertainty before the test, viz. .40 bits. 16. 2.3 X 5 X 60 = 690 bits/minute 17. Begin by computing the output uncertainty. The probabilities of receiving each signal are obtained as the sum of receiving it correctly (0.2 for Nos. 1 and 4, .178 for 2 and 3, .198 for 5) plus the addition due to errors (1/4 of the errors, for each erroneous transmission). This procedure yields //(out) = 2.32 bits. Next, compute the ambiguities. These are zero for symbols no. 1 and 4. For 2 and 3, the ambiguity can be computed as the sum of the information needed to ascertain that an error has occurred (—0.11 loga 0.11 — 0.89 loga 0.89) plus the information needed to find out which of the possible and equiprobable four errors has occurred, which is 0.11 x 2.0 bits/symbol. Symbol no. 5 is treated similarly. The average of the ambi- guities is 0.31 bits, hence T equals 2.32 — 0.31 or 2.01 bits — a loss of about one-sixth of the input information. 18. One solution is the following: 11000 10101 OHIO 00011 A single error will result in the reception of a word which is not in the code book. If one follows the rule of substituting that message in the code book which differs from the received one by one digit only, then every error (provided there is only one!) will be corrected. A five-digit binary message can carry five bits of information. If it is known that one error has occurred somewhere in a group of five symbols, then the information needed to locate the error is loga 5 = 2.33 bits. With maximum efficiency, one should use only 2.33/5 or 46.5 per cent of redundant information (which could be achieved by coding large sequences of five-digit words!). In our case, the redundant information is 3/5 or 60 per cent, and we trans- mit with an efficiency of 40/53.5 = 75 per cent. (Observe that there is less uncertainty if it is known that there is one error in every five-symbol word, than when it is only known that the error rate is 20 per cent !) A Primer on Information Theory 49 REFERENCES 1. L. Szilard: tJber die Entropieverminderiing einem thermodynamischen System bei Eingriffen intelligenter Wesen. Z. Phys. 53, 840-856 (1929). 2. R.A.Fisher: On the mathematical foundations of theoretical statistics. Phil. Tram. {A) 222, 309-368 (1922). 3. R. V. L. Hartley: Transmission of information. Bell Syst. Tech. J. 7, 535-563 (1928). 4. N. Wiener: Cybernetics, J. Wiley and Sons, New York (1948). 5. C. E. Shannon: A mathematical theory of communication. Bell Syst. Tech. J. 27, 379-423, 623-656 (1948). 6. C. E. Shannon and W. Weaver : The Mathematical Theory of Communication, University of Illinois Press, Urbana (1949). 7. L. N. Ridenour: Computer memories. Sci. Amer. 192, 92-100 (1955). 8. R. M. Fang: The transmission of information. Tech. Rep. Mass. Inst. Tech. Res. Lab. Electron., no. 65 (1949) 9. L. Brillouin: Science and Information Theory, Academic Press, New York (1956). 10. L. DoLANSKY and M. Dolansky: Tables of log-, Ijp, etc.. Tech. Rep. Mass. Inst. Tech. Res. Lab. Electron., no. 227 (1952). 11. E. Klemmer: Tables for computing informational measures. Tech. Rep. A. F. Cam- bridge Research Center, ARDC. 12. Articles by J. E. Golay, P. Elias, I. S. Reed, R. A. Silverman, and M. Balser in: Trans- actions of the I.R.E. Professional Group on Information Theory (1954). 13. A.M.Turing: On computable numbers, with an application to the Entscheidungs- problem. Proc. Lond. Math. Soc. 42, 230-265 (1937). 14. H. Quastler: Studies of human channel capacity. In: Information Theory, ed. by C. Cherry, Academic Press, New York (1956). 15. Wm. McGill and H. Quastler: Standardized nomenclature : an attempt. In: Infor- mation Theory in Psychology, ed. by H. Quastler, Free Press, Glencoe, 111. (1955). 16. H. Quastler: Information theory terms and their psychological correlates, ibid. 17. J. von Neumann and O. Morgenstern: Theory of Games and Economic Behavior, Princeton University Press, Princeton (1947). 18. H. Quastler, H. H. Chase, W. Montagna, M. V. Edds, Jr., P. F. Fenton, and P. B. Weisz: Essays on biological unitization. Rep. Control Systems Laboratory, Univ. 111., No. R-52(1953). 19. A. A. Blank and H. Quastler: Notes on the estimation of information measures. Rep. Control Systems Laboratoiy, Univ. 111., no. R-56 (1954). 20. F. Fritz and G. W. Grier, Jr.: Pragmatic communication. In: Information Theory in Psychology, ed. by H. Quastler, 232-243, Free Press, Glencoe, 111. (1955). 21. C.E.Shannon: Prediction and entropy of printed English. Bell Syst. Tech. J. 30, 50-64 (1951). SOME INTRODUCTORY IDEAS CONCERNING THE APPLICATION OF INFORMATION THEORY IN BIOLOGY Hubert P. Yockey Oak Ridge National Laboratory, Oak Ridge, Tennessee Abstract — The model of protein synthesis in the cell which has been built up as the result of the work of many researchers has been used as a basis for applying the principles of infor- mation theory in biology. The main Une of the argument has been the role of noise in the genome. The discussion has been kept as independent as possible of special models. It was shown that in a real organism noise must exist in the genome and that an ensemble of organisms may be represented by a probability distribution in H, p{H, A). Individuality is thus incorporated in a very natural way. Dancoff 's principle requires that there be a lower limit for viability for this distribution. Ha. The action of a deleterious agent which induces errors in the genome by acting on nucleo- tide pairs is assumed to be represented by an equation of the first order: ^ = -j(X)p,(j) + ija) where /(A) measures the effectiveness of the deleterious agent, of which A is a measure, in producing defects. A differential equation for H(X) is derived and it is shown that {dHldX)E^ as a function of A behaves like J{,X). I. INTRODUCTION Information theory finds its place in biological thought through its ability to deal quantitatively with organization and specificity. The importance of these concepts has long been recognized in biology, but this realization is rather sterile unless a quantitative form of expression can be found. One is reminded of a quotation from Lord Kelvin, 'When you can measure what you are speaking about and express it in numbers, you know something about it, but when you cannot measure it, when you cannot express it in numbers, your knowledge is of a meagre and unsatisfactory kind.' The need for expressing biological quantities in numbers is clear but solving the problem of how to do it is very much like belling the cat. Biology doesn't seem to have any problems both really simple and terribly important such as some which occur in the physical sciences. The application of first principles has come much more slowly in biology for perhaps this reason. That ideas of great general application do exist in biology is exemplified by Mendel's laws and by the theory of evolution. One of the purposes of this article, and indeed one of the purposes of this book, is to explore the practical and theoretical consequences that may be found in the discovery that biochemical specificity of proteins is carried, largely at least, by the exact order of twenty amino-acid residues. The suggestion of 50 Some Introductory Ideas Concerning the Application of Information Theory in Biology 51 Watson and Crick (1) that genetical infomiation is carried by the exact order of four kinds of nucleotide pairs provides a molecular vehicle for the genetic control of protein specificity. Gamow (2) was the first to see that this control implied the existence of a four-letter to twenty-letter code. Thus by following the logical consequences of purely biological, or perhaps biochemical, problems one is lead directly to a problem purely mathematical in character. This notion of the role of order, which is basic to information theory, is worth pursuing in biology since it provides a way of measuring what we are speaking about and expressing it in numbers. Furthermore, from the results of applying the theory to specific problems, we may obtain an experimental check on the validity of these ideas as first principles. In this article we shall apply these considerations to the storage and transfer of biochemical specificity. We shall explore, in particular, the role of noise in the genetical message. In my article in Part V the theory is applied to the practical problem of calculating and understanding survivorship curves. The present status of the means of storage and transfer of specificity is given by Gamow, by Ycas and by Augenstine in their respective articles in this volume. The question of the exact way in which information is destroyed by read-off error, radiation damage, aging, thermal fluctuations, biochemical side reactions, and so forth, is of equal importance. This problem is also discussed in this volume but no final and detailed account can be given at this writing. Nevertheless, since there is virtue in attempt, we shall attempt the development of a mathematical formahsm which is information theoretic in character. Most animals and plants exist at one time, at least, in the form of a single cell; we can consider that cell to contain a substantial part of the directions for the development of the organism. Since infonnation is conserved unless lost due to noise, it shall be assumed that the mature organism is characterized by substantially the same information content as the fertilized egg or seed. In order to fix the idea we shall develop the formalism on the basis of Watson and Crick's suggestion concerning the role of DNA. It should be remembered that the central ideas of this paper are independent of much of the detail embodied in Watson and Crick's papers and are dependent only on the possi- bility of genetical endowment being conveyed by a series of structures composing an information bearing molecule. Suppose we imagine the symbols A, B, C, D (Gamow's predilection is to the less prosaic spades, clubs, hearts, diamonds!) arranged in one-to-one correspondence with the nucleotide pairs of the DNA found in a particular given cell. The cell will have been selected from a number of similar but not identical cells in a colony under study. This colony may be thought of as being indefinitely large, so that in principle we may consider the ensemble of all possible organisms identifiable as being members of the colony. Since the number of nucleotides in DNA is finite, the number of elements in this ensemble is also finite. Because of this one-to-one correspondence it will be seen that the set of symbol sequences, which is the mathematical model of the ensemble of organisms, will contain the informational or specificity properties of the ensemble of organisms. 52 Hubert P. Yockey The importance or value of a theory lies, among other things, in its capability of treating a wide variety of phenomena from a single point of view. It is well to think, at the start, of the field of validity this theory may have and, if it should fail, the significance of its failure. If it should be discovered that Watson and Crick's suggestion has very little bearing or applicability then this development, while negative, is still a valuable result. One would then perforce search for another explanation for the great detail and specificity characteristic of any biological phenomenon. At present it is the most detailed proposal based specifically on molecular chemistry. The theory here developed is essentially statistical and may be expected to express its results in the form of expectation values, probabihty distributions, and their functions. The statistical character of the theory is directly in the line of thinking of both modern biology and modern physics. It should be kept clearly in mind that information theory deals with organizational problems and so some aspects of organisms will be outside its scope. In this sense it may be that the role information theory will play in biology will parallel that played by thermo- dynamics in physics and chemistry. II. NOISE IN THE GENETICAL INFORMATION The Instability of a Perfect System Let us consider an ensemble of organisms and discuss the communication of information from the DNA to protein. There is evidence discussed by Gamow and by Ycas in this volume that the code which translates information from the four-symbol DNA code via RNA to the twenty-symbol protein code is based on triads of nucleotide pairs. Indeed it can be seen that it must be at least the triads since a twenty-symbol alphabet carries 4.32 bits per symbol whereas the pairs in a four-symbol alphabet carry exactly four bits per symbol, assuming no intersymbol constraints. The triads carry six bits per symbol and so this represents some inherent redundance. It would be desirable to express this formalism in terms of the DNA triads of nucleotide pairs ; however, this requires a knowledge of the DNA to protein code. These data are missing. Our objective is to develop the mathematical fomialism in as simple a way as possible so it appears more appropriate to consider the communication of specificity from DNA to RNA. Here we are dealing with a coding between two four-symbol alphabets. Suppose we are considering an ensemble of organisms which is isogenic, and further that this means that each organism is characterized by exactly the same order of nucleotides in the DNA of its nucleus. We shall now show that this situation is unstable and that therefore a real ensemble of organisms will be represented by an ensemble of messages recorded in its DNA. From this it will follow that there is a distribution in the message entropy, characteristic of any ensemble of organisms, even one which is isogenic. The message entropy is H=H,-H, (1) where H^ is the message entropy of the genetical information and H„ is the loss of information due to noise. That is, //„ is the loss of information from Some Introductory Ideas Concerning the Application of Information Theory in Biology 53 some fault cither in the duplication process in the germ line or the somatic line or from incorrect rcad-o(T of any kind. //„ may be expressed in terms of the read-off or transition probabilities (3) of a letter of kind / to a letter of kindy, Piij). The probability of letter / is p{i). H=H,-\-y p{i) p^ij) log2 p,{j) (2) Consider the case where these probabilities are a function of some variable 1. In the application of these considerations A is the measure of some deleterious influence such as dose of ionizing radiation. Form the derivative dHjd?.: ciHldX = log2 e 2 (MO ic¥>^) P.ij) + Pii) loge pSi) {dIdX) p,{j) + P.(j)ioi,p,{J){dldX)p{i)] (3) The absolute value of dHfdX will become indefinitely large because of the second term in equation (3) as any p^{j) approaches zero if p{i) ^ 0 and (dldX) pi{j) 7^ 0. This may happen, in particular, if any p/ij) approaches one for then SL\lpi{k), (j ^ k) approach zero. This situation {p,{j) = 1) corresponds to the assumption that there is always a correct reproduction in the DNA duplication or in the RNA read-off. Under these circumstances the first term is finite and the third term is zero. Watson and Crick regard a mutation as being reflected by a change in order of the nucleotide bases in DNA. This is apparently always possible; they have suggested a biochemical scheme by which this can be affected. This means that in a real biological system p{i) ^ 0 and {djdX) p/ij) 7^ 0. A real ensemble of organisms will be represented by an ensemble of genetic messages. This will be true even if the ensemble is isogenic. Some noise must exist in the genetical information; if the noise is less than equilibrium it is quickly intro- duced. There is some experimental evidence in support of this conclusion. Burdette (4) prepared populations of isogenic Drosophila. One strain had the same low incidence of tumors in both sexes (about 4 per cent) and the other had a high incidence (about 60 to 80 per cent) even greater in males than in females. The tumor incidence of the isogenic strains was initially much lower in each case than the stock from which it originated. But in each case, by the twelfth genera- tion, the tumor incidence of the isogenic strain had returned to about the same rate as that of the original stock. Tumor incidence is a morphological mal- function and, as shown in this and other experiments, is under genetic control. The fact that all flies were not tumor bearing and the gradual return of the isogenic strains to the tumor incidence of the strains from which they were selected, reflects the accumulation of errors in the genome. The results of the experiment are in accord with the proposition proved above. Representation of the Ensemble 0/ Organisms by a Probability Distribution in H: piH, A) If we grant that perfect systems do not exist, the other side of the coin is, how imperfect may they be? This question was first discussed by Dancoff and QuASTLER (5) and their conclusion, which is known as Dancoff's principle, states that the amount of redundance is just that required to reduce the error 54 Hubert P. Yockey rate to a tolerable level. According to this principle, we may expect that errors will continue to accumulate in the genome of a given organism until at some point serious difficulty including death will occur. This will be reflected by some value of H, which we call //^, limited by viabihty. An argument for a lower limit H^^ has been given previously (6). Errors will accumulate in the genome but at the same time there is a favorable selection for those members of the ensemble which have low equivocation. This represents a certain reserve capacity to withstand the insults of existence. It may therefore be expected in general that the message entropy of the ensemble of organisms will be described by a probability distribution. This distribution can, perhaps, be calculated from first principles, at least for simple cases, when more is known about the storage and transfer of genetical information. Death of an organism is defined in different ways in various fields of biology. Permanent loss of reproductive power is the definition of death usually expressed or implied in bacteriology (7). This is the definition chosen in spite of the fact that there are many inteiTnediate stages between the active living cell and the dead cell. It is known that yeast cells which have lost the power to multiply may still be able to fennent (8). Zelle and Hollaender (7) have recently pointed out that attempts to explain the bactericidal effects of irradiation on the basis of one mechanism are unrealistic. In the case of animals the cessation of metabolism, not the loss of fertility, is the criterion of death. These criteria of death are not really different or antagonistic. Since loss of function is implied by loss of information content any experimentally convenient definition of lethality may be used to suit the problem at hand. The lower end of the distribu- tion in message entropy will therefore be determined by the specificity required by the environment. A communications analogy may clarify the notion further. Suppose we have a message, with redundance, which is sent through a communication channel with a small but finite noise level. The message contains instructions to perform some necessary task. A recording is made and the message is sent through again, and so forth. Eventually, depending on the noise level of the channel and the redundance in the message, it will be just barely intelligible. No further recordings can be made without loss of part of the required information content. The ensemble of recordings is analogous to the ensemble of organisms. It will be seen in either case that there is a distribution of information content among the elements of the ensemble. Individuality finds a place in the theory developed here in a very natural way. This feature corresponds more to reality (9) than theories which must explain non-uniform response as fluctuations. Besides the experiments of Burdette mentioned above it will suffice to note one other example of biological individuality. Consider the experiments of Schott (10, 11), Hetzer (12), Lambert (13), GowEN (14), discussed by Gowen (15), on Salmonella tvphimurium in mice and Salmonella gallinarum in fowl. The host population is exposed to the pathogen and the survivors are chosen for further breeding. The case for mice is typical. The survival ratio improved from 18 per cent to 93 per cent in six generations, but remained nearly constant after that. One hundred per cent survival was not achieved. The survival ratio is characteristic of the ensemble not of the Some Introductory Ideas Concerning the Application of Information Theory in Biology 55 individual. Gowen (15) also prepared six strains of mice by sibling malings for twenty or more generations. When survival was tested the survival ratios were 1, 14, 34, 63, 64, 83 and 88 per cent. These results again stress the importance of individuality as Gowen pointed out.* Point Mutations and Chromosome Aberrations We have now arrived, via our discussion, at territory familiar to the radiation biologist. This is the controversy over the role played by point mutations and chromosomal aberrations induced by deleterious agents such as x-rays. This subject has been ably discussed recently by Muller, Kaufmann, Giles, Carlson, SwANSON and Stadler, and by Kimball (16). The point of view of these authors varies. Kimball takes the stand with Lea (17) that the death of cells is due to chromosome aberrations which become effective at cell division. Swanson and Stadler point out that the two effects occur together and that a clear cut separation has not yet been accomplished. Muller points out some difficulties with the mutation by breakage interpretation. Russell (18) states that gross chromosomal aberrations, although they cause early death of embryos, are probably not an important radiation hazard to man. From the point of view of this article each of these effects is a way of intro- ducing disorganization in the genome. The point mutation mechanism is the biological analogue of the 'white noise' of the communications engineer. The other extreme is not found in communication engineering but involves a strong correlation between errors and is reflected as a loss of whole paragraphs or other gross mutilation of the message. Each of these extreme cases will be important in applications of information theory in biology. Unfortunately, the second case has not been studied mathematically and so it is not known how to calculate the equivocation it introduces. It is therefore necessary to proceed with the calculation of only the part of the equivocation which corresponds to point mutations. Since one of our objectives is to develop a fundamental theoretical treatment of radiation hazard to man, Russell's comment encourages one to think that this procedure is v/orthwhile. It should be remembered that equivocation from these two extreme conditions may have the same dependence on the deleterious influence. This is a point which requires further mathematical study. The Interaction of the Deleterious Agent nith DMA and the Decay of H According to the Watson and Crick model of DNA there seems to be no biochemical reason why there should be an interaction between nucleotide pairs. The biological requirements for protein specificity do not seem to demand an intersymbol influence (19). The matter is not closed, but the evidence favors regarding the interaction of a deleterious agent with a nucleotide pair to be of the first order. We have previously suggested that the action of ionizing radiation or other deleterious agent may be such that the nucleotide pair is altered in such a way that it mimes another symbol as far as protein synthesis is concerned (6). It * Individuality as an integral feature in biology has been emphasized recently by Rcxier J. Williams: in Biochemical IncUvidiiality, J. Wiley and Sons, New York, Chapman & Hall, London (1956). 56 Hubert P. Yockey may be thrown into an excited tautomeric form from which it recovers by relaxation. Possibly one can account for biological recovery by such a mechanism. The consideration of recovery is omitted from this paper for simplicity and we shall need only the notion expressed in the first sentence of this paragraph. In view of the above remarks we may write the following equation for the rate of change of /?,()) with A: idldX) p,ij) = -y,,(A) p,{j) + c,,(A) (4) The first terni represents the loss in nucleotides responsible for the {i,j) transition. The second term is due to the gain in nucleotides engaging in the (i,j) transition coming from other nucleotides altered by the deleterious agent. This can be brought into sharper focus by thinking of the binary case. Suppose q is the correct and p is the incorrect read-off probability. We are calculating the equivocation, or damage to the message, resulting from point errors. This means that, accordingly, a letter is not deleted but is read off either correctly or incorrectly. This letter switching process may continue until half the letters are correct and half are incorrect; at that point p = Ijl and q = 1/2. The infor- mation content vanishes. In the case of a four letter alphabet a letter which is acted upon and which may therefore change or may retain its original read-off character has an a priori probabiUty of 1/4 to remain or to become a correct letter. Thus the second term is required by the normalization condition. Equation (4) describes the effect of the interaction of the deleterious agent, say the x-ray dose, with the information bearing molecules in the cell. It corresponds to current views of reaction kinetics. Should it be discovered that some effect, for example, inter-symbol influence, should be taken into account then equation (4) may be altered suitably. The following argument would then still be cogent except that the new form of equation (4) would be used. Present experimental evidence substantiates equation (4) and we have no present justification for greater complication. In fact the /./A) and c-j{X) represent more detail than is available. Sum equation (4) over ally: 2 (d/dX) pij) = - 2 JM plj) + 1 cM (5) Since J J IPi(j)=l; I(dldX)p,(j) = 0 (6) j j o = -2^a)AO')+2c.>a) (7) 3 3 If the /,/A) and the c^/A) may be replaced by an average value J(X) and c(l), equation (7) becomes, for a four-letter alphabet: 0 = -J{X) + 4 cU) (8) c(X) = +yiX) (9) Equation (4) may be written as follows: (dldX) p,(j) - -7(A) p,{j) + i/(A) ( 1 0) Some Introductory Ideas Concerning the Application of Information Theory in Biology 57 Given (dldX) p^d) as some function of A, equation (3) may be regarded as a differential equation for //(A). This equation has a simple form if the y,v,(A) and the c,,(/l) may be replaced by their averages y(/l) and iJ{?>.). {dHldX) = log2 e 2 {p{i)J{X)[ p^ij) + I] -{-p{i)m [-^p,{j) + I] loge/;,(;) -\-p,{j)\og,pij){dicrA)p{i)] (11) {dHldXy= -J{X) log2 e lp{i)p,{j) loge/^(7) +i J(A) log2 e 2 p{i) loge PiCj) +log2 e 2 Piij) HePiij) {dldX) pii) (12) Substituting equation (2) in equation (12) and rearranging we have {dHldX) + J{X)H = J{X)H, + ya) 2 pii) \o^z Piij) i.j + lPi(j)iog,p,(j)(dld?i)p{i) (13) i.j The third term on the right of equation (13) is negligible for biological systems. To show this we must discuss first the method of calculating the {dldX) pU). By definition (3) the following relation holds: p{i)=lpii)p^ii)- (14) i Form the derivative with respect to A and substitute equation (4) : {dldX) p{i) = llpij) {dldX) p,{i) + pAi) (dIdX) p(j)] ( 1 5) j {djdX) pii) = - 2 y,, /.,(/•) pij) + 2 q. pij) + 2 pM) i^m p(j) ( 1 6) j i J The equations (16) are a set of differential equations for the p{i). They may be rearranged in the usual form: {dldX) pii) - 2 PjO) idldX) pij) = - 2 Jji PiU) pij) + 1 c,i pij) i 1 7) j j J We are interested in the conditions when the id/dX) pii) vanish. The condition is of course that the terms on the right of the equations (17) are all equal and that the detenninant of the coefficients of the idjdX) pii) be different from zero. Among the circumstances in which this will occur are those where all p^ii) = q and all Pi{k) = p ii j^ k). That is, all letters are equally probable and one kind of error is as likely as the other. In my paper in Part V the behavior of dH/dX under the much stronger conditions that the J^j and c^j vanish at A = 0 will be needed. Then, of course, providing that the determinant of the coefficients of the idldX) pii) be different from zero, all id I dX) pii) = 0. It may therefore be expected that except under most exceptional and special conditions the idfdX) pii) will be very small or will vanish. It can be further shown that for a nearly perfect system the coefficients of the idldX)pii) in equation (13) are small compared to one. Dancoff and 58 Hubert P. Yockey QuASTLER (5) have estimated the error rate per cell per generation to be some 10-1 tQ jo-2 times the spontaneous mutation rate per cell generation (10^* to 10~i^). Taking this to mean that Piii) = q^{\-p) and p^ij) = p ^ \Q~^ (i ^j) we see that /'^(01og2/7,(0 = +log2(l -p) ^-p = -10-6 p,(j) log^Piij) = -6 X 10-« log2 10 ^ -10-5 (18) Because of the discussion given above this term in equation (13) may be neglected. Equation (13) gives the value of {dHjdX) at the values of /?/;) corresponding to Hg^. Let these values be plij). dH = J{X)[H, -H, + 12 P{i) log2 p'm ( ' 9) dH The coefficient of /(A) will be a constant so that — dl will behave as a function Ha of X like J{X). This result will be needed in my article in Part V. REFERENCES 1 . J. D. Watson and F. H. C. Crick : Genetical implications of the structure of deoxyribose nucleic acid. Nature, Lo/tcl. Ill, 964-967 (1953). /. D. Watson and F. H. C. Crick: The structure of DNA. Cold Spr. Harb. Symp. Quant. Biol. 18, 123-131 (1953). J. D. Watson and F. H. C. Crick: Molecular structure of nucleic acids. A structure for deoxyribose nucleic acid. Nature, Lond. 171, 737-738 (1953). 2. G. Gamow: Possible mathematical relation between deoxyribonucleic acids and proteins. Biol. Mcdd., Kbh. 22, (3), 1-13 (1954). 3. C. Shannon and W. Weaver: The Mathematical Theory of Communication, University of Illinois Press, Urbana (1949). 4. W. J. Burdette: Incidence of tumors in isogenic strains. /. Nat. Cancer Inst. 12, 709-714(1952). 5. S. M. Dancoff and H. Quastler: The information content and error rate of living things. In: Information Theory in Biology, ed. by Henry Quastler, 263-373, University of Illinois Press, Urbana (1953). 6. H. P. Yockey: An application of information theory to the physics of tissue damage. Radiat. Res. 5, 146-155 (1956). 7. M. R. Zelle and A. Hollaender: Effects of radiation on bacteria. In: Radiation Biology, ed. by A. Hollaender, Vol. 2, chap. 10, McGraw-Hill, New York (1955). 8. O. Rahn: The order of death of organisms larger than bacteria. J. Gen. Physiol. 14, 315-337 (1930). 9. R. Pearl: On the distribution of differences in vitality among individuals. Amer. Nat. 61, 113-131 (1927). 10. R. G. Schott: The inheritance of resistance to Salmonella aertrycke in various strains of mice. Thesis, Iowa State College Library, 1-59 (1931). 11. R. G. Schott: The inheritance of resistance to Salmonella aertrycke in various strains of mice. Genetics 17, 203-229 (1932). 12. H. O. Hetzer: The genetic basis for resistance and susceptibility to Salmonella aertrycke in mice. Genetics 22, 264-283 (1937). Some Introductory Ideas Concerning the Application of Information Theory in Biology 59 13. W. V. Lambert: Genetic investigations of resistance and susceptibility to disease in laboratory animals. Rep. Agric. Res., Iowa Agric. Exp. Sla., 89-90 (1931); 91-92 (1932); 115(1933); 142-143(1934); 158-159(1935); 147-148(1936). 14. J. W. Gowen: Genetic investigations of resistance and susceptibility to disease in laboratory animals. Rep. Agric. Res., Iowa St. Coll. Agric. Exp. Sta., 158-159 (1937); 151-153 (1938); 156-160 (1939); 192-194 (1940); 171-172 (1941); 189-190 (1942); 178-182 (1943); 204-210 (1944); 278-283 (1945); 257-260 (1946); 230-232 (1947). 15. J. W. Gowen: Significance and utilization of animal individuality in disease research. /. Nat. Cancer Inst. 15, 555-570 (1954). 16. A. HoLLAENDER (cd.): Radiation Biology, McGraw-Hill, New York (1955). 17. D. E. Lea: Actions of Radiations on Living Cells, Cambridge University Press, Cambridge, England (1955). 18. W. L. Russell: Genetic effects in mammals. In: Radiation Biology, ed. by A. Hollaender, Chap. 12, McGraw-Hill, New York (1955). 19. M. Ycas: The protein text. This volume. PART II STORAGE AND TRANSFER OF INFORMATION A CENTRAL issue in modern biology, which touches in some degree all branches of that science, is the problem of species specificity and its relation to protein- specificity and synthesis. The subject can be approached from many points of view but the one adopted by the authors of the papers in Part II is to seek the solution in terms of the properties of a communication system The justification for considering, from this point of view, a phenomenon which looks, at first sight, to be purely biochemical lies in the recent discovery that protein specificity is expressed as an exact order of amino acid residues. If this is even substantially the case then it is germane to discuss such problems in these terms. In fact, a number of current papers on protein synthesis and specificity have recourse, at one point or another, to the language of information theory. Since the specificity of proteins is thought to be coded in the exact order of pairs of nucleotide bases in DNA, the relationship of DNA, RNA, and proteins can be considered from aspects which are mathematical rather than purely bio- chemical. Gamow was the first to notice these mathematical aspects. He and Ycas pursue in this part some of the issues which they reveal. The influence one hopes these considerations will have on the experimentalist is clear. Additional data on the amino-acid residue sequences and other structural data for a large number of proteins can be put to immediate practical use in solving for the protein code, and therefore in understanding more about protein synthesis. Unfortunately, mainly due to the lack of sufficient protein text, few definite answers can be given. But it is possible to eliminate some past errors and to phrase the question in a sharper fashion than before. The notion that an abstract quantity such as information is stored in the genetic material and is transferred to proteins during their synthesis raises immediate questions as to how this is done, how much is transferred, and how this quantity is aff'ected by changing experimental conditions. These questions are attacked from diff"erent analytical and experimental points of view by the papers by Augenstine, by Mahler, Walter, Bulbenko and Allmann, and by Koch and by Glinos. The information theoretic properties of communication systems of particular concern to the papers in this part are the coding problem, the representation theorem, and redundance. Each paper deals with issues of its own but in terms of these ideas to a greater or lesser degree. It is in this way, among others, that information theory may grow to be as useful to the biologist as thermodynamics is to the chemist, whether his subject is clearly one in communication as is that of Frishkopf and Rosenblith or somewhat less clearly that of protein specificity. H. P. Y. 61 THE CRYPTOGRAPHIC APPROACH TO THE PROBLEM OF PROTEIN SYNTHESIS George Gamow and Martynas Ycas University of Colorado and University of New York Abstract — The Watson and Crick suggestion concerning the role of DNA in replication, mutation, and protein synthesis requires a coding between the four-letter DNA alphabet and the twenty-letter protein alphabet. An attempt has been made to discover this code by crypto- graphic methods. Various schemes have been worked out but no success obtained at this writing. There is hope that as the number of protein sequences increases this problem will be solved. Speaking about information storage and transfer in a living cell, one always likes to compare the cell with a large factory. The cell nucleus is the manager's office, directing the work of the factory, and the chromosomes are the file cabinets in which all blue prints and production plans are stored. The cytoplasm is the plant itself with the workers and machinery carrying out the actual production ; those are, of course, the enzymes catalyzing various biochemical reactions. If something goes wrong with the information stored in the chromosome, the corresponding enzyme will also do a wrong thing. Consider, for example, an enzyme which produces the pigment necessary for color vision. If the particular section of chromosome carrying the directions for producing that pigment is defective, the enzyme will not get the correct instructions, and will not produce the right type of pigment. As a result, the individual will be color blind. The materials of chromosomes and of enzymes are chemically different, except that in both cases we deal with long molecular chains formed by the repetition of a comparatively small number of different units. DNA (deoxyribo- nucleic acid), forming the chromosomes, is a sequence o^ four different units or 'bases': namely, adenine, thymine, guanine, and cytosine. For sake of picturesque presentation, we may associate them with four suits of cards: spades, clubs, diamonds and hearts. Each DNA molecule is equivalent to a sequence of cards many thousand units long, and the way in which different suits follow each other contains, in code form, the instructions to the original cell (fertilized ovum) and its descendants to develop into a rosebush, a skunk, or a man. The first question is this. How is information which is carried by DNA molecules of the chromosomes duplicated when the cell goes through the process of division? An answer can be given on the basis of the model of DNA proposed about three years ago by J. Watson and F. Crick (1). They started with the fact, first noticed by E. Chargaff (2), that the number of adenines in any given DNA molecule is always equal to the number of thymines, while the number of guanines is always equal to the number of cytosines (3). In the playing card analogy there are as many spades as there are clubs, and as many diamonds as hearts. This suggests that we deal here with a double-stranded 63 64 George Gamow and Martynas YCas ^\. / / / \ ^\ -U 1 \ I / \ ^o_: u -J. / -u c *c < ♦ o X-U-X -^ u=u. I I K r/ ~o 0=u -u — X \-/ \ /^" \ -u — ^ ■. -U X ./ -o \ / \ ^-\ /■ O"^ / X — u o / X I / /r -u ./ -u- -U — X ° I I — u \ 4> c 'c o 3 c m o ■«-> U The Cryptographic Approach to the Problem of Protein Synthesis 65 sequence in which red and the black cards are paired together. A heart is always paired with a diainond (and vice versa), while a spade is always paired with a club (and vice versa). The fact that DNA molecules also contain one sugar (ribose) and one phosphate for each 'base' suggests a molecular model similar to a rope ladder. The vertical ropes on both sides are formed by 'sugar- phosphate- sugar-phosphate-' sequences, while the paired bases form rigid horizontal steps attached to sugars on both sides. The reason why the above- mentioned pairing of bases takes place is two-fold. Cytosine and thymine (hearts and clubs) are 'pyrimidines', being formed by a single C — N — ring with different atomic groups attached to them. Adenine and guanine (spades and diamonds) are 'purines', and contain in their structure two connected rings, one with six atoms, and the other with five. The chain shown in Fig. 1 is a sequence of sugars and phosphates. To each sugar is attached a 'base', and in tliis section of the molecule you see four different bases. Two of them (hearts and clubs) are short, and two others (spades and diamonds) are long. Now, in order to run the second strand beside it in the parallel way, we should attach short bases to long ones, and long bases to short ones. Of course, in the playing card analogy again, one could also join a heart to a spade and a club to a diamond. But this is excluded because in these cases hydrogen atoms will be in the wrong places to form proper hydrogen bonds between these two bases. The evidence supplied by an x-ray diffraction pattern indicates in addition that the DNA molecule has a helical shape, being twisted around its central axis by 36° each step. Thus, it makes a complete turn each 10 steps. The Watson and Crick (4) theory of dupHcation of DNA molecules proceeds as follows. When the cell is ready to divide, there appears a large number of free nucleotides in the nucleoplasm surrounding the chromosomes. A nucleotide is defined as one of the four bases with a sugar and a phosphate attached to it. At that time the double stranded DNA molecule splits into two single strands along its main axis, and each strand is regenerated by catching the corresponding free nucleotides from the surrounding medium. Thus, each heart separated by splitting from its diamond gets another diamond from the solution, and each diamond gets another heart. As the results, we get two new double stranded DNA molecules, each identical with the original one. Once in a while a mistake may be made in this duplication process, and we call it a mutation. So much for the structure and functioning of DNA molecules. Now we come to the problem of information transfer from the chromosomes to the enzymes. How does the sequence of bases (card suits) in DNA determine the structure of the enzyme? Enzymes are proteins, and are formed by long sequences of twenty different chemical groups known as amino acids. It is well known that there are as many as twenty-four or twenty-five amino acids, but, as Dr Yeas tells us in more detail in the next paper, one can show that the extra ones in the original protein synthesis are modifications of the original twenty which take place after the protein molecule is synthesized. Thus, for example, 'proline' is an original amino acid used in protein synthesis, whereas 'hydroxyproline' is its postsynthetic modification. Since we symbolized four bases of nucleic acid molecule by four playing card suits, it is reasonable to symbolize the twenty basic amino acids, which have complicated chemical 66 George Gamow and Martynas Ycas names, by twenty letters of a (reduced) English alphabet. Thus, one protein molecule may look like: . . .arreducesugarreducesug. . . and another like: . . . akeacoloruisionpigmentma . . . Just to give an example of how the sequence of amino acids in protein molecules may affect their biochemical activity, we will give the example of two closely related hormones: oxytocine and vasopressin. Both are formed by a sequence of only nine amino acids: Oxytocine — Cys-Tyr-Z/ew-GIun-Aspn-Cys-Pro-Lew-Gly Vasopressin — Cys-Tyr-PAe-Glun-Aspn-Cys-Pro-vlr^-Gly The two sequences are identical except for the substitutions in the third and eighth place. However, their functions are rather different. Oxytocine has the property of causing the contraction of the uterus in the process of childbirth. If you inject it into the blood of a cow, even if the cow is not pregnant it will go through all motions it would go through if a calf were to be born. Vaso- pressin, on the other hand, has rather different properties: it contracts the blood vessels and causes increased blood pressure. Thus, simply by changing two amino acids out of nine, the action of the hormone is completely changed. Whereas replacement of some amino acids in a protein may completely change its biological function, there also exist replacements which distinguish the same protein taken from different species of animals. Thus, for example, insulin A, which is formed by a sequence of amino acids with twenty-one members, differs for cattle and swine in the eighth and tenth place. Human insulin, which has not yet been analyzed, possibly differs slightly from that extracted from cattle and swine. Nevertheless, the latter are successfully used on human patients. Since there must exist a definite relation between the sequence of bases in nucleic acid and the sequence of amino acids in proteins, we can ask ourselves what this relation is. Here we have to return to our analogy of a factory. The v/orkers from the factory do not walk into the manager's of^ce to find out what to do, and the manager also does not go to the plant to instruct workers per- sonally. There are people, called foremen, who get the information from the manager's ofiice and tell the workers. In the cell the role of foreman is carried out by RNA molecules (ribonucleic acid) which are, presumably, very similar to the molecules of DNA. They are different only in that one oxygen atom is missing in each sugar of DNA, and there is a slight change in one of the four bases, which in RNA is called urosil instead of thymine. RNA is presumably synthesized by DNA inside the nucleus and receives the set of instructions carried by DNA. Then it passes out into the cytoplasm, and is incorporated into the so-called microsomes, i.e. foremen's offices, where the synthesis of proteins takes place. We do not yet have a model of the RNA molecule. It seems, however, that in this case the pairing rules of adenine to thymine (urosil), and guanine to cytosine do not hold, which suggests that RNA molecules are single-stranded. The Cryptographic Approach to the Problem of Protein Synthesis 67 Since RNA serves as an intermediary between DNA and proteins, we have here two problems. First, how is RNA formed by DNA ? Second, how are proteins synthesized by RNA? The first problem may turn out not to be very difficult because of the close similarity between the two molecules. For example, RNA may be a non-regenerated half of DNA with small changes in sugars and in one of the bases. It may be that the absence of the oxygen atom in RNA's sugar is responsible for the failure to form a double-stranded configuration. However, we still do not know the answer to this question. The second problem concerning the synthesis of proteins by RNA mole- cules presents more challenge to the imagination. How can a sequence formed by four different units (four bases) be translated in a unique way into a sequence formed by twenty units (twenty amino acids)? Here is a possibility which seems to us to be very likely. Suppose one plays a game of poker in which only three cards are dealt, and pays attention only to the suit of the card. How many different hands will one have? Well, one can have a 'flush', i.e. three cards of the same suit. There are four different flushes: three hearts, three spades, etc. Then one can have a 'pair', i.e. two cards of the same kind, and one different. How many of those are there? One has four choices for the suit of the pair, and three choices for the third card. Thus, there are altogether twelve possibilities. The poorest hand will be a 'bust', i.e. three different suits. There are four different busts: no hearts, no diamonds, etc. We have altogether twenty different possibilities. This 'magic number' 20 is just the number of amino acids participating in the primary process of protein synthesis. We may imagine that each amino acid in the synthesized protein is determined by a triplet of bases in the RNA template. Since the distances between neighboring amino acids in the extended polypeptide chain are equal to the distances of neighboring bases in the poly- nucleotide chain (both being equal to 37 A), it was at first natural to suppose that the correlation between the two chains looks in a way shown in Fig. 2, RNA-Template where individual bases are shown by circles and the amino acids by triangles. This represents the so-called over-lapping code in which the neighboring amino acids have in common two bases in the RNA template. If the transfer of information from nucleic acid to protein is carried out according to such an overlapping code, there must exist a definite inter-symbol correlation between the amino acids constituting protein molecules. Thus, for example, if a certain amino acid is determined by two adenines and some other base, its neighbors will be preferably amino acids which also contain adenine in their template transcript. In order to see whether or not such a correlation between the 68 George Gamow and Martynas Ycas neighbors really exists in the known protein sequences, it is necessary to test all possible assignments between the twenty amino acids and the twenty possible base triplets. The number of all possible assignments of that type is 20! = 3.10^^, Since 3.10^^ represents the age of our universe (5 bilhon years) expressed in seconds, the straightforward test of that kind would require quite a consider- able time even if we could test one assignment each second ! However, as it often happens in cryptographic problems, one can sometimes find parts of the message which reduce quite considerably the amount of necessary work. Thus the code messages sent by German spies during the war were likely to contain the combinations of letters corresponding to various possible ports of embark- ation of American expeditionary troops. The same happens in protein sequence. For example, the adrenocorticotropin molecule contains the sequence: — Lys — Lys — Arg — Arg— Pro — Val — Lys — Val — In this sequence there are two identical amino acids in succession followed by another pair of identical ones. In the English language there are not many words having such a property. (Tennessee is one of the rare examples!) Then lys repeats again three steps later, and has identical neighbors (val) on both sides. These facts simplify the problem to such an extent that, instead of spending five billion years, it was possible to find a single assignment between the amino acids in the above sequence, and the base triplets in the course of an afternoon. At first it was thought the problem had been solved, but, when one tried to extend these assignments to the other parts of the ACTH molecule and to the other known protein sequences, one was led to direct contra- dictions. In the course of subsequent decoding work, other examples leading to similar contradictions were found, and it became clear that the thing just will not work. In fact, as Dr Yeas discusses in the following article, it seems that there is no correlation between the neighboring amino acids whatsoever. This negative result can only mean that the original hypothesis represented in Fig. 2 was incorrect, and that in the process of protein synthesis the nucleic acid molecule is not present in its extended form. If, as seems to be true, we deal here with a "non-overlapping code" in which each amino acid is determined by an individual base triplet of its own (Fig. 3), we are forced to assume that RNA-Template Fig. 3. the RNA molecule is shrunk by a factor of three. We can imagine, for example, that during the process of protein synthesis the RNA molecule has the shape of a spiral as shown in Fig. 4. Closely connected with the problem of a non-overlapping code is the problem The Cryptographic Approach to the Problem of Protein Synthesis 69 of "punctuation". Indeed, a sequence of bases can be broken into a set of non-oveiiapping triplets in three different ways depending upon the base with which we start. The three dilTerent readings of tiie same template can be des- cribed mathematically as 3n, 3n/l, and 3n/2 (3n/3 being the same as 3n). A| 3.6 A As was suggested by Dr Barbara Law, three possible readings of the same RNA template may explain an interesting regularity first noticed by Dr Martynas Yeas. He observed about two years ago that, in a case of seven proteins for which the sequences of amino acids were known, the total number of amino acids in the protein molecule was a multiple of three : nine amino acids in oxytocine and vasopressin, twenty-one in insulin A, thirty in insulin B, thirty-nine in ACTH, 126 in ribonuclease, etc. This could be explained if one assumes that each RNA template synthesizes the proteins in all three possible vv'ays, and that these three different readings are afterwards united in one linear sequence. If this were true, there must exist a cryptographic correlation between the first, second, and third "thirds" of each protein molecule. One thinks of how such a correlation could be checked, but it seems to be very difficult indeed. Recently, though, the existence of such a correlation became rather doubtful, since two protein sequences published recently contain 29 and 124 amino acids. In summing up, we should say that the problem of finding the nature of the correlation between polynucleotide chains of nucleic acids, and the polypeptide chains of the proteins is still unsolved, although various methods for establishing such a correlation have been worked out. We may hope, however, that with the increased number of known protein sequences, this problem will be solved in one way or another. REFERENCES 1. J. D. Watson and F. H. C. Crick: Molecular structure of nucleic acids. Nature, Lond. 171, 737-738 (1953). 2. E. Chargaff: for reference see S. Zamenhof, G. Brawerman, and E. Chargaff: On the deoxypentose nucleic acids from several micro-organisms. Biochim. Biophys. Acta. 9, 402-405 (1952). 3. G. R. Wyatt: Nucleic acids of some insect viruses. /. C^'//. P/(V.v/o/. 36, 201-205 (1952). 4. J. D. Watson and F. H. C. Crick: Genetical implications of the structure of deoxyri- bose nucleic acid. Nature, Lond. 171, 964-967 (1953). 6 THE PROTEIN TEXT Martynas Ycas Department of Microbiology, State University of New York Upstate Medical Center, Syracuse, New York And strange to tell, among that Earthen Lot Some could articulate, while others not: And suddenly one more impatient cried — 'Who is the Potter, pray, and who the Pot ?' The Book of Pots Abstract — The sequence of residues in proteins, regarded as a text written in a twenty symbol alphabet, is examined. The following tentative conclusions are drawn: 1. Twenty amino acids are distinguished by the protein-forming mechanism. Super- numerary amino acids arise from the regular twenty by secondary modification of protein- bound residues. 2. Each residue in the protein has a separate genetic representation. 3. There is no intersymbol correlation between adjacent residues. 4. Natural selection is not the only factor determining the frequency of occurrence of the various kinds of residues. It is suggested that the method of encoding protein sequence information in nucleic acid imposes differences in frequency of occurrence on the different kinds of residues. 5. Peptide chains are not multiples of some fixed number of residues. The encoding and transfer of genetic (DNA) information to RNA and protein is discussed, as well as the problem of the independent reproduction of RNA viruses. While the data set certain limits on the possible ways of encoding and transferring information, they are not sufficient for a unique solution of these problems. Ribonucleic acid of Tobacco Mosaic Virus (TMV) has been shown to deter- mine the sequence of amino acid residues in the protein of the virus (1, 2, 3). It seems logical therefore to believe that the sequence of other proteins is also determined by RNA.* Since RNA is essentially a linear sequence of four kinds of nucleotides, while proteins are linear sequences of about twenty kinds of amino acid residues, the RNA molecule can be regarded as a text, written in a four-symbol alphabet, which encodes another text, the protein, written with about twenty symbols. * The following abbreviations will be employed. RNA — ribonucleic acid; DNA — deoxy- ribonucleic acid; Ad — adenylic acid; Gu — guanylic acid; Cy — cytidylic acid; Ur — uridylic acid; ala — alanine; arg — arginine; asp — aspartic acid ; aspn — asparagine; asx — asparticacid or asparagine; cys — cysteine; glu — glutamic acid; glun — glutamine; glx — glutamic acid or glutamine; gly — glycine; his — histidine; ileu — isoleucine; leu — leucine; lys — lysine; met — methionine; phe — phenylalanine; pro — proline; ser — serine; thr — threonine; try — trypto- phan; tyr — tyrosine; val — valine; Hlys — hydroxylysine ; Hpro — hydroxyproline; serP — phosphoserine. Peptides are written with the amino group to the left, the symbols being connected by a dash ( — ). The sign (*) signifies a terminal residue. Sequences considered uncertain are in parentheses ( ). Symbols in parentheses, with commas between (ala, gly) mean that the sequence is not known. 70 The Protein Text 71 Several attempts, none completely convincing, have been made to determine the coding system employed (4, 5, 6, 7). Cryptography must be based on a study of texts, and 1 shall therefore attempt an examination of protein molecules from this point of view. The following aspects of protein structure will be examined : 1. The number of kinds of amino acids which occur in proteins. 2. The effect of mutations on amino acid sequence. 3. Whether intersymbol correlations exist between adjacent residues. 4. The frequency of occurrence of the various amino acid residues. 5. Whether any restrictions exist on the length of peptide chains. After considering the empirical evidence, I shall indicate its bearing on the problem of encoding protein sequence information into the RNA molecule. I. THE NUMBER OF AMINO ACIDS OCCURRING IN PROTEINS In previous studies (6, 7) it has been assumed that proteins are composed of exactly twenty different kinds of residues. Since in fact more than twenty kinds of residues occur in proteins, the assumption requires some justification. All organisms, from viruses to mammals, use the same building blocks for their proteins. With minor qualifications this is also true of the nucleic acids, but not true of the third major class of biologically-occurring high polymers, the polysaccharides. The amino acids which invariably occur in all organisms and virtually all proteins are the following: ala, arg, asp, aspn, cys, glu, glun, gly, his, ileu, leu, lys, met, phe, pro, ser, thr, try, tyr, val. The number in this list is exactly twenty. It will be noted that I omit cystine from this list. Because of its structure, cystine corresponds to two residues. The structure of insulin (8) shows that one cystinyl residue can occupy non-adjacent positions in a peptide chain or even participate in two different chains. Cystine is best regarded as an oxidation product of cysteine, formed after incorporation of the cysteinyl residue into the peptide chain. This view is supported by the recent discovery of an enzyme which reversibly catalyzes the reaction 2 cysteinyl :^ cystinyl when these residues are protein bound (9). Another example of such a reaction may be the cyclic oxidation and reduction of protein SH groups during the various stages of cell division (10). In addition to the above twenty, other alpha amino acids occur in nature. Some of these, such as homocysteine, citruline and ornithine are well known biochemical intermediates but do not occur in proteins. It is clear that the number of amino acids which occur in proteins is limited by an inability to incorporate, rather than make, amino acids. Hydroxyglutamic acid and norleucine, previously believed to be protein constituents, have been shown not to exist as natural products (11). Alpha amino-adipic acid has been isolated from an impure protein hydrolyzate, but it has not been demonstrated that it is a protein constituent in the same way as other amino acids (12). Diamino pimelic acid, commonly occurring in bacteria, appears to be associated with the polysaccharide material of the cell wall (13, 14). Nevertheless, there are amino acids, other than the twenty enumerated, 72 Martynas Ycas which certainly occur in proteins. These include hydroxylysine and hydro- xyprohne (in collagen), phosphoserine (in a number of different proteins (15)), thyroxine (in thyroglobulin) and tyrosine — O — sulphate (in fibrinogen) (16). The distribution of these amino acids is different from the regular twenty. Whereas the twenty amino acids occur in virtually all proteins, the super- numerary ones have an erratic distribution, being confined to one or to a few. The suggestion was first made by Crick, that the supernumerary amino acids are the result of modifications of some of the regularly occurring amino acids after these have been incorporated into a peptide chain. The biochemical evidence for this is as follows. When one of the twenty regularly occurring amino acids is presented labeled to an organism, it is rapidly incorporated into protein and most of the label is found in the corresponding residue. It should be noted that glutamine and glutamic acid are separately incorporated and do not arise one from another by addition or subtraction of amide groups after incorporation (17). (A similar demonstration for the analogous case of asparagine and aspartic acid is still lacking.) Clearly, therefore, these amino acids are the precursors of the corresponding protein-bound residues. The supernumerary amino acids behave differently. Thus lysine is the precursor of hydroxylysine (18), but C^* or tritium-labeled hydroxylysine is not incorporated into collagen (19). Similarly, proline is the precursor of hydroxyprohne, but proline is a much better precursor of the hydroxyprolyl of collagen than is hydroxyprohne itself (20, 21). These amino acids, then, are not incorporated as such, but presumably are formed by oxidation of protein-bound proline and lysine. Phosphoserine likewise is formed by phos- phorylation of protein-bound serine (22). Thyroxine is apparently formed from the tyrosine residues of thyroglobuhn (23). There is no information at present on the metabolism of tyrosine — O — sulfate. Since not all appropriate residues are secondarily modified, this inter- pretation imphes that the enzymes catalyzing such conversions show specificity for sequence in the protein. At least one enzyme is known which shows such specificity. Prostatic phosphatase dephosphorylates phosphoserine in the sequence asx-serP-glx-ileu-ala, but not in glx-serP-ala (24). It is therefore suggestive of some enzyme specificity that hydroxyprohne in collagen occurs mainly, if not exclusively, before glycine (25) (Table IV). Other amino acids, as shown later, shovv' no such neighbor preferences. The region determining whether proline is to be oxidized or not probably includes more than three residues, as indicated by the isolation from collagen of the tripeptides ala- pro-gly; ala-Hpro-gly and ser-pro-gly; ser-Hpro-gly (Table IV). The biochemical evidence thus appears to indicate that the protein-forming mechanism selects exactly twenty different kinds of amino acids, and that the supernumerary ones arise by secondary modification of protein-bound residues. A possible cause for error in this conclusion should be noted. It is virtually certain that amino acids are not incorporated as such, but in the form of some sort of activated derivative. If the same amino acid were to form more than one derivative, the number of items to be selected would of course exceed twenty. There is no evidence for this at present, and only further advances in biochemistry can decide whether this is the case. Tlie Protein Text 73 II. GENETIC EFFECTS ON PROTEINS There is an increasing body of evidence indicating that tiie details of protein structure are genetically determined. A study of the effect of mutations on proteins should therefore tell us something both about the nature of mutations and the protein forming mechanism. Known cases of genetic effects on proteins are listed below. 1. In man hemoglobin occurs in several electrophoretically distinguishable forms, the presence of each being apparently controlled by alleles of a single gene (26). Hemoglobin C differs significantly in amino acid composition from hemoglobin A (27). Hemoglobin A and S have been degraded in a controlled fashion with trypsin and the resulting peptides separated. The difference between these hemoglobins is apparently confined to a short section of the molecule (28). 2. Two electrophoretically different hemoglobins occur in sheep. Their presence is determined by alleles of a single gene (29). 3. Two forms of lactoglobulin occur in cow's milk, and like the hemo- globins are determined by different alleles of one gene. Crystallographic investigations indicate unit cells of the same size, but there are very slight differences in the diffraction pattern, which the investigators attribute, possibly, to the substitution of a few amino acid residues by others (30). 4. Mutants of Neurospora and Escherichia co/i produce abnormally heat- labile forms of tyrosinase (31) and a panthothenic acid synthesizing enzyme (32), respectively. It is clear that a change in the proteins has occurred, but unfor- tunately there is no further information on its physico-chemical nature. The genetic evidence indicates that there is no interaction between alleles controlling the synthesis of different variants of one protein. If both alleles are present, both types of protein are formed. A possible exception should be noted. The N-terminal groups of wheat gliadin are reported to be phe, of rye gliadin phe and glx, but unexpectedly the ghadin of wheat x rye hybrids was found to have no amino or carboxyl terminal ends, indicating, possibly, a cyclic protein (33). This case obviously needs further study*. The evidence cited above shov/s that the properties of proteins are gene- determined, but it does not indicate clearly what these properties are. More detailed information is available on this point from a comparison of homo- logous proteins of related species, if it is assumed, as is usually done, that species differences are the result of gene mutations. Available evidence on amino acid sequence of homologous proteins is * There is considerable confusion as to the N-terminal residues of wheat gliadin. Fraenkel- CoNRAT (51) misquotes Deich and Soreni (33) as stating that the N-terminal residues are phenylalanine and histidine, apparently because of a misunderstanding in Chemical Abstracts (138). KoROS, whose paper I was able to consult only in abstract (139), reports histidine as N-terminal. Ramachandran and McConnell (140), working with wheat gliadin but failing to specify the species, also find histidine. Deutsch (the same as Deich quoted above, the differ- ence in spelling being due to transliteration from the Cyrillic) reports that gliadin from Triticiim durum and Triticum milgare has N-terminal phenylalanine (141). This is misquoted as tyrosine, and tyrosine and glutamic acid, respectively, by Ramachandran and McConnell (140). The original paper of Deutsch (141) was also unavailable to me. 74 Martynas Ycas collected in Table I. Mutations (as inferred from differences between homo- logous proteins) do not produce a general scrambling of protein sequence, but a replacement of one or more residues, leaving the rest of the sequence unchanged. Since homologous proteins can differ by a one residue replacement, it is clear that individual residues, rather than groups of residues, are represented in the genetic material. Table I. Sequences in Homologous Proteins from Different Species Protein Species Insulin (34) . cys-thr-ser-ileu-cys . . cys-ala-ser-val-cys . . . cys-ala-gly-val-cys . . . cys-thr-gly-ileu-cys . . cys-thr-ser-ileu-cys . Pig Cattle Sheep Horse Whale Myoglobin (35) *val . . . *val . . . *giy . . . *gly . . . Finback whale Sperm whale Horse Seal {Phoca vitulina) Protamine (36) (Composition, not sequence) glyasefaalaovaliileui glyaseraalagvalaileuo Salmo irhleus Salnio trutta Serum albumin (37, 38) *asp-ala .... leu* *asp-thr .... ala* Man Cattle • Cytochrome c (39) . . . cys-ala-glun . . . . . . cys-ser-glun . . . Horse, Cattle, Pig, Salmon Chicken Vasopressin (40) . . . pro-arg-gly-NHo* . . . pro-lys-gly-NHa* Cattle Pig Protein The Protein Text Species 75 Hemoglobin (41) *val-leu . *val-gly . *val-glun *val-leii . *val-gly . *val-asx . *val-leu . *mct-gly . *val-Ieu . *val-ser . , *val-asx . *val-leu . *val-gly . *val-leu . Horse, Pig Dog Cattle, Goat, Sheep Guinea pig Rabbit, Snake Chicken Gliadin (33) *phe . . . *phe . . . Wheat *phe . . . *glx . . . Rye Fibrinogen (42) *tyr . . . *ala . . . Man *tyr . . . *glx . . . Cattle ACTH (43, 44, 45) . . . pro-ala-gly-glu . . . . . . pro-gly-ala-glu . . . Sheep Pig . . . glu-ala-ser-glu . . . . . . glu-leu-ala-glu . . . Sheep Pig Hypertensive p-ptide (46, 47) . . . val . . . Cattle . . . ileu . . . Horse 76 Martynas Ycas Protein Species Virus (48) . . . thr-ser-gly-pro-ala-thr* TMV (M, YA strains) . . . thr(thr,ala)pro-ala-thr* TMV (HR strains) It is possible that a mutation may suppress an amino acid determining site altogether. This is indicated by the tentative finding of Akabori (quoted in (41)), that the 'B' chain offish insuHn has the sequence . . . pro-lys*, as compared with the sequence . , . pro-lys-ala* in cattle. In some cases (ACTH, TMV), two adjacent replacements differentiate one homologous protein from another. It is not probable that this is due to two independent but adjacent mutations, but rather that a single mutational event has affected two residue-determining sites. Such a view is made plausible by the work of Benzer (49). He has shown that mutations in bacteriophage involve small sections of DNA, of molecular dimensions, but that these sections can be of diflferent lengths. Presumably the length of the mutated section deter- mines the number of residues changed in the protein. It is perhaps not too sanguine to hope that eventually it may become possible to measure crossover values in terms of distance in residues along a protein chain, and thus obtain an estimate of the number of bases in DNA determining a single residue selecting site. The present difficulties of such an approach are of course obvious (50). It would be of interest to determine if there are any restrictions on the replacement process. Restrictions might be expected on the following grounds. More than one nucleotide must determine an amino acid site. If the process of mutation were predominantly to change some, but not all nucleotides determining a site, then obviously not all sites would be interconvertible in one step. A study of any such restrictions would be of great value, since their nature would depend on the coding principle and could be used to infer the latter. Table II. Replacements Inferred from Table I and their Frequency of Occurrence Occurrence 3 2 2 2 2 2 Replacement val <->ileu ala <->thr ala ■<—> ser ala <->gly ala <-> leu ser <-^gly ala <-^gIx val 4^gly val <->met phe <-^glx slur i<-^ asx arg <->lys The Protein Text 77 Known replacements in homologous proteins are collected in Table II. In the small sample we have (nineteen replacements), half recur twice or more, suggesting strongly that the process, as observed, is not a random one. Unfor- tunately, the sample is not unbiased. Certain replacements arc lethal or semi- lethal (hemoglobin S, for example), and are, without doubt, selected against. What we actually observe has therefore passed through the sieve of selection. The direct genetic approach to this problem is tedious, because of the difficulty of determining the phenotype (the amino acid sequence), and rapid progress is scarcely to be expected. A much larger body of data on homologous proteins may, however, enable us to reach a decision on whether the replacement process is intrinsically restricted or not. An additional point emerges from a consideration of such protein mole- cules as consist of more than one chain (Table III). It will be noted that there Table III. Terminal Residues of Proteins having more than one Peptide Chain (The exact number of chains is not indicated.) Protein N-terminal C-terminal Reference Cytochrome c fhis jhis (51) Growth hormone [ala phel phej (51) Triosephosphate- dehydrogenase fval Ival met \ met (51) Collagen giy ala J (51) Gliadin (wheat) phe (33) Glladin (rye) fphe Iglx (33) ^ lactoglobulin peu |leu ileu \ ileu (51) Fibrinogen (man) ftyr lala (51) Fibrinogen (cattle) ftyr glx (51) Hemoglobin (horse) fval jval (41) Hemoglobin (cattle) fval \met (41) 78 Martynas Ycas is a strong tendency for the terminal residues of such proteins to be identical. This is certainly not due to the chains being identical in all cases, since the hemoglobins, for example, do differ in the penultimate positions (Table I). Rather it appears to indicate that multi-chain proteins arise by reduplication of genetic material, so that the several chains start out by being identical, but gradually diverge in the course of evolution in the same way as homo- logous proteins of different species. This hypothesis, as applied to the hemo- globins and insulin, has been previously discussed (6). Determinations of the residue sequence along different chains of one protein may therefore throw additional light on the replacement process. Table I shows that the process by which replacements become established is very slow. Elucidation of the sequence of homologous proteins may therefore make it possible to determine phylogenetic relations between large groups such as phyla, which cannot now be certainly determined from morphological and embryological evidence. III. CORRELATIONS BETWEEN ADJACENT RESIDUES Are there any forbidden combinations of adjacent residues? An examination of the sequence of residues in proteins (Table IV) could provide an answer to this question. <0 Q.Q. too 3>- CO _iq: men >-_» _i_i — << < < oo oo X 3(oHuJOttQ:>-a:_i ui>-iiJXD:uxir>4 ALA ARG ASP ASPN CYS GLU GLUN GLY H IS ILEU LEU LYS MET PHE PRO SER THR TRY TYR VAL •• •• • • • •• • 9 •• •• •• • • • •• •• • • • •• • • •• •• •• •• • •• #9 •• • •• •o •• •• •• • •• • •• • • — •• • •• •• • • •• •• • • • •• •• •• •• •• • • •• • •• •• •• •• •• • • • • « • •e •• •• •• •• •• •• • •• • •• •• • • •• • • •• • •• • • 9 • •• •• •• •• •• •• • • • • Fig. I. Dipeptlde sequences now known to occur in proteins, compiled from Table IV. The N-terminal amino acids are plotted in the rows, the C-terminal in the columns. There are of course 400 possible pairs of the twenty amino acids. The known protein sequences in Table IV have been broken down in the following way. A sequence, say, of ala-arg-gly is broken down into the dipeptides ala-arg, arg-gly, and the appropriate cells in Fig. 1 are then filled, the N-terminal residues being represented by the rows, the C-terminal by the columns. Using all the data available in Table IV, Fig. 1 shows that somewhat more than half of all possible dipeptide combinations are known to occur. The question The Protein Text 79 Table IV. List of Known Sequences in Proteins Actin (52) . . . his-ileu-phe* Adrenocorticotropin (45) *ser-tyr-ser-met-glu-his-phe-arg-try-gly-lys-pro-val-gly-lys-lys-arg-arg-pro-val-lys-val-tyr-pro- ala-gly-glu-asp-asp-glu-ala-ser-glu-ala-phe-pro-leu-glu-phe* Carboxypeptidase (53) *aspn-ser; ser-thr a Casein (54, 55) serP-glx ; *lys-ieu-val-ala-glx-asx Chymotrypsinogen (56, 57) leu-ser-arg-ileu-val ; aspn-ser-gly-(glun-ala) Clupein (58, 59) *pro-ser-arg; ser-ala-arg-arg* ; arg-arg-arg-arg; Collagen (60, 61) ala-Hpio-gly; ala-pro-gly; glx-arg; glx-Hpro-gly; gly-asx-gly; gly-glx; gly-pro-ala; gly-pro-glx; gly-pro-gly; gly-pro-Hpro ; ser-Hpro-gly ; ser-pro-gly; ala-gly-ala; gly-gly; ser-gly; thr-gly; ala-asx; asx-asx; asx-glx; asx-gly; glx-ala; glx-glx; glx-gly; glx-gly-gly; glx-met; glx-phe; ser-asx; val-glx; ala-arg; arg-gly-gly; arg-val-gly; ser-arg; val-arg; ala-lys; asx-arg; lys-gly; pro-ser; pro-thr; ser-ala; thr-ala; lys-pro-gly ; leu-ala ; ala-ala-gly ; Cytochrome c (39) . . . val-glun-lys-cys-ala-glun-cys-his-thr-val-glu r Globulin (rabbit) (62) *ala-leu-val-as\ . . . Glucagon (63) *his-ser-glun-gly-thr-phe-thr-ser-asp-tyr-ser-lys-tyr-leu-asp-ser-arg-arg-ala-glun-asp-phe-val- glun-try-leu-mct-aspn-thr* Hemoglobin (41) *val-glun-leu; *val-leu; (horse). *val-gly; *met-gly; *val-ser; *val-glx; *val-asx; (various species, see Table 1 .) 80 Martynas Ycas Hypertensive peptide (46) *asp-arg-val-tyr-val-his-pro-phe-his-leu* Insulin (cattle) (8) 'A' chain: *gly-ileu-val-glu-glun-cys-cys-ala-ser-val-cys-ser-leu-tyr-glun-leu-glu-aspn-tyr-cys- aspn* 'B' chain: *phe-val-aspn-glun-his-leu-cys-gly-ser-his-leu-val-glu-ala-leu-tyr-leu-val-cys-gly- glu-arg-gly-phe-phe-tyr-thr-pro-lys-ala* /5 Lactoglobulin (64) his-ileu* Lysozyme (65, 66, 67, 68) thr-asx-val-glx-ala ; ileu-glx-leu-ala-leu; asx-glx-ala; leu-thr-ala; glx-asx-ileu ; thr-glx-ala-gly ; ser-asx-gly-met-asx; asx-ala-met-lys-cys-arg; val-thr-pro-gly-ala ; ser-asx-arg; lys-phe-glx-gly ; arg-cys-glx-ala ; ser-phe-asx-glx ; thr-asx-arg-arg ; thr-gly-asx-val ; ser-val-cys-ala-lys-gly ; gly-cys-asx ; leu-gly-ala-val ; asx-ileu-pro-cys ; arg-cys-lys-gly ; ser-val-asx-cys-ala ; asx-leu-cys-asx ; arg-asx-cys-ileu; ser-arg-leu; ser-asx-cys-arg-Ieu ; arg-asx; arg-gly; asx-asx; gly-leu; ileu-arg; ileu-asx; ileu-val; leu-leu; ser-ala; ser-leu; val-ala; *lys-val-phe-gly-arg; arg-his-lys; asx-gly-ala-asx-leu* ; glx-ser-phe-asx ; ala-lys-phe-glx; asx-tyr-arg-gly ; arg-gly-tyr-ileu-leu ; asx-ala-tyr-gly-ser-leu-asx; leu-pro; ala-ala-met ; Melanophore expanding hormone (69, 70) *asp-glu-gly-pro-tyr-lys-met-glu-his-phe-arg-try-gly-ser-pro-pro-lys-asp* Myoglobin (71) * gly-leu Ovalbumin (72, 15, 73, 74, 64) val-ser-pro* ; asx-serP-glx-ileu-ala ; glx-serP-ala; ala-gly-val-asx-ala-ala ; cys-ala; cys-val; cys-gly; cys-phe; thr-cys; ser-cys; cys-glx; glx-cys; phe-cys; asx-cys; val-cys; Oxytocin (75) *cys-tyr-ileu-glun-aspn-cys-pro-Ieu-gly-NH2°' Papain (76) *ileu-pro-glu Pepsin (77, 15) *leu-gly-asx-asx-his-glx ; thr-serP-glx ; Prolactin (78) *thr-pro-val The Protein Text 81 Ribonuclease (79) *lys-glu-thr-ala-ala-ala-lys-phc-glun-aig ; lys-ser-arg-aspn-leu-lhr-lys-asp-aig ; lys-aspn ; tyr-glun-ser-tyr ; tyr-Iys; lys-his; asp-ala-ser-val* Salmine(80, 81) *pro-arg-arg; arg-pro-val-arg-arg; pro-ileu-arg; val-gly; arg-val-ser-arg ; arg-ileu-arg; arg-ala-ser-arg ; arg-gly-gly-arg; arg-ser-ser-arg ; val-gly; Serum albumin (37) *asp-ala (man); *asp-thr (cattle); Silk fibroin (Bombyx) (82, 83, 84) gly-ala-gly-ala-gly-[ser-gly-(ala-gly)„]8-ser-gly-ala-ala-gly-tyr n usually 2, mean value always 2. gly-val-gly; tyr-gly; phe-gly; gly-ser-pro-tyr-pro ; tyr-pro-ser-tyr Tobacco mosaic virus (48) thr-ser-gly-pro-ala-thr* Tropomyosin (52) ala-ileu-met-thr-ser-ileu"'' Trypsinogen (85) *val-asp-asp-asp-asp-lys-ileu Vasopressin (40) *cys-tyr-phe-glun-aspn-cys-pro-arg-gly-NH2* Wool (86) ser-cys; gly-cys; thr-cys; ala-cys; leu-cys; cys-gly; cys-thr; cys-ala; cys-val; cys-leu; cys-phe ; remains whether any of the blank cells represent forbidden combinations, or whether they are merely the result of accidents of sampling. To answer this question statistically, the frequencies of occurrence of various combinations have been plotted in Fig. 2. There are more blank cells here than in Fig. 1, as a portion of the data has been discarded to avoid obvious sources of bias. Thus the sequences of silk, collagen, wool and protamine have been omitted, since these proteins have an obviously aberrant structure. Likewise, sequences of less than three residues have not been used, since the ease of 82 Martynas Ycas isolation of various dipeptides varies, making it possible that the frequencies of some peptides have been systematically over- or underestimated. Figure 2 can now be treated as a contingency table with 761 degrees of < < O a: < CL tn < Z Q- < >- o _i o Z Z) -I o >- -I O X 13 U.' LJ -I CO >- -I 1- UJ 2 I a. O (E Ol UJ CO 1- V CC >- H -1 < > 0 ALA 4 1 2 3 1 3 3 2 1 3 1 1 1 26 ARG 1 3 2 2 4 1 2 1 1 1 18 ASP 1 2 4 2 1 1 1 1 13 ASPN 2 2 4 3 2 2 3 1 1 2 2 24 CYS 4 2 3 1 1 2 1 1 1 1 1 1 19 GLU 4 1 1 2 1 1 1 1 1 1 16 GLUN 4 1 1 1 1 2 1 3 1 2 1 18 GLY 3 1 1 1 2 1 1 2 1 1 2 2 1 1 1 21 H IS 1 3 1 1 1 1 8 ILEU 1 2 1 1 1 2 2 10 LEU 1 1 1 2 2 3 1 1 2 2 4 20 LYS 1 1 2 2 1 2 1 1 1 3 1 1 1 2 20 MET 2 1 1 1 5 PHE 1 1 4 1 1 1 1 1 1 2 14 PRO 2 1 1 1 1 2 2 1 1 1 3 16 SER 1 4 1 3 1 1 2 1 1 2 1 1 1 2 2 4 28 THR 2 2 1 1 1 1 3 4 1 16 TRY 1 1 2 TYR 1 1 2 1 2 2 1 1 1 2 1 1 16 VAL 1 1 2 3 3 4 1 1 1 1 1 1 2 22 b c S2 58 20 38 II 24 21 45 19 3B 13 27 22 40 27 48 6 14 11 21 23 43 16 36 6 II 15 29 16 32 20 48 II 27 2 4 15 31 24 46 330 Fig. 2. Frequencies of occurrence of dipeptide sequences in proteins, plotted as in Fig. 1. The sequences of clupein, collagen, salmine, silk fibroin and wool have not been used. Sequences of less than three residues, as well as those where the acid and amide forms of glx and asx are not differentiated, were also not used. On the basis of the study of Ohno (68), glx and asx in lysozyme are assigned to glun and aspn, respectively. The seven-residue sequence common to ACTH and MEH was counted only once, a — marginal totals of rows ; b — marginal totals of columns c — marginal totals of rows and columns. freedom, and the null hypothesis, that there is no correlation between adjacent residues, tested. The deviation X from the expected distribution in Fig. 2 is calculated as: {a^_ -f- a,^{Oj, + ^.;)1^ (1) {a,^ -I a.i){aj. + a.j) An where n is the sum of the marginal totals (330), a^j the value of a cell in column / and row j, Oj and a,, the marginal totals in column and row respectively of the residue defining the column, <7._, and a_,. the analogous values for the residue defining the row. For computational purposes (1) reduces to: ^="(2'K+«i. +«.)"') From Fig. 2, A = 392. The value of /, which is calculated from is 0.414, which is less than 1.645, the 5 per cent confidence limit. (2) (3) The Protein Text 83 It may therefore be concluded that there is no evidence for any intersymbol correlation between nearest neighbors. Inspection of sequences reveals like- wise no obvious correlations of residues more than one removed from each other, but to decide this question definitely will require more knowledge of longer sequences than is now available. Gamow, Rich and Ycas (6) have previously studied this question of intersymbol correlation. They examined a grid diagram, similar to Fig. 2 but embodying fewer data, to see whether the frequencies of entries follow the PoissoN distribution. This method is invalid, since it does not take into account the fact that different amino acids occur with very different frequencies. I am glad to avail myself of this opportunity to correct these authors. IV. FREQUENCY OF OCCURRENCE OF DIFFERENT AMINO ACIDS Amino acids occur with different frequencies in proteins. Some, like leucine, are consistently abundant, others, like methionine, consistently rare. The frequency of occurrence of the various amino acids in the bulk protein of a whole organism, Escherichia coli, is shown in Fig. 3. aiatt , , E.COLI PROTEIN 10 +H \ • AMINO ACID L^„ + TRIPLET j; \ \ ucA S6f _i 6 _ o +i°" 2 2 1 1 1 10 RANK 20 Fig. 3. Composition of bulk protein of Escherichia coli (87), amino acids arran- ged in order of abundance. The vakies for glu, glun and asp, aspn arbitrarily taken as half of glx and asx, respectively. The value of cysteine taken from Roberts and Cowie (88). 'Triplets' refers to the frequencies of triplets of nucleotides, calculated according to the hypothesis of Gamow and Ycas (7) from the composition of E. coli RNA (89). Data on the composition of twenty-three proteins are summarized in Table V. This table shows that the composition of individual proteins is not too different from that of bulk protein. The most abundant amino acid usually has a frequency of about 0.10 to 0.12, the least 0.005 to 0.01. Table V suggests the possibility that the differences in composition of various proteins may be merely the result of chance fluctuations from a mean, and not importantly related to biological function. This notion may not be as far-fetched as might appear at first sight. The most important function of proteins is catalysis, and the enzymatically active site probably involves only a few amino acids. In addition, proteins of a given organism appear to have 84 Martynas Ycas an important mutually complementary relation to each other which enables them to be retained by the cells. This is shown by experiments with injected catalase. Homologous catalase injected into guinea pigs is absorbed by the tissues, but heterologous catalase is rejected (108). Similarly, homologous antibodies readily pass the fetal barriers in rabbits, heterologous pass much less readily (109). This phenomenon is probably connected with the anti- genicity of proteins. The antigenically active sites of proteins are probably also small, and therefore the exact sequence and composition of the major part of the protein m.ay be irrelevant to function. It might be expected, then, that the exact structure of small parts of a protein molecule would be rigidly determined, and any mutation affecting this portion would be eliminated by selection. Mutations affecting the 'irrelevant' portions may not affect the viabihty of the organism, and the same protein in different species may therefore diverge by a process of 'evolutionary drift.' That this process is real is strongly suggested by the facts known about cytochrome c. This enzyme serves the same function and has the same prosthetic group in both yeast and mammalian tissues, but the two cytochromes have very different elution volumes from ion exchange resin columns (110), almost certainly indicating a large difference in amino acid composition. If for each kind of residue there is a characteristic rate of replacement by mutation, the proteins should approach a definite equilibrium composition, if selection is a minor factor. More definitely, each protein will constitute a 'random grab' from a universe of amino acids, the frequencies of the amino acids in this universe being determined by the equilibrium condition. Qualitative considerations suggest that there is something other than selection which tends to make a given amino acid occur with a certain frequency. Certain amino acids, alanine, leucine, isoleucine and valine have aliphatic side chains lacking any obvious reactive functional group. The data on replacements (Table II) indicate, apparently, that one is as good as another, as far as their function in a protein is concerned. Yet leucine is systematically more abundant than isoleucine. These two amino acids are so similar that it is difficult to separate them by paper chromatography. Each of the other aliphatic amino acids has its own characteristic frequency, likewise. Quantitatively, if a sample of « items is drawn at random from a population where an item of type A occurs with frequency p, the distribution of A in a large series of samples is given by the binomial (p + q)", where q = \ — p. In particular, the variance cr^ of the distribution of A is given by o- == npq (4) If the hypothesis of a 'random grab' is correct, then in a collection of proteins the variances of amino acids should be related to the mean value of their frequencies and to the size of the proteins, expressed as the number of residues per molecule. An immediate difficulty is that the sizes of the proteins listed in Table V are not known, and these certainly differ one from another. It should be particularly noted that the relevant size is not necessarily that obtained from physical measurements of diffusion, osmotic pressure and sedimentation. This is because there is ample evidence that physical molecules can be the result of The Protein Text 85 aggregation of smaller, chemically identical units. Furthermore, from the evidence presented in Table III, the several peptide chains constituting some proteins may not be identical, but are nevertheless quite similar. The statistically relevant size of hemoglobin would then be somewhere between 600, the approxi- mate number of residues in the whole molecule, and 150, the average size of the four subunits. Disregarding this difficulty, 1 have plotted the variance of each amino acid, calculated from Table V, against pq (Fig. 4). All points (except glx) fall within O.IO 0.05 • GLX • PRO gly^^'leu CYS» /ARG, 0.05 0.10 pq Fig. 4. Plot of variances of amino acids against/?^, where/? = mean frequency of occurrence of amino acid, ^ = 1 — /?. Line n = 100 calculated variance for sample size (protein) of 100 residues, (glx) is plot with the values from tropo- mysin and y casein omitted. (or very close to) one standard error of the line for n — 100. The fact that the sizes of the proteins are not identical tends to scatter the points, making agree- ment with the hypothesis somewhat more significant. The large deviation of glx is due to its abundance in two proteins, y casein and tropomyosin. If these are omitted the agreement is good. The evidence therefore permits (but of course does not prove) the hypothesis that the composition of proteins is mainly determined not by selection, but rather approximates to a 'random grab' from a single universe of amino acids. There is of course no question that selection can produce proteins of very 86 Martynas Ycas i to > .^ I ft.- <3 02 « R 3 o cr •~ -' ■^ S o V ;^' g E (Z.01) 3SBi;(joqdsoqj (90l) uriunq[B uinjag (901) u!inqoi§opBi ^ (50 1 ) uqnqoiS BuiopXpv (t'Ol) UpjOjd 3AJ3^ (6/,) asEspnuoqi-a (COl) 3SBU3§Ojp;(q3p 3p^q3piBJ33/(ir) (eoi) asBiopiv (301) ssBpijdadXxoqjB^ (10 1 ) aiuXzosXi E^edBtj (001) ujinqoio 'i (66) uiBdHd (86) UI3SE3 / (^6) SSBJ/CuiB iCjBAIlBS (96) 3SEU3§ojpXq3Q loqoDiv (S6) uiLunqjBpBq » (^^6) auiyfeos/Cq (e6) uiqoi§oXj>j (f6) uiqoiSoiuaH (36) uiso/duodojx (36) uiPV (16) uiqiuojqiojd (06) 9SB[/CjoqdsoqdsuBJx r^TtCTvio — oOfNmoor^-^— ^Ov^f^fNr^^ pOOOfNOOOOOOO-^OOOOO d d d d d d d d d d d o o o o o d d qp-^0-;OOOOOOOOOOOOo voododd-^'^cs^d^rirtiricnTt'-^'^r-^d odrnOsvdrnfSroHddd'^-^'^iO^-''^''^^ vodf^^Hr-^d'^ON'-^ Ttdio'^^io'--^rnr~-r~- oomoo'ovOr^ON'^'—'— 1 lod'^fodod—'^'^vd in(Nr-_poooN"^oqri- I ri>ororn-<^-^' Tt — i' fsj -^' t-^ i/S vd ts Tt Tt ■*' (N vS /^rnOO'0-^0\<-nO\rfirioooO'^oo tNfNod-^'iio-4^dr-^ nTt--ir)oofNi^a\vo c»TtON^--r-^(Nvdt~^vdf<-)<^v-ivovd— 'ro-^'r-^ fOvo(-^i/-ir4rn>nO\ vo— ••^r^O' (S On 0 rn q ■^ CO 0 w^ _ 0 0 ro r-; r) q ^— ( On NO t- 1- 00 ro d ^ iri 00 CO t-' >n VD 0 — q NO t^ q NO r~- d -^ oo' — 00 00 ro ND On ■* <^ NO ro q Tl- NO "* q NO r^ ON " 06 r-^ fN NO 00 fS r-^ NO V-) 00 ^ r^ 00 rn o\ V^ On CO ■* 1-^ — " Tt lY-i "■t T}- -H NO — + -* 00 n 0 >/-i r»-i On 0 00 en — ro d oo' (N CN "* 2 Tt ^ (S NO /^ 0 r-.55 « CO « u oij ao, 45 " »-S ~ C l-l-c{2-iI/3-f-'*J+- > 03 The Protein Text 87 unusual composition. This occurs mainly in cases where the mechanical pro- perties of protein fibers are important, as in keratin, collagen and silk. These have been omitted from Tabic V. The most extreme case known to me is the silk of the Congolese moth Anaphe nialoneyi, where glycine and alanine together constitute 94 per cent of the entire protein (1 1 1). Fox and Homeyer (1 12) have also noted the general similarity of composition of various proteins, but have interpreted it in a quite novel manner. Their suggestion is that proteins are similar because the time that has elapsed since the origin of life has been too short to allow more differences to develop between the various proteins, all of which are presumed to be descendants of a single molecule. I believe the composition of silk tends to indicate that there has been ample time for any conceivable differentiation. V. LENGTH OF PEPTIDE CHAINS I have previously called attention to the apparent fact that the number of residues in naturally occurring peptide chains is an exact multiple of three (113). Since then, a more exact determination of the composition of ribonuclease (79) and the elucidation of the structure of glucagon (63) have shown that this statement is incorrect (Table VI). In view of the predominance of chain lengths Table VI. Length of Protein and Peptide Chains in Number of Residues (Note: Cystine counted as two cysteine residues.) Protein or peptide Number of residues Reference Oxytocin Vasopressin Melanophore expanding hormone I (hog) Insulin 'A' chain Glucagon Insulin 'B' chain Melanophore expanding hormone II (hcg) Melanophore expanding hormone (ox) Ribonuclease 9 9 18 21 29 30 30 48 124 (75) (40) (69, 70) (8) (63) (8) (114) (114) (79) that are multiples of three, it might perhaps be suspected that the exceptions are due to secondary removal of residues, as occurs, for example, in the activa- tion of pepsinogen, trypsinogen, chymotrypsinogen and fibrinogen. The tenta- tive finding of Akabori (quoted in (41)), that the B chain of fish insulin has twenty-nine residues, rather than the thirty found in cattle insulin, makes it doubtful that secondary removal of residues is the explanation. Since twenty- nine (the number of residues in glucagon) is a prime number, and not a factor in the chain lengths of other peptides, it seems reasonable to conclude that peptide chains are not multiples of some fixed number of residues. VI. THE CODING PROBLEM Having examined the protein text, we can now discuss what conclusions we may draw as to the storage, transfer and replication of the information contained in the protein molecule. 88 Martynas Yc^as The gene, and by inference DNA, is thought to contain the infoiTnation which eventually appears as a sequence of amino acid residues in the corre- sponding protein. As shown by a study both of the replacement process and of the amino acid sequences, each residue has an independent genetic representa- tion. These representations are presumably aligned in linear order on the DNA molecule. There is in fact no evidence at present that the gene is anything other than a linear sequence of amino acid determining sites, although the possibility that it may also determine the structure of immunopolysaccharides in an analogous fashion cannot yet be dismissed. Recent biochemical evidence (which I shall not discuss here) indicates that it is RNA, not DNA, which is directly involved in the process of protein forma- tion. Transfer of information therefore involves at least two steps : DNA to RNA, and RNA to protein. The straightforward inference would thus be that DNA serves as a template for the formation of RNA. Absence of cytoplasmic inheritance supports the view that RNA is not a self-replicating structure. This is also supported by four lines of biochemical evidence : 1. The initial rate of incorporation of labeled precursors into nuclear RNA is much greater than into cytoplasmic RNA (115). 2. In Amoeba depleted of RNA, RNA only regenerates if a nucleus is present (116). 3. A one-way flow of RNA from nucleus to cytoplasm can be demonstrated (117). 4. The rate of RNA fomiation is minimal at the time DNA is replicating (118). Unfortunately, this conclusion may be an oversimplification. There is no lack of biochemical evidence pointing in the opposite direction: 1. The composition of nuclear and cytoplasmic RNA is not identical (119). 2. The time curves of precursor incorporation into RNA do not indicate that the nuclear fraction is the precursor of the cytoplasmic (115). 3. Radioactive precursor is incorporated into the RNA of enucleated Acetabularia plants (120). 4. Different strains of RNA viruses are self-replicating. This is difficult to explain if RNA is the product of a DNA template. The problem is to reconcile these apparently discordant facts. Consider first the determination of RNA structure by DNA. Since both DNA and RNA are texts written in a four symbol alphabet, it is natural to suppose that the coding problem is very simple. It is sufficient to assume that one nucleotide of DNA determines one nucleotide of RNA (121). Recent evidence indicates, however, that this is incorrect. It is possible to suppress protein synthesis in susceptible bacteria with chloramphenicol. When this is done using amino acid-requiring strains, it can be demonstrated that amino acids are required for RNA synthesis, even though no protein synthesis is taking place (122, 123, 124). The natural inference, supported by several converging lines of evidence, is that it is not the nucleotides themselves which are the precursors of RNA, but rather compounds containing both a nucleotide and an amino acid. This leads to a unitary picture of the synthesis of RNA and of protein. When such precursors are lined up on a The Protein Text 89 protein-synthesizing template (RNA), the amino acids polymerize to form protein; when lined up on DNA, the nucleotide portions polymerize to form RNA (Fig. 5). If this is correct, an obvious conclusion follows. Since omission of a single amino acid stops RNA synthesis, the RNA-fonning mechanism must distinguish not four, but a minimum of twenty different kinds of items. But since the product contains only four, the RNA in general must contain less information than the template that made it. Several nucleotides in DNA must be involved in selecting a single nucleotide of RNA. Since the template must contain more information than the product, RNA cannot be the template for itself; i.e. it cannot be self-replicating. There is an important exception to this statement. AMINO ACIDS NUCLEOTIDES TEMPLATE Fig. 5. Schematic representation of the synthesis of RNA and protein from common precursors (see text). The nature of the template is presumed to deter- mine whether the aligned precursors polymerize to produce protein or RNA. If the information in the template is reduced below a certain level, it is possible to obtain a product identical to the template itself. The formalization is as follows. While in process of formation, the RNA molecule can be visualized as a sequence of nucleotides to which amino acids are attached (Fig. 5). Before removal of amino acids on polymerization the informational content of the 'proto-RNA,' of length n, is n loga 20. After removal of the amino acids the information content is reduced to n logg 4. If restrictions of some kind exist on the number of combinations allowed, the number possible for 'proto-RNA' will be reduced to b[n loga 20]; (b < 1). Such restrictions on 'proto-RNA' will result in less severe restrictions on the RNA itself, since in general one con- figuration of RNA can correspond to numerous different configurations of 'proto-RNA'. Therefore, if there are 20*" possible configurations of 'proto- RNA', RNA itself has 4*^" possible configurations available (1 > c > b). The information content of RNA will equal that of 'proto-RNA' bn log2 20 = en log2 4 (5 when 1 > c^ 2.166. Since the information content of 'proto-RNA' is now the same as that of RNA, an RNA template could, fonnally, be self-replicating. It is now possible to reconcile the genetic and biochemical facts outlined above. Assume that the synthesis of RNA proceeds in two steps. At the first step, a strand of RNA is synthesized using a DNA template. Information is thus transferred from DNA to RNA. The next step is supposed to occur in the cytoplasin. RNA material is added to the nuclear-synthesized RNA, but in a manner which does not add to the informational content. A model for 90 Martynas Ycas this process could be the building up of a complementary strand of DNA, as in the Watson and Crick scheme for DNA reproduction (125).* Normally, the process stops at this stage, since the RNA molecule has insufficient information to act as a template for itself. In the case of viruses, however, the cytoplasmic process of adding new material to the original RNA Table VII. The Composition of the Protein and RNA of Viruses Composition of protein in moles per cent, of RNA as fractions of 1. t value assumed. It should be noted the influenza virus contains lipid, and the protein analysed may in part be of host provenance. Tobacco Tomato Turnip Southern Influenza Protein Mosaic Bushy Stunt Yellows Bean Mosaic A (126) (127) (128) (129) (130) ala 9.6 8.5 6.6 7.5 5.9 arg 6.4 5.3 1.6 6.4 6.0 asx 10.3 11.1 4.2 7.3 11.7 cys 0.7 0.8 2.3 0.9 — gbc 8.7 5.7 7.1 6.8 7.0 giy 3.9 8.6 4.2 8.9 7.0 his 0.0 1.2 1.5 1.3 1.9 ileu 5.2 3.3 9.0 6.2 8.3 leu 7.1 10.9 8.6 8.3 8.5 lys 1.1 3.4 8.0 3.0 5.2 met 0.0 0.8 2.1 2.6 3.2 phe 5.7 3.6 2.5 3.5 4.7 pro 5.5 3.9 10.2 5.3 4.7 ser 10.0 8.6 8.4 8.7 4.4 thr 11.6 11.0 13.9 11.5 6.5 try 1.1 0.5 0.6 i.ot 1.1 tyr 2.4 2.8 1.5 4.1 3.6 val 10.8 10.0 7.9 6.7 6.1 amide 12.7 11.4 8.0 — — RNA (131) (127) (132) (131) (133) Ad 0.30 0.26 0.22 0.26 0.23 Gu 0.25 0.29 0.18 0.26 0.20 cy 0.19 0.21 0.38 0.23 0.24 Ur 0.27 0.26 0.22 0.25 0.33 results in the production of material identical to the template itself. From this point of view, an RNA virus can be regarded as a specialized RNA molecule, which because of restrictions on the sequence of 'proto-RNA' can act as its own template, utilizing the normal RNA-synthesizing mechanism of its host. The composition of the RNA of viruses lends some support to these ideas. * It is obvious that until more is known about RNA structure the question of its replication can be discussed only in general terms. If RNA is a double-stranded structure, the nucleotide composition shows that bases in the two chains cannot be uniquely paired as in DNA, but each base must pair with one of two others, as shown by the equality of 6-keto and 6-amino groups (89). In attempting to elucidate the details of RNA reproduction information on the number of strands, whether each strand contains all the information of the whole structure, and where the complementary strand is synthesized, is of crucial importance. The Protein Text 91 Normally the number of 6-keto (Gu + Ur) and 6-amino (Ad + Cy) groups in RNA is equal (89). Virus RNA does not necessarily obey this rule, indicating that it differs in this respect, at least, from all the others (Table VII). This hypothetical scheme is presented to show that the apparent contradic- tions of the genetic and biochemical evidence do not make it logically necessary to abandon a unitary view of RNA reproduction. The coding of protein information into RNA has attracted considerable attention, but cannot as yet be considered as solved. Study of the protein text indicates that any solution will have to meet several requirements. Firstly, since exactly twenty amino acids are incorporated into protein, it is clear that at least three nucleotides are needed to determine an amino acid. Gamow (134) has proposed that 20 is a 'magic' number, which is the result of the existence of twenty possible sites of three nucleotides each. Four kinds of items, taken three at a time, give twenty different combinations, if order is disregarded. Crick, Griffith and Orgel (135) point out, however, that there is at least one other way of deriving a 'magic' 20 number. They start by considering the problem of what it is that delimits one amino acid-determining site from another, the 'punctuation mark problem'. Assuming that three bases determine a site, it is a problem why the 3n + U 3n + 2, 3« + 3 bases represent a site, while 3« + 2, 3n + 3, 3« + 4 do not. They solve this problem by assuming that only certain triplets of nucleotides correspond to an amino acid (sense sites), while others do not (non-sense sites). The criterion separating these two types of sites is the following. The set of sense sites are all triplets which, when placed next to each other in any possible combination, give sense sites only at positions 3/z + 1, 2?i + 2, 3n -j- 3, but not otherwise. For example, the triplet AAA is a non-sense site, since when placed next to itself it gives the sequence AAAAAA. The site is not unambiguously defined, as AAA occurs both at the 1-3 position and at the 2-4 position. They find that there are exactly twenty triplets (out of sixty-four) which satisfy the criterion of sense sites, as follows : ABA BCA ADC BDD ABB BCB ADD CDA ACA BCC BDA CDB ACB ADA BDB CDC ACC ADB BDC CDD Other ways of selecting twenty sense sites are also possible. The sense sites, these authors suggest, may correspond to amino acid-selecting sites of RNA. The 'punctuation mark problem' could, of course, also be solved if amino acids were selected in a sequential manner starting from one end of the template. Secondly, besides the requirement that at least three nucleotides are required to determine an amino acid site, the study of proteins indicates that these amino acid determining sites are independent and share no nucleotides with their neighbors. This conclusion follows from the absence of any intersymbol correlations in the protein text, and also from the fact that a mutation (as inferred from a study of homologous proteins) can result in a change at one site only, leaving the rest of the sequence unchanged. The number of nucleotides 92 Martynas Ycas in the template must therefore exceed the number of residues in the correspond- ing protein by a factor of at least three. Absence of intersymbol correlation shows that the 'overlapping' codes discussed by Gamow, Rich and Ycas (6) do not correspond to reality. The third requirement is somewhat more hypothetical. From the evidence presented above, it would appear that selection is not the sole factor determining the frequency of occurrence of the various amino acids. This is strongly suggested by the different frequencies of amino acids with aliphatic side chains, and particularly by the characteristic preponderance of leucine over isoleucine. It is therefore reasonable to believe that the coding principle itself imposes certain differences in frequency on the various amino acids. If only one configuration of nucleotides corresponds to each amino acid, the coding per se cannot make some amino acids frequent and others rare. This can be done, however, if some amino acids have more than one configura- tion of nucleotides to which they correspond. For this reason I am inclined to believe that the type of coding proposed by Crick, Griffith and Orgel (135) does not correspond to reality. Gamow and Ycas (7) have proposed a code that formally meets these three requirements. An amino acid is presumed to be determined by three nucleotides, taken without regard to order. In addition, the number of nucleotides in the RNA is assumed to be three times the number of amino acid residues in the corresponding protein. This has the following consequences: 1 . There are twenty such triplets, the same as the number of amino acids. 2. Neighboring triplets share no nucleotides between them. Any sequence of amino acids is thus permitted. 3. The frequencies of various amino acids, calculated on the assumption that the sequence in RNA is random, are unequal. This is because the expected frequency of any triplet is given by the product of the frequencies of the com- ponent nucleotides and the number of configurations for the given composition. Thus there are six triplets (all presumed to determine the same amino acid) of the type ABC, three of AAB and one of AAA. The pattern of frequency distribution of the various triplets, calculated in this manner, corresponds very closely to the amino acid distribution, as shown, for example, in Fig. 3 for the case of E. coli. I believe that this type of coding, even if not itself the one wliich actually occurs, is similar to the one that corresponds to reality. The most striking defect is that it provides no explanation, in fact contradicts, the requirement that in RNA the number of 6-keto groups should equal the number of 6-amino groups. H. A. Simon (136) has proposed a modification to take care of this difficulty. If RNA is a paired structure, somewhat similar to DNA, and 6-keto bases pair with 6-amino ones, then the following four pairs of nucleotides exist (again disregarding order) : Ad-Gu; Ad-Ur; Cy-Gu; Cy-Ur. If one takes these pairs, rather than the individual nucleotides, as units, one can maintain an hypothesis of determination by sextuplets, analogous to determination by triplets. The frequency distribution of sextuplets, calculated for a random RNA sequence, is very similar to that obtained for the triplet The Protein Text 93 distribution. This suggests that a whole series of codes of this type may exist, all having similar general properties. At present the major difficulty is not to produce a coding principle that explains the known facts, but rather to make a choice between the many that are possible. The correctness of a coding principle can, in general, be ascertained from a consistency of correspondence of the RNA and protein texts. Unfortunately, such a direct approach is not at present possible. Except perhaps in the case of RNA viruses, it is not possible to isolate a pure RNA corresponding to a pure protein, and were this possible, the sequence of nucleotides could not be deter- mined by any method currently available. If the composition only of a series of RNA's and the corresponding proteins is known, it is theoretically possible to check some coding schemes as follows: If the coding scheme is correct, the various configurations of nucleotides can be assigned to the amino acids in such a manner as to give, when summed over the protein, the experimentally determined RNA composition, and this con- sistently for all RNA-protein pairs. No assumption need be made that the RNA sequence is random. Actual application of this method requires a large number of RNA protein pairs of accurately determined composition, obviously diftering as much as possible from each other, and the facilities of an electronic computer. The electronic computer is much the easier of the two to provide. At present the data are hopelessly inadequate, although analyses of the proteins and RNA's of viruses may eventually make such an approach possible. However, in attempting a correlation of viral RNA and protein (Table VII), it should be remembered that some viral RNA's do not show the equality Ad + Cy = Gu + Ur characteristic of non-viral RNA (89). This suggests that normal RNA may be multi-stranded, while viral may not be. It is therefore not im- possible that viral RNA may contain all the information, but not all the material of a protein determining structure, and hence differ in composition from it. An additional difficulty is that it is not certain that all viral RNA is concerned in the determination of the protein which eventually appears in the virus particle. In lieu of anything better, I have attempted to make consistent assignments of triplets to amino acids on the assumption that the sequence in RNA is random. The random frequencies of triplets were calculated for liver (Fig. 5), Tobacco Mosaic and Turnip Yellow virus. I then tried to assign each triplet to an amino acid in such a manner that each member of the pair would have approximately the same frequency in the three cases. No satisfactorily consistent assignments could be obtained by this method. Assuming that the RNA's and proteins actually correspond, failure indicates one or more of the following: 1 . The coding principle used is false. 2. The RNA is not a random sequence. 3. The proteins of viruses are so small that relatively large deviations from expected frequencies may be found. The molecular weight of TMV protein is about 17000 (48, 137), that of Southern Bean mosaic about 26000 (129), Several of the amino acids occur as only a few residues per molecule, so that a 94 Martynas YCas difference of one or two residues from the statistically expected value produces very large relative deviations. Since the frequency of occurrence of an individual amino acid is small, even a larger protein such as hemoglobin may be too small to be a statistically valid sample for the purpose of calculating frequencies on the basis of a random RNA sequence. The following case is of interest. The RNA's of liver and of reticulocytes are virtually identical in composition, and therefore the proteins (bulk liver protein and hemoglobin) would be expected to have a very similar composition. Actually, this is not the case (Fig. 6). Considerable differences RNA RNA RAT LIVER RETICULOCYTES A D 18.4 17.5 6 U 33.1 34.7 C Y 30.5 29.9 U R 18.0 17.9 4 6 8 LIVER PROTEIN Fig. 6. The composition of bulk liver protein (142) and hemoglobin (93). The RNA composition of liver from (89) of reticulocytes (143). All in moles per cent. exist, as can be seen from the deviations of the points from the line of slope 1 . It would be better to use for this purpose the bulk RNA's and proteins of whole organisms and organs, were it not for the fact that bulk protein and RNA from various sources is so similar that no strong check on the coding principle is possible. The method of assignments from the assumption of a random RNA sequence fails, then, either strongly to confirm or to deny any proposed coding principle. It is possible that as more information becomes available some light may be thrown on the coding problem from a study of replacements of residues in homologous proteins, if replacements prove to be nonrandom. The reader will not fail to notice that the inadequacy of the data render most of my conclusions tentative. More information of the type considered The Protein Text 95 here will, of course, become available in the future and will not fail to clarify matters. 1 have attempted to organize and analyse such data as exist, in the hope that the value of this sort of information might become clearer, and in order to facilitate their examination as more become available. Obviously, data on composition and sequence are not the only possible sources of information bearing on coding. Strong hints will eventually be obtained from a study of RNA structure and sequence, as well as from other, more conventional, biochemical approaches. The solution of these problems will surely not be long delayed. Acknowledgment— \i is a pleasure to acknowledge the collaboration of George Gamow. I have profited from discussions of various aspects of these problems with Drs F. H. C. Crick, Beatrice S. Magdoff and Herbert A. Simon. Dr Louis J. Cote has given valuable assistance with statistical problems. The errors, of course, remain my own. REFERENCES 1. H. Fraenkel-Conrat : The role of the nucleic acid in the reconstitution of active tobacco mosaic virus. /. Amer. Cheni. Soc. 78, 882-883 (1956). 2. A. GiERER and G. Schramm: Infectivity of ribonucleic acid from tobacco mosaic virus. Nature, Lond. 177, 702-703 (1956). 3. A. GiERER and G. Schramm: Die Infektiositat der nucleinsaure aus tabakmosaikvirus. Z. Natiirf. lib, 138-142 (1956). 4. A. L. Dounce: Duplicating mechanism for peptide chain and nucleic acid synthesis. Enzymologia 15, 251-258 (1952). 5. D. Schwartz: Speculations on gene action and protein specificity. Proc. Nat. Acad. Sci., Wash. 41, 300-307 (1955). 6. G. Gamow, A. Rich, and M. Ycas: The problem of information transfer from the nucleic acids to proteins. Advances in Biological and Medical Physics 4, 23-68, Academic Press, New York (1956). 7. G. Gamow and M. Ycas: Statistical correlation of protein and ribonucleic acid com- position. Proc. Nat. Acad. Sci., Wash. 41, 1011-1019 (1955). 8. A. P. Ryle, F. Sanger, L. F. Smith, and R. Kitai: The disulfide bonds of insulin. Biochem. J. 60, 541-556 (1955). 9. W. J. NiCKERSON and G. Falcone: Identification of protein disulfide reductase as a cellular division enzyme in yeasts. Science 124, lll-lli (1956). 10. D. Mazia: Materials for the biophysical and biochemical study of cell division. Ad- vances in Biological and Medical Physics 4, 69-118, Academic Press, New York (1956). 11. E. E. Howe: Properties of amino acids. In Amino Acids and Proteins, ed. by D. M. Greenberg, pp. 13-55, Charles C. Thomas, Springfield, 111. (1951). 12. E. Windsor: a-Amino adipic acid as a constituent of a corn protein. /. Biol. Cheni. 192, 595-606(1951). 13. E. Work and D. L. Dewey: The distribution of a, e diaminopimelic acid among various micro-organisms. J. Gen. Microbiol. 9, 394-409 (1953). 14. H. Smith, R. E. Strange, and H. T. Zwartouw: a, e-Diaminopimelic acid in the peptide moiety of the cell wall polysaccharide of bacillus anthracis. Nature, Lond. 178, 856-866 (1956). 15. M. Flavin: The linkage of phosphate to protein in pepsin and ovalbumin. J. Biol. Chem. 210,771-784(1954). 96 Martynas Ycas 16. F. R. Bettelheim: Tyrosine-o-sulfate in a peptide from fibrinogen. /. Amer. Cheni. Soc. 76, 2838-2839 (1954). 17. J. M. Barry: Use of glutamine by the mammary gland for the synthesis of casein. Nature, Lond. 174, 315-316 (1954). 18. F. M. SiNEX and D. D. Van Slyke: The source and state of the hydroxylysine of collagen. J. Biol. Chem. 216, 245-250 (1955). 19. F. M. SiNEx: personal communication. 20. M. R. Stetten: Some aspects of the metabolism of hydroxyproline, studied with the aid of isotopic nitrogen. /. Biol. Chem. 181, 31-37 (1949). 21. G. Wolf, W. W. Heck, and J. C. Leak: The metabolism of hydroxyproline-a-C" in the intact rat. Radioactivity in amino acids from proteins. /. Biol. Chem. 223, 95-105 (1956). 22. G. Burnett and E. P. Kennedy : The enzymatic phosphorylation of proteins. /. Biol. Chem. Ill, 969-979 (1954). 23. J. Roche and R. Michel: Thyroid hormones and iodine metabolism. Amm. Rev. Biochem. 23, 481-500 (1954). 24. M. Flavin: The linkage of phosphate to protein in pepsin and ovalbumin. /. Biol. Chem. 210, 771-784 (1954). 25. A. Rich and F. H. C. Crick: The structure of collagen. Nature, Lond. 176, 915-916 (1955). 26. H. A. iTANo: The hemoglobins. Annu. Rev. Biochem. 25, 331-348 (1956). 27. T. H. J. HuiSMAN, J. H. P. Jonix, and P. C. Van Der Schaaf: Amino acid composition of four different kinds of human haemoglobin. Nature, Lond. 175, 902-903 (1955). 28. V. M. Ingram: A specific chemical difference between the globins of normal human and sickle-cell anaemia haemoglobin. Nature, Lond. 178, 792-794 (1956). 29. J. V. Evans, J. W. B. King, B. L. Cohen, H. Harris, and F. L. Warren: Genetics of haemoglobin and blood potassium differences in sheep. Nature, Lond. 178, 849-850 (1956). 30. D. W. Green, A. C. T. North, and R. Aschaffenburg : Crystallography of the ^-lactoglobulins of cow's milk. Biochim. Biophys. Acta. 21, 583-585 (1956). 31. N. H. Horowitz and M. Fling: The role of the genes in the synthesis of enzymes. In Enzymes: Units of Biological Structure and Function, ed. by O. H. Gaebler, pp. 139-145 Academic Press, New York (1956). 32. W. K. Maas and B. D. Davis: Production of an altered pantothenate-synthesizing enzyme by a temperature-sensitive mutant of Eschericia coli. Proc. Nat. Acad. Sci., Wash. 38, 785-797 (1952). 33. T. L. Deich and E. T. Soreni: Aminokontsevie gruppi gliadinov i ich izmenenie pod vlianiem mezhrodovoi gibridizatsii. C. R. Acad. Sci. U.R.S.S. 98, 623-626 (1954). 34. J. I. Harris, F. Sanger, and M. A. Naughton: Species differences in insulin. Arch. Biochem. Biophys. 65, 427-438 (1956). 35. J. C. Kendrew, R. G. Parrish, J. C. Marack, and E. S. Orlans: The species specificity of myoglobin. Nature, Lond. 174, 946-949 (1954). 36. K.Felix: Zur chemie des zelkerns. Experiential, 'i\2-'i\l {\952). 37. E. O. P. Thompson: The N-terminal sequence of serum albumins; observations on the thiohydantoin method. /. Biol. Chem. 208, 565-572 (1954). 38. W. F. White, J. Shields, and K. C. Robbins: C-terminal sequence of crystalline bovine and human serum albumins: relationship of C-terminus to antigenic determinants of bovine serum albumin. J. Amer. Chem. Soc. 11, 1267-1269 (1955). 39. H. TuppY and S. Paleus: Study of a peptic degradation product of cytochrome c. I. Purification and chemical composition. Acta Chem. Scand. 9, 353-364 (1955). 40. V. Du Vigneaud, H. C. Lawler, and E. A. Popenoe: Enzymatic cleavage of glycinamide from vasopressin and a proposed structure for this pressorantidiuretic hormone of the posterior pituitary. /. Amer. Chem. Soc. 75, 1880-1881 (1953). 41. H. Ozawa and K. Satake: On the species difference of N-terminal amino acid sequence in hemoglobin I. /. Biochem. 42, 641-648 (1955). The Protein Text 97 42. L. LoRAND and W. R. Middlebrook: Species specificity of fibrinogen as revealed by end-group studies. Science 118, 515-516 (1953). 43. P. H. Bell: Purification and structure of /^-corticotropin. J. Amer. Cfiem. Soc. 76, 5565-5567 (1954). 44. W. F. White and W. A. Landmann: Studies on adrcnocorticotropin. XL A pre- liminary comparison of corticotropin-A with /3-corticotropin. J. Amer. Chem. Soc. 11, 1711-1712(1955). 45. C. H. Li, L L Geschwind, L D. Raacke, J. L Harris, and J. J. Dixon: Amino acid sequence of alpha-corticotropin. Nature, Lond. 176, 687-689 (1955). 47. W. S. Peart: Composition of a hypertensin peptide. Nature, Lond. Ill, 132 (1956). 46. L. T. Skeggs, W. H. Marsh, J. R. Kahn, and N. P. Shumway: Amino acid composition and electrophoretic properties of hypertensin L /. Exp. Med. 102, 435-440 (1955). 48. C. L Niu and H. Fraenkel-Conrat: C-terminal amino acid sequences of four strains of tobacco mosaic virus. Arch. Biochem. Biophys. 59, 538-540 (1955). 49. S. Benzer: Fine structure of a genetic region in bacteriophage. Proc. Nat. Acad. Sci., Wash. 41, 344-354 (1955). 50. G. Pontecorvo and J. H. Roper: Resolving power of genetic analysis. Nature, Lond. 178, 83-84 (1956). 51. H. Fraenkel-Conrat: The chemistry of proteins and peptides. Annu. Rev. Biochem. 25, 291-330 (1956). 52. R. H. Locker: C-terminal groups in myosin, tropomyosin and actin. Biochim. Biophys. . Acta. 14, 533-542 (1954). 53. E. O. P. Thompson: The N-terminal sequence of carboxypeptidase. Biochim. Biophys. Acta. 10, 633-634 (1953). 54. T. Posternak and H. Pollaczek: De la protection contre I'hydrolyse enzymatique exercee par les groupes phosphoryles. Etude de la degradation enzymatique d'un peptide et d'un polyose phosphoryles. Helv. Chim. Acta. 24, 921-930 (1941). 55. N. Seno, K. Murai, and K. Shimura: Studies on the N-terminal lysylpeptides of a- casein. /. Biochem. 42, 699-704 (1955). 56. H. Neurath and W. J. Dreyer: The activation of chymotrypsinogen. Isolation and identification of a peptide liberated during activation. J. Biol. Chem. 217, 527-539 (1955). 57. F. Turba and G. Gundalch: Aminosaure sequenz in der umgebung des reaktiven serinrestes in chymotrypsin-molekiil. Biochem. Z. "ill, 186-188 (1955). 58. (This article consists of paragraphs by different authors) Erweitertes makromolekulares kolloquium. Angew. Chem. 65, 349-352 (1953). 59. K. Felix, R.Hirohata, and K.Dirr: Uberclupein. Hoppe-Seyl.Z.2l%,169-H9{\91^). 60. W. A. ScHROEDER, L. M. Kay, J. LeGette, L. Honnen, and F. C. Green: The con- stitution of gelatin. Separation and estimation of peptides in partial hydrolysates. /. Amer. Chem. Soc. 76, 3556-3564 (1954). 61. T. D. Kroner, W. Tabroff, and J. J. McGarr: Peptides isolated from a partial hydro- lysate of steer hide collagen. II. Evidence for the prolylhydroxyproline linkage of collagen. J. Amer. Chem. Soc. 11, 3356-3359 (1955). 62. R. R. Porter: A chemical study of rabbit antiovalbumin. Biochem. J. 46, 473-478 (1950). 63. W. W. Bromer, L. G. Sinn, A. Staub, and O. K. Behrens: The amino acid sequence of glucagon. J. Amer. Chem. Soc. 78, 3858-3860 (1956). 64. C. I. Niu and H. Fraenkel-Conrat: Determination of C-terminal amino acids and peptides by hydrazinolysis. /. Amer. Chem. Soc. 11, 5882-5885 (1955). 65. A. R. Thompson: Amino acid sequence in lysozyme. 2. Elution chromatography of peptides on ion-exchange resins. Biochem. J. 61, 253-263 (1955). 66. R. AcHER, U. R. Laurila, and C. Fromageot: Contribution a Tetude de la structure du lysozyme d'oeuf de poule. Peptide aromatique obtenue par hydrolyse enzymatique. Biochim. Biophys. Acta 19, 97-109 (1956). 98 Martynas Ycas 67. K. Ohno: On the structure of lysozyme. III. On the carboxyl-terminal peptide. /. Biochem. 42, 615-625 (1955). 68. K. Ohno: On the structure of lysozyme II. Characterization of aspartyl, asparaginyl, and glutaminyl residues in lysozyme. /. Biochem. 41, 345-350 (1954). 69. I. I. Geschwind, C. H. Li, and L. Barnafi: Isolation and structure of melanocyte- stimulating hormone from porcine pituitary glands. J. Amer. Chein. Soc. 78, 4494-4495 (1956). 70. J. I. Harris and P. Roos: Amino-acid sequence of a melanophore stimulating peptide. Nature, Lond. 178, 90 (1956). 71. V. M. Ingram: The application of Edman's peptide degradation method to horse myoglobin and haemoglobin. Biochim. Biophys. Acta 16, 599-600 (1955). 72. M. Flavin and C. B. Anfinsen: The isolation and characterization of cysteic acid peptides in studies on ovalbumin synthesis. /. Biol. Chem. 211, 375-390 (1954). 73. M. Otessen and A. Wollenberg: Stepwise degradation of the peptides liberated in the transformation of ovalbumin to plakelbumin. C R. Lab. Carhberg (Chim.) 28, 463^75 (1953). 74. M. Flavin: Cysteine and phosphoserine containing peptide sequences of ovalbumin. Nature, Lond. 173, 214 (1954). 75. V. Du Vigneaud, C. Ressler, and S. Trippett: The sequence of amino acids in oxytocin, with a proposal for the structure of oxytocin. /. Biol. Chem. 205, 949-957 (1953). 76. E.O. P. Thompson: Crystalline papain. IV. Free amino groups and N-terminal sequence. J. Biol. Chem. 207, 563-574 (1954). 77. M. B. Williamson and J. M. Passman: The amino acid sequence at the N terminus of pepsin. /. Biol. Chem. Ill, 151-157 (1956). 78. D. Cole and C. H. Li: N-terminal sequence of prolactin. Fed. Proc. 14, 195 (1955). 79. C. H. W. HiNS, W. H. Stein, and S. Moore: Peptides obtained by chymotryptic hy- drolysis of performic acid-oxidized ribonuclease. A partial structural formula for the oxidized protein. J. Biol. Chem. Ill, 151-169 (1956). 80. R. Monier and M. Jutisz: Contribution a Tetude de la structure de la salmine d'Oncor- hyncus. Biochim. Biophys. Acta 14, 551-558 (1954). 81. R. Monier and M. Jutisz: Contribution a I'etude de la structure de la salmine d'On- chorhynchus. II. Etude de quelques peptides resultant de I'hydrolyse trypsique. Biochim. Biophys. Acta 15, 62-68 (1955). 82. F. Lucas, J. T. B. Shaw, and S. G. Smith: Amino-acid sequence in a fraction of Bombyx silk fibroin. Nature, Lond. 178, 861 (1956). 83. L. M. Kay and W. A. Schroeder: The chromatographic separation and identification of some peptides in partial hydrolysates of silk fibroin. /. Amer. Chem. Soc. 76, 3564- 3568 (1954). 84. E. Abderhalden and A. Bahn: Isolierung von tyrosyl-seryl-prolyl-tyrosin beim stufenweisen abbau von seidenfibroin (Bombyx Mori). Hoppe-Seyl. Z. 219, 72-81 (1933). 85. E. W. Davie and H. Neurath: Identification of a peptide released during autocatalytic activation of trypsinogen. J. Biol. Chem. Ill, 515-529 (1955). 86. R. Consden and A. H. Gordon: A study of the peptides of cystine in partial hydro- lysates of wool. Biochem. J. 46, 8-20 (1950). 87. A. Polson: Quantitative partition chromatography and the composition of E. coli. Biochim. Biophys. Acta 2, 575-581 (1948). 88. R. B. Roberts, D. B. Cowie, P. H. Abelson, E. T. Bolton, and R. J. Britten: Studies of biosynthesis in Escherichia coli. Carnegie Institution of Washington Publication 607, p. 28, Washington, D.C. (1955). 89. D. Elson and E. Chargaff: Evidence of common regularities in the composition of pentose nucleic acids. Biochim. Biophys. Acta 17, 367-376 (1955). 90. F. Friedberg: The amino acid composition of adenosine triphosphate-creatine trans- phosphorylase. Arch. Biochem. Biophys. 61, 263-266 (1956). The Protein Text 99 91. K. Laki, D. R. Komintz, P. Symonds, L. Lorand, and W. H. Seegers: The amino acid composition of bovine prothrombin. Arch. Biochem. Biophys. 49, 276-282 (1954). 92. D. R. Komintz, A. Hough, P. Symonds, and K. Laki: The amino acid composition of actin, myosin, tropomyosin and the meromyosins. Arch. Biochem. Biophys. 50, 148-159 (1954). 93. A. Rossi-Fanelli, D. Cavallini, and L. De Marko: Amino acid composition of human crystallized myoglobin and hemoglobin. Biochim. Biophys. Acta 17, 377-381 (1955). 94. G. JoLLES and C. Fromageot: La proteine lysante II de la rate du lapin. II, Composition en acides amines. Biochim. Biophys. Acta 14, 219-227 (1954). 95. W. G. Gordon and J. Ziegler: Amino acid composition of crystalline a-lactalbumin. Arch. Biochem. Biophys. 57, 80-86 (1955). 96. K. Lange: Aminosaurezusammensetzung kristallisierte alkohol-dehydrogenase aus backerhefe. Hoppe-Seyl. Z. 303, 272-275 (1956). 97. J. Muus: The amino acid composition of human salivary amylase. /. Amer. Chem. ^oc. 76, 5163-5165(1954). 98. W. G. Gorden, W. F. Semmett, and M. Bender: Amino acid composition of y-casein. /. Amer. Chem. Soc.15, 1678-1679 (1953). 99. E. L. Smith, A. Stockell, and J. R. Kimmel: Crystalline papain. III. Amino acid composition. J. Biol. Chem. 207, 551-561 (1954). 100. E. L. Smtth, M. L. McFadden, A. Stockell, and V. Buettner-Janusch : Amino acid composition of four rabbit antibodies. J. Biol. Chem. 214, 197-207 (1955). 101. E. L. Smith, J. R. Kimmel, D. M. Brown, and E. O. P. Thompson: Isolation and pro- perties of a crystalline mercury derivative of a lysozyme from papaya latex. /. Biol. Chem. 215, 67-89 (1955). 102. E. L. Smith and A. Stockell: Amino acid composition of crystalline carboxypeptidase. J. Biol. Chem. 207, 501-514 (1954). 103. S. F. Velick and E. Ronzoni: The amino acid composition of aldolase and d-glyceralde- hyde phosphate dehydrogenase. /. Biol. Chem. 173, 627-639 (1948). 104. B. A. KoECHLiN and H. D. Parish: The amino acid composition of a protein isolated from lobster nerve. /. Biol. Chem. 205, 597-604 (1953). 105. E. L. Smith, D. M. Brown, M. L. McFadden, V. Buettner-Janusch, and B. U. Jager: Physical, chemical and immunological studies on globulin from multiple myeloma. J. Biol. Chem. 216, 601-620 (1955). 106. W. H. Stein and S. Moore: Amino acid composition of lactoglobulin and bovine serum albumin. /. Biol. Chem. 178, 79-91 (1949). 107. J. F. Velick and L. F. Wicks: The amino acid composition of phosphorylase. /. Biol. Chem. 190, 741-751 (1951). 108. R. N. Feinstein, M. Hampton, and G. J. Cotter: Species specificity of catalase. Enzymologia 16, 219-225 (1953). 109. I. Batty, F. W. R. Brambell, W. A. Hemmings, and C. L. Oakley: Selection of antitoxins by the foetal membranes of rabbits. Proc. Roy. Soc. B 142, 452-471 (1954). 110. B. Hagihara, T. Horio, M. Nozaki, I. Sekuzu, J. Yamashita, and K. Okunuki: Comparison of properties of crystalline cytochrome c from yeast, beef heart and pig heart. Nature, Lond. 178, 631-632 (1956). 111. F. Lucas, J. T. B. Shaw, and S. G. Smith: The chemical constitution of some silk fibroins and its bearing on their physical properties. Shirley Institute Memoirs 28, 77-89 (1955). 112. S. W. Fox and P. G. Homeyer: A statistical evaluation of the kinship of protein mole- cules. Amer. Nat. 89, 163-168 (1955). 113. M. YcAS: Numerology of peptide chains. Naturwissenschaften 43, 197-198 (1956). 114. B. J. Benfrey and J. L. Purvis: Purification and amino acid analysis of melanophor- expanding hormone from hog and ox pituitary glands. Biochem. J. 62, 588-593 (1956). 100 Martynas Y6as 115. R. M. S. Smellie: The metabolism of the nucleic acids. The Nucleic Acids, ed. by E. Chargaff and J. N. Davison, vol. 2, pp. 393-434, Academic Press, New York (1955). 116. J. Bracket: Action of ribonuclease and ribonucleic acid on living amoebae. Nature, Lond. 175, 851-853 (1955). 117. L. Goldstein and W. Plaut: Direct evidence for nuclear synthesis of cytoplasmic ribose nucleic acid. Proc. Nat. Acad. Sci., Wash. 41, 874-880 (1955). 118. K. Lark and O. Maaloe: Nucleic acid synthesis and the division cycle of Salmonella typhimurium. Biochim. Biophys. Acta 21, 448-458 (1956). 119. B. Magasanik: Isolation and composition of the pentose nucleic acids and the cor- responding nucleo-proteins. 77/^ Nucleic Acids, ed. by E. Chargaff and J. N. Davidson, vol. 1, pp. 373^07. Academic Press, New York (1955). 120. J. Bracket and D. Szafarz: L' incorporation d'acide orotique radioactif dans des fragments nuclees et anuclees d'Acetabularia mediterranea. Biochim Biophys. Acta 12, 588-589 (1953). 121. L. S. LocKiNGEN and A. G. DeBusk: A model for intracellular transfer of DNA (gene) specificity. Proc. Nat. Acad. Sci., Wash. 41, 925-934 (1955). 122. A. B. Pardee and L. S. Prestidge: The dependence of nucleic acid synthesis on the presence of amino acids in Escherichia coli. J. Bad. 71, 677-683 (1956). 123. F. Gros and F. Gros: Role des aminoacides dans la synthese des acides nucleiques chez Escherichia coli. Biochim. Biophys. Acta 22, 200-201 (1956). 124. M. YcAS and G. Brawerman: Interrelations between nucleic acid and protein bio- synthesis in micro organisms. Arch. Biochem. Biophys. 68, 118-129 (1957) 125. J. D. Watson and F. H. C. Crick: Genetical implications of the structure of deoxyri- bose nucleic acid. Nature, Loud. Ill, 964-967 (1953). 126. F. L. Black and C. A. Knight: A comparison of some mutants of tobacco mosaic virus. /. Biol. Chem. 202, 51-57 (1953). 127. D. De Fremery and C. A. Knight: A chemical comparison of three strains of tomato bushy stunt virus. /. Biol. Chem. 214, 559-566 (1955). 128. E. Roberts and G. B. Ramasarma: Amino acids of turnip yellow virus. Proc. Sac. Exp. Biol. Med. 80, 101-103 (1952). 129. B. S. Magdoff R. J. Block, and D. B. Montie: Amino acid composition of southern bean mosaic virus. Contr. Boyce Thompson Inst. 18, 371-375 (1956). 130. C. A. Knight: Amino acid composition of highly purified viral particles of influenza A and B. /. Exp. Med. 86, 125-129 (1947). 131. R. W. Dorner and C. A. Knight: The preparation and properties of some plant virus nucleic acids. /. Biol. Chem. 205, 959-967 (1953). 132. R. Markham and J. D. Smith: Chromatographic studies of nucleic acids. 4. The nucleic acid of the turnip yellow mosaic virus, including a note on the nucleic acid of the tomato bushy stunt virus. Biochem. J. 49, 401^06 (1951). 133. G. L. Ada and B. T. Perry: Specific differences in the nucleic acids from A and B strains of influenza virus. Nature, Lond. 175, 854 (1955). 134. G. Gamow: Possible mathematical relation between deoxyribonucleic acid and protein. Dansk. Biol. Medd. 22, No. 3 (1954). 135. F. H. C. Crick, J. S. Griffith, and L. E. Orgel: Codes without commas. Proc. Nat. Acad. Sci., Wash. 43, 416^21 (1957). 136. H. A. Simon: personal communication. 137. I. Harris and C. A. Knight: Studies on the action of carboxypeptidase on tobacco mosaic virus. /. Biol. Chem. 214, 215-230 (1955). 138. T. L. Deich and E. T. Szorenyi: Chem. Abstr. 49, 1882 (1955). 139. ZoTAN KoROs: Free amino groups of gliadin. Magyar Kim. Folyoirat 56, 131-136 (1950). 140. L. K. Ramachandran and W. B. McConnell: The terminal amino acids of wheat gliadin. Canad. J. Chem. 33, 1463-1466 (1955). The Protein Text 101 141. T. Deutsch: N-terminal amino acids of gliadins from wheal and rye. Ada Physiol. Acad. Sci. Hiwg. 6, 209-224 (1954). 142. B. S. ScHWEiGERT, B. T. GuTHNECK, J. M. Price, J. A. MiLiER, and E. C. Milier: Amino acid composition of morphological fractions of rat livers and induced liver tumors. Proc. Soc. Exp. Biol. Med. 72, 495-501 (1949). 143. G. Rost: Zusammensetzung der ribonucleinsaure der reticulozyten. Naturwissenschaften 43,499(1956). DISCUSSION Koch: I should like to comment on the result of some recent tracer experiments that have been conducted in Dr Swick's laboratory at the Argonne National Laboratory (1, 2, 3). What we have tried to do is to ask ourselves something about the total balance of the turnover of RNA, DNA, and protein in the tissue which is most often studied by the biochemist; namely, rat liver. The interesting thing that comes out of this is that when suitable tracer experiments are done, you can make the definite statement that in a single cell DNA is syn- thesized when it is produced and DNA stays as a cell compound until the death of the ceil, whereas on the other hand it is very easy to show that all of the RNA in the cell is turned over, and it is turned over essentially with about the same half-life that all of the proteins are turned over in the ceil ; that is, there are no special classes of proteins that are not turned over, especially classes of RNA that are not turned over in this tissue. The immediate conclusion from this is that, inasmuch as the amount of protein is many times more than the amount of RNA, on a molar or other basis, there can be no one-to-one hand-off of this kind. In other words, you cannot take the DNA and make the RNA from it without using it over and over again in a different way than has been suggested here. YcAS : While it may be true that there is turnover of RNA in rat liver, I believe, on the basis of work with micro-organisms, that there is no obligatory turnover of RNA associated with protein synthesis. The RNA, which is part of the protein forming mechanism, is a passive template, and apparent coupling or dissociation of protein and RNA turnover is adequately explained, I think, by the assumption that both have common precursors. Koch: I would just like to add that in the case of micro-organisms it is fairly clear that the protein turnover does not occur (4). It is also pretty well established that DNA and RNA turnover do not occur in an actively growing culture. So the concept of turnover in the micro- organism is not a relevant one. But what it does mean is that you cannot accept some of the proposals that have been described that inherently require the obligatory breakdown of some- thing (RNA), concomitant to the synthesis of another type of molecule (protein). MoROWiTz: I would like to introduce some evidence for an alternative approach to the problem of intersymbol influence. In some work recently published by Sidney Fox (5) analyses are reported on the total protein of soybean, corn, wheat, and rye. These analyses indicate that a very high proportion of the protein molecules have lysine in an N-terminal position and arginine in the next position. This approach to statistical constraints involves an experimental analysis of a population of proteins from a single source as contrasted to Dr Ycas' theoretical analysis of a population of unrelated proteins. We have attempted to determine if any constraints are to be found in E. coli protein. The preliminary results indicate that methionine is found in N-terminal positions in a proportion consistent with a chance distribution. Cystine and cysteine in N-terminal positions may show a considerably greater constraint. YcAs: I think that the method used by Fox and yourself introduces an obvious source of bias, if what you are trying to do is look for intersymbol correlations. The abundances of different species of protein in a cell are not equal, and more abundant proteins contribute more end groups. You have to examine the proteins one by one, giving the same statistical weight to each. A similarity in end groups of proteins from related species indicates not an effect of inter- symbol correlation, but rather descent from a common ancestor. As can be seen from the data I summarized, proteins change only slowly in evolution. Branson: There is one question which has been opened up by Dr Gamow's and Dr Ycas' comments; namely, the whole problem of redundancy in protein molecules. The evidence is fairly conclusive, I believe, that so far as the antigenic action of a protein is concerned, the 102 Martynas Ycas active region is approximately 1 5 A on a side. If the same is true of other biological functions, a great deal of surface area in a protein is passive. At least it is passive for a given specific function. Thus it is reasonable to inquire how much of a protein molecule you can whittle away and keep a given biological property. There is a fairly convincing teleological explanation for this redundancy. In the early history of living systems, the membranes containing the living material might have been rather leaky. Thus to retain the small biologically-active components within the cell, they had to be associated with a large but inactive structure which would not pass out through the large spaces. In the evolutionary scheme, then, there remain many large units where really the functional part is relatively small. So that when one amino acid is taken out and another put in, the sub- stitution does not make much difference so long as it is not in the essential small functioning unit of the protein molecule. YcAS : I am also of the opinion that mere size of an enzyme may be quite important for the totality of its biological functions, even if it seems to make no difference to the catalytic function as measured in a test tube. Which part of a protein is significant and which is not is a matter of what function we are measuring. I doubt that at present we know all the functions of a protein from the point of view of the organism itself. REFERENCES 1. R. W. SwiCK and D. T. Handa: The distribution of fixed carbon in amino acids. /. Biol. Chem. 218, 557 (1956). 2. R. W. SwiCK, A. L. Koch, and D. T. Handa: The measurement of nucleic acid turnover in rat liver. Arch. Biochem. Biophys. 63, 226-242 (1956). 3. R. W. SwiCK and A. L. Koch: The measurement of nucleic acid phosphorus turnover in rat liver by the constant exposure technique. Arch. Biochem. Biophys. 67, 59-73 (1957). 4. A. L. Koch and H. R. Levy: Protein turnover in growing cultures of Escherichia coli. J. Biol. Chem. 217, 947-951 (1955). 5. S. Fox: Evolution of protein molecules and thermal synthesis of biochemical substances. Amer. Sclent. 44, 347-359 (1956). PROTEIN STRUCTURE AND INFORMATION CONTENT* L. G. AUGENSTINE Brookhaven National Laboratory, Upton, New York I. INTRODUCTION In stating that a given system has an information content of a certain number of bits, care must be taken to specify not only the context within which this number has been derived but also an attempt must be made to give meaning and utility to this measure. Specifying the context is particularly important since for most systems there are many levels at which the information content can be derived. For example, the information content for a cell is very low, if one is concerned only whether it is living or dead, but it is very large if one is interested in specifying the parameters of each of its individual elementary particles. In this article, estimates will be made of the information content of given proteins by taking into account that they are a sequence of amino acids which can assume only a discrete number of configurations. An attempt will be made to study some of the factors which affect the infonnation content and the types of constraints which must operate in the elaboration of proteins. Some idea of the magnitude and types of the constraints pertinent to proteins can be obtained from parallel studies on proteins and printed English (for which the constraints are known). Finally, the information content based upon structure will be compared with estimates of information content obtained within the context of protein function. Although the fact has not always been fully appreciated, information measures are usually more effective in selecting among alternative hypotheses than in suggesting new ones. This particular trait arises from the fact that information estimates, which depend only upon the probabilities associated with a class of experimental outcomes, will often describe the degree to which a number of variables interact but indicate little or nothing about the behavior of the individual variables. As a result no novel synthetic procedures or selection principles are advanced here to explain the manner in which polypep- tide sequences and/or configurations are determined. Rather, in this paper information theory considerations have been used to evaluate alternative explanations of some aspects of protein construction. II. ESTIMATION OF STRUCTURAL INFORMATION CONTENT AND CONSTRAINTS At the structural level the total information content (/() of a protein will be treated as the sum of two terms; one (/,) depends upon the amino acid sequence * Research carried out at Brookhaven National Laboratory under the auspices of the U.S. Atomic Energy Commission. 103 104 L. G. AUGENSTINE • VALUES CALCULATED FROM PROTEINS X VALUES CALCULATED FROM ENGLISH PARAGRAPHS 800 700 '~^ MYOSIN • TROPOMYOSIN 600 500 I ALBUMIN rO''^^"^"^'^^^^"^ 400 — • INSULIN (48,000) • SILK FIBROIN EDESTIN '• • OVALBUMIN ZE1N» « •GROWTH HORMONE "^PEPSIN CHYMOTRYPSINOGEN 300 — ;3-LACT OGLOBULIN^' > 200 • GLIADIN .BOVINE SERUM ALBUMIN ^ •lactogenic HORMONE 100 HORSE MYOGLOBIN ♦•^^TH -RIBONUCLEASE >(» INSULIN (12,000) X •SALMINE X N 1 1 1 1 1 lyd^jmox 0.50 0.60 0.70 0.80 0.90 1.00 Fig. 1 . Values of /s/C/Jmax as a function of the number of symbols, A^ in proteins and paragraphs. N/m Fig. 2. Distribution of the normalized frequency, ^—^ of letters and amino acids in the language and protein samples. See the text for further discussion. Protein Structure and Information Content 105 and the other (I^) upon the configurations of the polypeptide chain in the native molecule. Treating sequence and configuration independently should lead to overestimates of 1„ since the pennissible configurations will depend upon the sequence. However, care has been taken to reduce the interaction of the two terms as much as possible, so that for the purposes of this paper no significant discrepancies should occur. Sequence' There are twenty amino acids which are most commonly incor- porated into proteins. Therefore the maximum value of /^ is 4.32 bits (logg 20) per amino acid residue.* It would occur when the twenty amino acids occur equiprobably. Values less than the maximum would occur due to any con- straints upon the amino acid sequence. Branson (I) calculated /, of twenty-six proteins for wliich the frequency of occurrence of the twenty amino acids had been determined (disregarding possible sequential dependencies). He found that those which formed part of a living structure of an organism had an ^ which was greater than 0.70 of the maximum value. His analysis is shown by the dots in Fig. 1. The X's show the result of a similar analysis on language samples. The language study was based on ten paragraphs chosen from diverse sources such as want ads, newspaper articles, textbooks, and magazines and differs from that usually used in analysis of language in that it is based on the paragraph rather than on large continuous samples.! In this case, letters have been treated like amino acids and paragraphs like proteins. Except for the single value of 0.99 the values from proteins and paragraphs agree quite well. Similarities between the distribution of amino acid frequencies and letters can be seen further in Fig. 2. There the ordinate indicates the number of times that a particular normalized frequency occurs ; the normalized frequency is the number of times, n^, that the /th symbol (either amino acid or letter) occurs, divided by N/m, the expected number of times that each type of symbol should occur if all m different kinds of symbols had equiprobable occurrence in the sample of TV symbols. As can be seen in Fig. 2 the distribution of the n ■ normalized frequencies -ttt- for the letters (solid fine) and the amino acids (shaded ^ A'//?; area) are almost identical except for the higher incidence of rarely-used letters in language. This small difference might not have occurred if some of the rarer amino acids, for which assays are difficult, had been included in the data. Constraints — The fact that the distribution of amino acids in non-structural proteins deviates from equiprobability about the same as (or possibly a little less than) the letters in written English, indicates that the constraints producing such unequal frequencies should be of the same order of magnitude as (or slightly less than) those governing English texts. However, this tells nothing about the * This value disregards any influence of residue 'complexions'. However, it is difficult to see how factors other than the identity of the residues can be very important, when one con- siders the freedom of rotation of the /^-groups with respect to the polypeptide chain. t It was felt that such a small-sample statistics study was preferable to one based upon large samples (such as a determination of confidence intervals for /, as a function of the paragraph size), since by essentially duplicating the analyses applied to proteins, insightas to the limita- tions of that procedure could be observed. 106 L. G. AUGENSTINE nature of the constraints or the manner in which they arise. The obvious question arises — is the unequal distribution due to unequal availabihty of the amino acids or is it due to constraints imposed in the processes of synthesis, i.e. by 'intersymbol influence' ?* Is the make-up of the pool of amino acids available to the protein-synthesizing centers indicative of the nature of the processes involved in amino acid synthesis or have these processes become adapted to the peculiar demands of the proteins being synthesized? This is essentially the same as looking at a collection of printer's type and asking the question, did the printer select his supply of type because this particular distribution of letters was all that was available to him or did he purposely purchase his particular assortment because he had found that it satisfied his needs? The possibility that the unequal availability of amino acids in the cellular pool may produce the unequal distribution does not seem likely. The experi- ments of Roberts, Cowie et al. (2, 3) at the Carnegie Institution indicate that it requires a five to thirty-fold excess of exogenous amino acids, such as valine, leucine and isoleucine, before the incorporation of these amino acids into protein is seriously affected in E. coli. In fact, once a substance has been incorporated into the amino acid pool of yeast, 1000 times the normal con- centration of exogenous amino acid does not affect its incorporation into protein (Cowie). Although these are excellent experiments they do suffer from problems of cell membrane permeability, intracellular diffusion, etc.; however, they, along with numerous experiments involving amino acid deficient mutants, suggest that as long as the minimum required amount of each amino acid is present the frequency distribution of the amino acids in the pool has a relatively small influence on the distribution of amino acids incorporated into protein. Two methods have been utilized in searching for intersymbol influence in proteins. In the first (reported previously (4)), the behavior of the normalized n- amino acid frequencies -rjj— were studied in individual proteins. The average normalized frequency of the individual amino acids for the twenty-six proteins was tabulated. Comparing the normalized frequency for the individual amino acids in particular proteins with the corresponding average value from the 26 proteins indicated large deviations in many cases. The gross deviations were examined for correlations between pairs of amino acids, both for positive and negative effects. Examination of the 26 proteins indicated that although there are some correlations between the frequencies of individual amino acids combined in single proteins, none was strong enough to be measurable with any degree of confidence for a sample as small as 26 proteins. Similar examinations of the normalized letter frequencies in paragraphs were investigated for significant deviations of pairs or groups of letters. Although strong intersymbol influences are known to exist between letters (e.g. between * 'Intersymbol influence' is a term commonly used to designate sequential dependencies, i.e. influences upon the identity of a particular element by neighbouring elements, which are not the only types of constraints which might be imposed by a synthesizing center. It is easy to imagine the possibility of unequal 'acceptability' for diff"erent symbols at individual sites on a template in which the factors affecting the specifications of each location are independent of the neighbors. Protein Structure and Information Content 107 q and u) no significant results were detected. Thus it can be concluded that such analyses do not exclude intersymbol influences of the same type or order of magnitude as those in language.* Gamow, Rich, and Ycas (5) have made a more exacting study of possible inter-symbol influences affecting amino acids. They treated the known amino acids as a series of dipeptides which they tallied into a 20 X 20 matrix similar to the 26 >: 26 digram matrices common in language analyses. The distribution for nonstructural proteins in such a 20 X 20 matrix followed quite closely a Poisson distribution. This they state is compatible with the assumption that the occurrence of a given amino acid does not affect the identity of its nearest neighbor. Their comparable analysis for English language gave a distribution which deviated from a Poisson. The Poisson distribution associated with the amino acid dipeptide analysis is not too significant since the sample of experimentally determined sequences is not necessarily a reliable representation of the bulk of amino acid sequences in nature. As Gamovv', Rich, and Ycas point out, their available sample is strongly affected by the composition of ACTH, lysozyme and insulin for which the complete sequences have been determined and the shorter sequences from other proteins are biased due to differential bond labilities within the protein which give rise preferentially to certain amino acids occurring as terminal peptides in the sequences isolated. It was felt that a possible explanation of the difference noted between digram analysis of letters and amino acids was that amino acids were also grouped into word-like structures but that the average number of symbols per 'word' was different than that found in English. Therefore, separate digram analyses were performed on English words having two to five letters, six to nine letters and those having ten or more. All the samples were selected so that the average cell density in the 26 x 26 matrix was 0.44, the same as that of Gamow, Rich, and Ycas, and these also all showed significant deviations from a Poisson distribution. MoROwiTZ (6) and some of the Biophysics group at Yale have been investi- gating the possibility that a polypeptide chain is a segment selected from either a single or a small number of repeating sequences which are invariant for a given chromosomal complement. The particular segments chosen and the unique fashion in which they are combined and folded would then account for the highly specific properties of the individual proteins. The possibiHty also exists that there was an initial long, or at least restricted, set of sequences from which present day polypeptide sequences have evolved in a manner similar to that by which organisms have evolved. Gamow^, Rich, and Ycas (5) have pointed out the most striking evidence for a "phylogenetically common ancestral sequence" in their comparison of the A and B chains of insulin, where the same amino acids occur in equivalent positions in both chains four times. The known sequences containing five amino acids or more (from Table I, ref. 5) were examined for repeating or matching sequences. (This was done by superposing the sequences in all possible permutations.) These data indicate that for proteins from a given species any single repeating sequence must * See the discussion by Dr Platt at the end of this paper. 108 L. G. AUGENSTINE be at least forty amino acid residues or longer. Comparing the sequences of different types of proteins indicated that (a) there is not a master sequence operating among species, or (b) evolution, i.e. amino acid substitution, has been so extensive as to make it undetectable, or (c) the master sequence is 200 residues or longer. The additional sequences (for hormones of sub-protein size) cited by Ycas (7) show that short polypeptide sequences with only minor amino acid differences do occur in cells of different species. Thus, the occurrence of repeating or a restricted number of amino acid sequences may be an explana- tion of the unequal amino acid frequencies observed. This possible restriction provides a basis for estimating the minimum value of Ig. A single, long, completely-detennined sequence would provide a situation of minimum infonnation content for polypeptides selected from it. To select A'^ residues from a sequence of S amino acids would require < log2 S bits to find N and < logo {S — N) bits to determine the starting point; or by another selection procedure, < logg (5" — 1) to find the starting point and roughly logg S/l to determine the end point. Either of these methods of selection gives an estimate of the minimum of /^ which is of the order of 2 log2 S bits. This is a very low minimum since according to the best present estimate (which is obviously too low) S f^ 200 and thus 2 logo 5 ^ 15. Therefore, the minimum of/,, is of the order of 0.1 bit/residue since A^ > 100 for proteins. Even if 5" is found to be 10^ (2 logg S)IN will still only be ~ 0.4 bits/residue. Thus, the search (6) for long master sequences of amino acids is of considerable interest with respect to information content considerations. Summarizing for 7^, we can say that for nonstructural proteins the potential information due to the amino acid sequence should be of the order of 0.85-0.95 of the possible maximum value. Although the constraints necessary to produce such an effect should be of the same order of magnitude as those in printed English, tests comparing language and the available proteins for which amino acid composition or sequences are known indicate that the constraints operating in the elaboration of proteins are probably different from those associated with language. Further, it seems unhkely that the unequal frequency of amino acids in proteins is due to unequal availability of the amino acids in the cellular pool. The possibility that polypeptide chains are segments selected from a single or restricted number of repeating sequences may be an explanation of the unequal frequencies, in which case /^residue would be close to zero. Configuration '• With the present state of knowledge the factors affecting /^ are much more difficult to assess. The number of states available to a poly- peptide chain whose bonds retained all of the lability they had as uncombined amino acids would be essentially innumerable. In fact, about the only con- figurations ruled out would be those resulting in closure of the chain upon itself. However the D- and L- forms do not both exist in nature and as has been pointed out by Pauling, Corey and Branson (8), the a-C, N and O group in the backbone of the polypeptide chain is essentially the planar, resonance O / / structure — C -N— . Other than these primary restrictions the polypeptide chain, in the absence of intramolecular or secondary bonding structures is essentially a random structure. Protein Structure and Information Content 109 Kauzmann (9) has given an excellent discussion of the known types of intramolecular bonds which are responsible for protein folding and which should therefore affect 7^. The most common type is the H— bond, especially those formed between the carboxyl O and the amide H. These are essentially non- specific bonds which can form between any pair of amino acid residues in which the C — O and N — H bonds are oriented at the proper angle. A stronger, more specific, but less common H^ — bond can form between the phenolic OH groups of tyrosine and the carboxyl group of glutamic or aspartic acid (9, 10). Another common type of bond stems from the van der Waals forces, which can exist between the atoms in different portions of the same or neigh- boring chains. The third type discussed by Kauzmann is the so-called hydro- phobic bond, which is distinct from the more commonly discussed van der Waals bonds. This results from the tendency of the more hydrophobic amino acid residues to avoid the aqueous phase and adhere together to form a sort of intramolecular micelle. These bonds, although they possess a low order of specificity, may contribute a good deal of stability since they arise as a result of the fact that the more hydrophobic amino acids cannot participate in the strong H-bonding with the solvent water molecules. Salt bridges, which are the ionic bonds formed between the negatively charged (glutamic and aspartic) and positively charged (lysine and argenine) residues, are another type. However, Jacobsen and Linderstrgm-Lang (11) have presented evidence which indicates that these bonds are of negligible importance as intramolecular protein bonds. One of the most important types of intramolecular bond (at least according to current theories (12)) is the highly specific S — S bond formed between cysteine residues in different portions of the same or neighboring chains. The formation of disulfide bonds as well as the 'strong' H-bonds greatly reduces the number of physical states available to the molecule since they can only be formed at a very few sites in the molecule. Since these two types of bond are the most specific of the intramolecular bonds, they are undoubtedly the most effective in determining variations in structure between different kinds of proteins. Repetitions Structures: Intramolecular bonds fonned in such a fashion as to produce repetitious structures reduce 4 tremendously. In the helical or pleated sheet structures proposed by Pauling, Corey and Branson (8) (and illustrated in (13)) the number of free parameters necessary to describe the configuration completely is extremely low and therefore the information content, /f, is also very lov/. In the helices it is only necessary to specify the length (that is, the total number of residues R), the pitch (3.7 or 5.1 residues per turn) and the exact orientation of the helix with respect to a reference point in the protein. An estimate of the lower bound of /^ can be obtained from these factors as follows: 1) To find the exact number of residues, /?, in a helix requires about 2 log2 R bits.* 2) The pitch requires 1 bit (3.7 or 5.1 residues/turn of the helix). * It is rather interesting that the determination of the value of any integer, either + or — (other than zero), requires exactly In bits, where 2"~^ < i? < 2" (which is close to 2 logs R): II bits are necessary to find that |^| is in the range indicated, // — 1 bits to find \R\ and 1 bit to determine R, i.e. the sign. For example, let R = —48: six questions which can be answered by yes or no will show that \R\ is 33-64; five more questions will determine that of the 32 possible values \R\ = 48 and one yes or no question determines R = —48. Thus, 2* < i? < 2« and 111 = 12 bits. 110 L. G. AUGENSTINE 3) A reasonable value for the number associated with specifying the interhelical bonds would seem to be 7?/2 bits. This arises by assuming RjA interhelical bonds, i.e. one bond per turn of the helix, and the previous discussion of intra- molecular bonding indicates that the identity of each interhelical bond requires about 2 bits of information. Another reasonable value for this factor is i?/4; this would occur for 1 one-bit interhelical bond or 1 two-bit bond every other turn, which attempts to take into account that disulfide and "strong" H-bonds are probably the most important interhelical bonds. Actually this factor could be zero since it may not be possible to specify interhelical bonds independent of the sequence. 4) The information necessary to specify the orientation of each helix with respect to some reference point in the protein is the most difficult factor to estimate. It may be almost zero, since the interhelical bonds may unequivocally determine the orientation of the helix. On the other hand, it should not be larger than (log2 R + 30) bits, where logg R bits is sufficient to determine a specific residue and 30 bits to specify its orientation. The 30 bits would be assigned to the six parameters associated with the two vectors necessary to specify orientation. An average 'grain' of 1 :32 is undoubtedly too coarse for specifying the orientation of a. single isolated helix, but is probably adequate for specifying a helix which is oriented in relation to others in the same molecule. The 7?/2 and 30 and the zero terms have been combined to give 'high' and 'low' values for the estimation of the minimum of /,.. These are calculated as /^residue (in bits) by /./residue - "!' LO W (1) and 30+3 logo R = 0.50H n ff^GH (2) The results as a function of R are shown in Fig. 3. Pauling, Corey and Branson (8) cite examples of heUcal polypeptides for which Ris 11, 18 and 36. The corresponding region of Fig. 3 has been shaded. From these considerations it would appear that the minimum value of /<. should be about 1 to 4 bits/residue depending upon R. Although many proteins appear to be helical in nature, there are others, such as ribonuclease (RNase), which from the available evidence would seem not to be. In RNase the structural specificity appears to be determined pre- dominantly by the S — S bonds with the other intramolecular bonds adding stabihty to the structure. A further discussion of the relative importance of the specific and non-specific intramolecular bonds in maintaining structure will be presented later. It is obvious that an upper limit cannot be assigned to /. as readily as to 7^.. However, since the structures proposed by Pauling, Corey and Branson probably represent polypeptide configurations for which /. is near minimum, it would appear that one bit/residue is a reasonable lower limit for 7^. From the estimates of 7, and 7^ presented here, it appears that for the proteins of general interest 7^ should have a value in excess of 4.5 bits/residue although Protein Structure and Information Content 111 if il is found that polypeptides are chosen from a single, long master sequence the value could be as low as 1 .0 bits/residue. Estimates of 4.5 bits per residue or greater at the structural level give a total information content, /,, for the non-structural proteins in excess of 500 bits (or in excess of 100 bits if the minimum estimate turns out to be the true one). Such an estimate is in sharp contrast to the estimates of 10 bits or less 1000 3 o UJ o z o UJ m 13 Z /RESIDUE-- 0.50 + 30+3 \oq, R 3 loq^ R Ij./ RESIDUE (IN Fig. 3. Limits for estimates of the minimum of 4 as a function of the number of residues per helix. The shaded area indicates helical polypeptide sizes reported by Pauling, Corey and Branson (8). obtained by Quastler and his co-workers (14) as the amount of information which must be transmitted for the proper functioning of most protein-controlled systems (e.g. enzymes, immune bodies). III. ESTIMATION OF STRUCTURAL INFORMATION CONTENT NECESSARY FOR FUNCTION A disparity of at least one order of magnitude or more in passing from one context or level of organization to another is of considerable interest. The ten-fold difference indicates that only a small part of the information potential is actually utilized in information transmission. Does this indicate that information transmission in such systems is very noisy and therefore organisms obtain good transmission by utilizing a very high degree of redundancy? Dancoff (15) proposed a principle of maximum 112 L. G. AUGENSTINE error in which he postulated that an organism (or for instance a protein- controlled system) will commit as many errors as are consistent with normal function, but that the inherent error rate, which is probably quite high for such reactions, is maintained at a tolerable level by the use of redundancy. Resorting again to the language analogy — a protein corresponds to a paragraph in complexity and its function may correspond to the thought which is conveyed by a paragraph. Does the difference in information content between the two contexts mean that in the process of evolution the organisms found that particular polypeptide configurations contained structures which could perform useful functions, but that these polypeptide permutations contained a large amount of excess and useless infomiation which has been perpetuated along with the small amount of information associated with the necessary structure ? Does it indicate that much of the protein structure is involved in secondary features of information transmission (e.g. the acquisition, concentration, and transport of energy) and only a small part of the total information content of the protein is intimately engaged in the process of information transmission ? Or does it indicate that each enzyme or p'rotein is capable of mediating many reactions and our experimental ingenuity has not been able to determine more than just a few of them ? (This is analogous to attempting to measure the information transmitted by a source wliich is transmitting through many channels, by monitoring only a single channel.) The discussion which follows will attempt to throw some hght on these questions. However, two important considerations must always be borne in mind when one is deahng with proteins. They are first and foremost colloidal in nature and therefore much of their activity falls in the realm of surface reactions. In the globular proteins it is quite likely that much of the total structural information content is in the interior of the molecules and therefore is unavailable to participate in information transfer occurring at their surface and can only participate in secondary operations similar to those mentioned above. The second consideration involves the question, just what is required for the transmission of one bit of information by a protein system? It seems very likely that one bit of potential structural information will not always transmit the same amount of information; rather, the efficiency of transmission will depend upon the context within which the performance is measured. For example, it is probably much simpler to attach either a hydroxyl or methyl group to a benzene molecule (which would involve one bit of determination) than it is to construct either a 3.7 or 5.1 helix (which also involves one bit of determination). This is somewhat analogous to the relative difficulties of determining whether a symbol is 0 or 1, or to determining whether one should get married or not ! Ig necessary: It appears in some cases that a fairly large fraction of the potential surface information due to the amino acids present is superfluous. For instance, it has been found in insulin that a large fraction of the residues cannot be critical for function. lodination, sulfonation and chelation, each of which can mask surface i?-groups, have been found not to affect insulin activity. Those residues which are species-specific can also be ruled out as being critical for function. Unfortunately, it is difficult to determine the exact Protein Structure and Information Content 113 degree to which a particular type of residue is masked by a given treatment, so that it is impossible to state exactly the fraction of surface residues which are not critical. In a similar manner, it is possible to mask the lysine and arginine residues on the surface of trypsin without destroying its activity (16). In fact, acetyltrypsin is available commercially (17) and has the ideal feature that with its lysine and arginine /^-groups masked, its ability to act as a substrate for other molecules of trypsin is decreased. Haurowitz (18) has also pointed out that some of the antigenic properties of proteins are in many cases not affected by iodination or sulfonation of receptive surface groups. The work of Raacke (19) has shown that a certain amount of surface heterogeneity (as demonstrated by electrophoretic behavior) is still compatible with a fully active protein. Her results plus the uncertainty found in the analyses of amino acid compositions indicate that an uncertainty of the order of 3 to 10 per cent can occur in the amino acid complement without loss in charac- teristic function. The results of Roberts and Cowie (mentioned previously) involving competition in the amino acid pool also indicate that about 3 to 20 per cent variabihty in amino acid incorporation can occur. However, it should be borne in mind that each position in the polypeptide sequence may not have a 3 to 10 per cent tolerance associated with it; rather, those residues which participate in active sites likely have a zero tolerance. /j necessary: Kalnitsky and Rogers (20) have reported that approximately 15 per cent of the ribonuclease molecule can be digested off with carboxy- peptidase before activity is lost. Work reported by Anfinsen (10, 21) indicates that this estimate may be a little high. Rather, he reports that the carboxy- tenninal three amino acids (valine, serine, alanine) can be removed with no loss in activity; but, that digestion with pepsin which splits off these three plus their neighbor, aspartic acid, and also ruptures a "strong" hydrogen bond in the vicinity produces loss in activity. Partial digestion by subtilisin (10, 22), which apparently digests central portions of the polypeptide chain, leaves the activity of the RNase intact as long as the digested portion is not oxidized. It is also known that fragments obtained either by hydrolysis or partial enzymatic degradation from myosin (23-25), trypsin (26), chymotrypsin (27, 28), lysozyme (29), papain (30) and pepsin (31, 32) retain their activity in certain situations. The results with pepsin and papain are particularly striking. Hill and Smith report no loss in the molar activity of papain (toward a synthetic substrate) after an average of 120 of its 180 residues had been removed by leucine-aminopeptidase (an N-terminal type enzyme). Perlmann has reported that some of the dialyzable fragments (which represent 20 per cent of the total original protein) resulting from pepsin auto-digestion retained 1 to 5 per cent of the original activity toward hemoglobin, but about 75 per cent of the activity of the intact pepsin when tested against the synthetic substrate acetyl 1-phenylalanyl diiodotyrosine. These latter results indicate strongly that pepsin, at least, has more than one active site and the site specific for pep- tide linkages adjacent to an aromatic amino acid depends upon the integrity of only a small portion of the molecule. 4 necessary: Of parallel interest to the above considerations is the question of how much configurational infonnation, I„ is necessary for function? The work of Anfinsen and others (10, 33) indicates that the configuration of RNase 114 L. G. AUGENSTINE can be considerably disrupted without loss in activity. They found that rever- sible denaturation in 8 M urea did not cause permanent loss in activity; in fact the RNase was still active in 8 M urea in which its specific viscosity was 8.9 as compared with 3.3 in aqueous solution. This large increase in specific viscosity indicates that the so-called native configuration can be opened con- siderably without destruction of activity. However, Anfinsen reports that oxidation with performic acid, which disrupts the disulfide bonds, causes irreversible inactivation and an increase in specific viscosity to 11.6. The phenomenon of complete loss in activity upon the appearance of the full sulfhydryl titer has been observed in most proteins. It has also been known for a number of years that different degrees of loss in characteristic activity can occur. A number of workers (34, 35) have studied reversible inactivation of enzymes in which it has been observed that a partial unfolding of the mole- cule can occur with a rise in specific viscosity, change in the optical rotation of the protein solutions, changes in solubility, etc., which upon the proper treatment can be reversed. The thermodynamics for reversible denaturation shown in Fig. 4 indicate that quite hkely the first step is common from protein to protein since AF* is remarkably constant for all proteins. Reversible denat- uration invariably shows an increase in entropy. However, AS* is not constant from protein to protein but varies by a large amount as shown by the unhatched areas to the right in Fig. 4. The author has proposed (12, 36) and discussed elsewhere in this volume (37) a hypothesis involving three steps, which attempts to explain this pheno- menon by ascribing the constant AF* to the initial opening of a disulfide bond. This first step is followed by the rupture of a number of neighboring intramolecular bonds (step 2) with a resulting opening of the molecule indicated by the increase in entropy. According to the proposal, this opening of the mole- cule is sufficient to disrupt the spatial arrangement of critical amino acids causing loss in activity, but enough stability and configuration is retained so that under the proper conditions the original native structure, or at least a structure compatible with activity, can restitute. In this hypothesis the rupture of a second disulfide bond (step 3) allows irreversible inactivation to proceed with essentially complete destruction of the characteristic protein structure. A conversion (using an equivalence derived in reference (38)) has been made in Fig. 4 from AS* to A/^. By assuming an average amino acid residue Table I Protein M.W. X 10-3 M.W. ^- 120 A/, (bits) A/./iV (bits/residue) Pepsin Trypsin Emulsin 36 20 38 300 167 317 78 30 48 0.26 0.18 0.15 Amylase Hemoglobin 59.5 67 496 558 36 110 0.07 0.20 Egg albumin 40 333 226 0.68 Lacto peroxidase Insulin 93 12 775 103 340 18 0.44 0.18 Protein Structure and Information Content 115 weight of 120, A/,/residue is given in Table I for those proteins in Fig. 4 for which the molecular weights are available. Thus A/, for the loss in specific activity is of the order of 0.25 bits/residue (the 0.68 value for egg albumin does not correspond to a loss in specific activity). This indicates that destruction of the right 5 to 25 per cent of /, (assuming /p is close to our minimum estimate of I to 4 bits/residue) causes loss of function, which may be reversible or irreversible depending upon which intramolecular bonds are disrupted. PEPSIN T= 298° K T= 323° K PROTEINASE TRYPSIN (KINASE) TRYPSIN INVERTASE INVERTASE VIBRIOLYSIN TETANOLYSIN HEMOLYSIN (GOAT) RENNIN T = 328°K LEUCOSIN INVERTASE (YEAST I NVERTASE T=333 EMULSIN (WET) AMYLASE (MALT) SOLAN IN HEMOGLOBIN TOSS" K EGG ALBUMIN T=343°K PEROXIDASE (MILK) T= 353° K INSULIN Fig. 4. The equivalence between AS* and A/^ for thermal inactivation. The shaded areas to the left represent AF* and the clear areas to the right AS*. (Adapted from Fig. 1, ref 12, by courtesy of University of Illinois Press.) Summary The above discussions indicate that redundancy considerations are not the explanation of the large excess of structural information content; rather, that only a small fraction of the potential information on the surface of the molecule is actively utilized in information transfer. Haurowitz (18), for instance, has pointed out that experiments with substituted antigens indicate that the antigenic specificity resides in an area on the surface of the protein which is approximately 10 to 15 A in diameter. Results cited here suggest 116 L. G. AUGENSTINE that the four or so amino acid residues which would occupy such a surface area (13) may occur as neighbors on the same chain (30-32). Other results mentioned previously (20, 21, 33) suggest that the critical amino acids do not occur in sequence in a single polypeptide chain. This follows from the con- sideration that digestion should be able to consume an average of about 50 per cent of the protein molecule before an active site composed of four or five adjacent amino acids would be encountered ; whereas one of four or five amino acids making up an active site should be encountered, on the average, after about a 20 to 25 per cent digestion of the molecule if the amino acids are distributed roughly at random. In addition, Kennedy and Koshland (39) has found that phospho-glucomutase when placed in 6 M urea loses its activity but recovers it upon dilution, which also indicates separated locations for the critical amino acids. Therefore it may not be possible to state a general rule concerning the relationship between the loci of critical amino acids within polypeptide chains. It seems that the role of intramolecular bonds is to insure that the amino acids which are critical for function are maintained in the proper spatial relationship to each other so that function can occur. Here again it is impossible to state a general rule as to how many of these intramolecular bonds can be disrupted before loss of function occurs, since apparently all of the hydrogen bonds can be broken in RNase without loss in function but not so in phospho- glucomutase. However, the integrity of the more specific secondary bonds (such as S — S) seems to be much more critical for the maintenance of function. The digestion experiments with pepsin and papain indicate further that it is important where in the molecule the bonds are destroyed. Other than ruling out redundancy as a possible reason for the discrepancy between the large potential information and the measured performance, it is difficult to choose among the other possibilities mentioned. The results with pepsin and papain, which have been mentioned, suggest strongly that much of the information content may be unnecessary for function, but has been perpetuated along with the critical content. However, the results with pepsin indicating that multiple sites do exist makes it impossible to assign a certain fraction of the information content as 'garbage'. How much of the polypeptide chain is involved in secondary features of information transmission and the structural complexity necessary for transmitting one bit of information are factors which are now being actively investigated by a number of workers. The various estimates of 7^, I^ and /<. are tallied in Table II. Table II I total ~ I sequence + 'configuration Maximum Plausible Minimum Necessary for performing a single specific function — 4.32 — >4.5 3.5 >1.0 1.0 15/A^ 1.0 10-90% 25% 35-90% Protein Structure and Information Content 117 IV. CONJECTURES Some of the results considered in preparing this paper lead to rather interest- ing speculation. The repetitious minimum entropy polypeptide structures proposed by Pauling, Corey and Branson (8) have already been mentioned. Such configurations may be generally applicable to macromolecules, since helical structures have also been proposed for desoxyribonucleic acid (DNA) polymers (40) and some viruses (41). Crane (42) states that helical configura- tions occur in linear (uni-dimensional) crystals, i.e. structures where progression from each sub-unit to its essentially identical neighbor is by a repeated process of translation and rotation. Lumry and Eyring (43) predict that once hydrogen- bonded secondary structures are formed the characteristic protein 'conformation' is determined by tertiary folding such that the free energy is minimized. How- ever, this does not explain why crystallization should initially occur and be maintained in solution; and to the author's knowledge no one has advanced arguments which provide a complete basis to account for the apparent preval- ence of minimum entropy biostructures, although there have been discussions of how living organisms produce 'order from disorder' or 'order as a result of order' (44). Considering the innumerable configurations available to bio- logical polymers, the question arises 'Are there criteria which determine that the seemingly improbable, highly ordered structures occur spontaneously?' or 'Are these structures imposed at some specific stage in biosynthesis?' Studies on the reversible denaturation of proteins (34, 35) suggest that the latter possibihty is more probable: that is, mild mistreatment can be reversed; whereas, once a certain molecular disarray or instability occurs, an unfolded state results from which the characteristic, native structure does not reconstitute. Neurath et al. (35) make the interesting point, that even if denaturation is complete enough so that physical properties such as solubility, crystallizing ability, or diffusion constants are seriously affected, some of the molecules may subsequently revert to a biologically active form; whereas, others will tend to reverse the molecular disarray by forming a more condensed state but without successfully restoring the native biological properties. This suggests that, although polypeptide chains have an inherent tendency to form semi- condensed configurations, the highly ordered, biologically-active structures are probably not only imposed during biosynthesis, but represent quasi-stable structures with built-in constraints which tend to cause small fluctuations to revert, i.e. a limited amount of disorder can be restrained without the inex- orable Second Law prevailing. Neurath (35) has also reported that the amount of disarray compatible with reversibility depends upon the type of denaturation. Further, denaturation is not reversible under all conditions but may await a change in pH or temperature. However, it is interesting that although an entropy increase is invariably associated with denaturation, removal of the denaturing agents can cause a decrease, which appears to contradict the Second Law; we will later resolve this apparent contradiction. The quasi-stability of native configurations is suggestive of the situation in diatomic molecules where stability conditions are readily depicted as a local 'weir (relative to the surroundings) or null area in a two-dimensional energy-configuration plot. However, since two dimensions would allow only 118 L. G. AUGENSTINE a very gross specification of the myriad degrees of freedom of macromolecules, some form of multi-dimensional space will be necessary to represent their stability conditions. The biologically significant portion of such a macro- molecular space will also be a 'well', but in a multi-dimensional surface rather than a line plot and will be centered near the locus of native structures in configu- ration space. A fraction of the well will represent conditions consistent with an active macromolecule and the remainder, conditions characteristic of reversible inactivation. Anything outside the well will correspond to states inconsistent with the restitution of a native configuration. The multi-dimensional space can be of sufficient dimensionahty so that all configurations differing by a 'single step' are neighbors. In such a 'fine- grain' specification each microstate and its probability density (as a function of energy, for example) can be represented. However, such a scheme has drawbacks: first, it has little novelty since any situation can be completely described by a sufficient number of parameters ; second, a model dealing only with microstates would be extremely diflftcult to test experimentally; and third, the excessive dimensionality makes it useless as an aid in envisioning possible mechanisms of macromolecular rearrangements. Thus, a 'coarse-grain' specification, which requires reducing the dimension- ality by transforming the microstates into a more useful set of macrostates, is desirable. This general operation can be schematized by the use of the follow- ing contingency table : Table III ■< Molecular Energy > ^1 ^2 ^k ^n °'lll '''lia '^llk '^lln °'l21 (^122 <^12i- ^12n ^m °'ij2 '^lik OCljn "'all °'212 '^21fc ^2ln '^iil '^iji ■ "■ijk • a,. A plausible specification for a multi -dimensional space is given in Table III, where a sufficient number of binary digits is used so that each microstate can be unequivocally identified, e.g. the two atoms involved in each bond as well as the bond length and angle could be identified. Each ol^j^. represents Protein Structure and Information Content 119 the probability density of a given microstate for molecular energy state E,^, where the ranges of /,y and k can be essentially infinite. A transformation to a 'coarse-grain' scheme which seems worth consider- ation is as follows. Each macrostate, M, (depicted by the leftmost column of digits in Table III) designates only which bonds exist in the macromolecule, e.g. sulfur atom no. 7 is hooked to carbon no. 179 and sulfur no. 11, C-563 to C-564 and N-201, etc. Mechanistically all microstates, w,;, contained in a given macrostate, M,, are grouped together by ordering the digits (or analogously ordering the axes in space). To complete the transformation other bond properties, e.g. length and orientation (the other column of digits in Table III), and their associated probabilities (the right hand portion of Table III) are lumped into two gross categories to provide an intuitively manage- able representation. This 'lumped fine structure' for each macrostate, Af,- can be represented on an 'energy-deviation' {ED,) plane at the locus (in trans- formed configuration space) corresponding to A/,: 'deviation' is a measure of instabihty, i.e. the extent to which individual microstates, /77,^, deviate from the configuration »7,^ corresponding to maximum stability for macrostate Mj. An example of a method for constructing such values is: (a) find the set of digits «?,s in the middle column of Table III which represents maximum stability for macrostate M^ and (b) determine how many of the corresponding digits of /;?,, and m^j differ. This number provides an excellent measure of 'deviation' because each microstate has a unique Z)-value and 'neighboring' microstates have adjacent Z)-values. Assigning probabilities to pairs of 'energy' and 'deviation' values completes the "fine" to 'coarse-grain' transformation. This requires summing the probabilities, a,^;;,, of those microstates associated with a particular D-value. The probability densities for E and D values can be arranged into contours of equal probability to avoid further complications of adding a third coordinate to the ED plane. These contours will possibly be quite irregular in shape and may well be discontinuous, since the only obvious restriction on their form is that they be non-intersecting. It should be noted that 'lumping' on to 'energy-entropy' planes would have provided a simpler transformation than that to the 'energy-deviation' planes. The microstates corresponding to a given 'deviation' can be equated to an entropy value by the usual — S/^jlog/), procedure, where the /7/s are the probabilities (properly normalized) associated with the microstates. Such a scheme was considered, but was found to be intuitively less useful than the ED transformation. The 'energy-deviation' scheme is of considerable interest when one con- siders possible mechanisms of both protein inactivation and enzymatic activity. Suppose, for instance, that the energy of a molecule in a native configuration is slowly raised, e.g. by external heat: the point representing 'molecular state' will be driven to new loci in multi-dimensional space. Undoubtedly a trajectory is followed such that the locus resides, 'statistically', on the contour which has the maximum probability permissible or consistent with its energy content and macrostate at any instant. This means that the locus first progresses over the EDj plane of the particular native configuration. A/,-. Eventually a locus will be reached where the probability contour occupied is lower than the corre- sponding contour on an adjacent ED plane. The molecular state will then 120 L. G. AUGENSTINE jump to that adjacent macrostate by some fomi of bond rearrangement.* Even without an immediate change in molecular energy due to external heat, the jump will likely be followed by an instantaneous migration of the molecular state locus on the new ED plane. This would be anticipated since the new locus might not be the position of maximum probability for that instantaneous molecular energy. A sufficient increase in temperature would eventually drive the trajectory out of the fraction of the null region corresponding to an active molecule: with sufficient mistreatment the locus would be driven com- pletely out of the null region into the portion of configuration space representing irreversibly inactivated molecules. Molecular energy will decrease when external heat is removed, and the molecular rearrangements will be reversed or not depending upon the sym- metry of the multi-dimensional surface of the well. Where denaturation is reversed merely by reversing the denaturing conditions, apparently the inacti- vation trajectory is retraced or else the null region is a smooth "well" with no intervening metastable positions in the reversal trajectory. Thus, for reactiva- tion the two trajectories would not have to be identical but need only form a •closed loop. Asymmetry in the probability contours of even one of the ED plots traversed, could cause the inactivation and reversal trajectories to diverge sufficiently so that metastable, non-active configurations would result. Such situations have been observed experimentally; for instance, thermal denaturation at alkaline pH is not reversed upon cooling until the pH is adjusted to acidic conditions (35). Since a change in pH should alter the ED contours it is easy to envision how it could make the reversal of denaturation more likely by changing the transition probabilities between macrostates and thus alter the reversal trajectory. Such an alteration would resolve the apparent contra- diction of the Second Law: a changed pH would act as a 'Maxwell Demon guiding the footsteps of the reversal trajectory'. Considering its likely statistical nature, it is probable that much of the trajectory of the locus of molecular states proceeds along essentially negligible probability gradients, not only with respect to transitions from one macrostate to another but more particularly with respect to instantaneous displacement from the locus of arrival on a new ED plane. Such transitions should be readily reversible and in general of limited consequence except as they lead to regions of larger gradients. However, a 'low-gradient' region would allow considerable leeway in trajectories. This would permit multiple pathways which would account for the spectrum of effects often observed following physical denatura- tion. In those transitions involving bonds which latch large segments of the molecule together (12) (e.g. interhelical bonds) gross molecular rearrangements could occur so that the trajectory would pass through regions of large probability gradients. Such transitions would not be instantaneously reversible and would therefore be relatively important in driving the trajectory away from the "active" portion of or even out of the 'well'. My proposed inactivation hypothesis discussed later (37) attempts to * Somewhat more rigorous discussions of factors aflfecting the trajectory of the locus of molecular state in similar multi-dimensional plots have been given by Teller (45) and Lumry and Eyring (46). Protein Structure and Information Content 121 specify the identity and sequence of high-gradient transitions. On this basis energy from an absorbed quantum, ionization or thermal process would migrate through the molecule in a fashion represented mainly by a 'low gradient' trajectory. However, once the energy or charge becomes localized in a bond of low ionization potential involved in latching large segments of the molecule together, a 'high-gradient' transition, not readily reversible, would occur. The inactivation efficiency of absorbed energy will thus be a function both of the locus of the molecular state at the time energy is absorbed as well as its resulting trajectory; where the trajectory depends upon the amount of energy introduced, the point of absorption and any external factors which affect the contours on the ED planes. For instance, the quantum efficiency of UV varies considerably with pH for a number of enzymes (47). The interdependence of energy, configuration and probability proposed here provides a formalism for depicting enzyme action. It is fairly typical of enzyme, as well as other types of catalysis, that reactions proceed which are normally not feasible because of steric or energetic hindrances. It is entirely possible that because of their large size, enzymes act as large energy reservoirs whose function is to "deliver" a quantity of energy to a particular site or com- plex in an irreversible fashion. Another possibility is that energy may not be delivered per se but as a change in configuration of the enzyme with a corresponding alteration in the spatial relationship between reactants complexed to the enzyme. Within these proposals the formation of the enzyme-substrate complex could have an important function. It could act as an external agent affecting the ED contours so as to cause a directed alteration in trajectory, leading finally to a completed enzyme catalysis. Effective, i.e. rapid and essentially irreversible, enzyme catalysis will likely depend upon (1) an E — S complex formation which involves a high-gradient transition, so as to enhance a drastic alteration in the trajectory of molecular state, and (2) the directed trajectory passing through a high-gradient region, preferably just before completion of catalysis, in order to make reversibility unlikely. REFERENCES 1. H. Branson: Information theory and the structure of proteins. In: Information Theory in Biology, ed. by H. Quastler, 84-104, University of Illinois Press, Urbana (1953). 2. R. Roberts: Carnegie Institution Yearbook:, No. 55, 110-148 (1956). 3. D. Cowie: Carnegie Institution Yearbook, No. 55, 110-148 (1956). 4. L. AuGENSTiNE, H. BRANSON, and E. Carver: A search for intersymbol influences in protein structure. In: Information Theory in Biology, ed. by H. Quastler, 105-118, University of Illinois Press, Urbana (1953). 5. G. Gamow, a. Rich, and M. Ycas: The problem of information transfer from the nucleic acids to proteins. In: Advances in Biological and Medical Physics 4, ed. by J. H. Lawrence and C. Tobias, 23-68 (1956). 6. H. MoROwiTZ, et al.: personal communication. 7. M. Ycas: (preceding paper in this volume). 8. L. Pauling, R. Corey, and H. Branson: The structure of proteins. Proc. Nat. Acad. Sci., Wash. 37, 205-211, 235-285 (1951). 9. W. Kauzmann: In: The Mechanism of Enzyme Action, ed. by W. McElroy and B. Glass, Johns Hopkins Press, Baltimore (1954). 122 L. G. AUGENSTINE 10. C. Anfinsen: Informal lecture at the Fourth Buena Vista Conference on Protein Synthesis (1956). C. Anfinsen: Advances in Protein Chemistry 11, 1-100 (1956). D. Steinberg, M. Vaughan, and C. Anfinsen: Kinetic aspects of assembly and degrada- tion of proteins. Science 124, 389-395 (1956). 11. C. F. Jacobsen and K. Linderstr0M-Lang : Salt hnkages in proteins. Nature, Loud. 164, 411^12(1949). 12. L. Augenstine: Structural interpretations of denaturation data. In: Information Theory in Biology, ed. by H. Quastler, 119-124, University of Illinois Press, Urbana (1953). "Trypsin monolayers at the air-water interface," PhD thesis at the University of Illinois (1956). 13. L. Augenstine: Remarks on Pauling's protein models. In: Information Theory in Biology, ed. by H. Quastler, 75-83, University of Illinois Press, Urbana (1953). 14. H. Quastler: The specificity of elementary biological functions. In: Information Theory in Biology, ed. by H. Quastler, 170-190, University of Illinois Press, Urbana (1953). 15. S. Dancoff and H. Quastler: The information content and error rate of living things. In: Information Theory in Biology, ed. by H. Quastler, 263-273, University of Illinois Press, Urbana (1953). 16. J. Sri Ram, L. Terminiello, M. Bier, and F. Nord: On the mechanism of enzyme action. LVIII. Acetyltrypsin, a stable trypsin derivative. Arch. Biochem. Biophys. 52, 464-477 (1954). 17. Worthington Biochemical Corp.: Descriptive manual 8, Freehold, N.J. 18. F. Haurowitz : Protein synthesis and immunochemistry. In : Information Theory in Biology, ed. by H. Quastler, 125-146, University of lUinois Press, Urbana (1953). 19. I. Raacke: Heterogeneity studies on several proteins by means of zone electrophoresis on starch. Arch. Biochem. Biophys. 62, 184-195 (1956). 20. G. Kalnitsky and W. Rogers: The activity of ribonuclease after digestion with carboxypeptidase. Biochim. Biophys. Acta 20, 378-386 (1956). 21. C. Anfinsen: The inactivation of ribonuclease by restricted pepsin digestion. Biochim. Biophys. Acta 17, 593-594 (1955). 22. S. Kalman, K. Linderstrom-Lang, M. Ottesen, and F. Richards: Degradation of ribonuclease by subtilisin. Biochim. Biophys. Acta 16, 297 (1955). 23. J. Gergeley: Studies on myosin-adenosinetriphosphatase. /. Biol. Chem. 200, 543-550 (1953). 24. A. Szent-Gyorgyi: Meromyosins, the subunits of myosin. Arch. Biochem. Biophys. 42, 305-320 (1953). 25. E. MiLHALYi: Trypsin digestion of muscle proteins. II. The kinetics of the digestion. J. Biol. Chem. 201, 197-209 (1953). E. MiLHALYi and A. Szent-Gyorgyi: Trypsin digestion of muscle proteins. I. Ultra- centrifugal analysis of the process. /. Biol. Chem. 201, 189-196 (1953). 26. F. Nord, M. Bier, and L. Terminiello: On the mechanism of enzyme action. LXI. The self digestion of trypsin, Ca-trypsin and acetyltrypsin. Arch. Biochem. Biophys. 65, 120-131 (1956). 27. M. Kunitz: Formation of new crystalline enzymes from chymotrypsin. /. Gen. Physiol. 22, 207-237 (1938). 28. J. Gladner and H. Neurath: Carboxyl terminal groups of proteolytic enzymes. II. Chymotrypsins. J. Biol. Chem. 206, 911-929 (1954). 29. J. Harris, C. Li, P. Condliffe, and N. Pon: Action of carboxypeptidase on hypophyseal growth hormone. J. Biol. Chem. 209, 133-143 (1954). 30. R. Hill and E. Smith: Crystalline papain. Biochim. Biophys. Acta 19, 376-377 (1956). 31. G. Perlmann: Formation of enzymatically active dialysable fragments during auto- digestion of pepsin. Nature, Lond. 173, 406 (1954). Protein Structure and Information Content 123 32. G. Perlmann: Discussion of paper by D. Koshland in J. Cell. Camp. Physiology, 47, Supplement 1, 217-234 (1956). 33. C. Anfinsen, W. F. Harrington, Aa. Hvidt, K. Linderstrom-Lang, M. Otteson, J. ScHELLMAN : Studies on the structural basis of ribonuclease activity. Biochim. Biophys. Acta 17, 141-142 (1955). 34. A. Stearn: Kinetics of biological reactions with special reference to enzymic processes. Advances in Enzymology 9, 25-75 (1949). 35. H. Neurath, G. Cooper, and J. Erickson: The denaturation of proteins and its apparent reversal. /. Phys. Chem. 46, 203-211 (1942). H. Neurath, J. Greenstein, F. Putnam, and J. Erickson: The chemistry of protein denaturation. Cliem. Rev. 34, 157-265 (1944). 36. L. AuGENSTiNEand R. Ray: Trypsin monolayers at the air-water interface. III. Structural postulates on inactivation. J. Phys. Chem. 61, 1385-1388 (1957). 37. L. Augenstine: Discussion of a proposed mechanism of protein inactivation. (Part IV of this volume.) 38. L. Augenstine: Information and thermodynamic entropy. In: Information Theory in Biology, ed. by H. Quastler, 16-20, University of Illinois Press, Urbana (1953). 39. E. Kennedy and D. Koshland: Properties of the phosphorylated active site of phos- phoglucomutase. J. Biol. Chem. 228, 419-433 (1957). 40. J. D. Watson and F. H. C. Crick : Molecular structure of nucleic acids. Nature, Land. Ill, 737-738 (1953). 41. H. Fraenkel-Conrat: Rebuilding a virus. Sci. Amer. 194 (6), 42^7 (1956). 42. H. Crane: Principles and problems of biological growth. Sci. Monthly 70, 376-389 (1950). 43. R. LuMRY and H. Eyring: Conformation changes of proteins. /. Phys. Chem. 58, 110-120 (1954). 44. E. ScHRODiNGER : What is Life? Cambridge University Press, Cambridge, England (1946). 45. E. Teller: The crossing of potential surfaces. /. Phys. Chem. 41, 109-116 (1937). 46. R. Lumry and H. Eyring: Energy exchange in photoreactions. In: Radiation Biology III, ed. by A. HoUaender, 1-70, McGraw-Hill Book Co. (1956). 47. P. Finkelstehm and A. McLaren: Photochemistry of proteins VI. pH dependence of quantum yield and UV absorption spectrum of chymotrypsin. /. Poly. Sci. 4, 573-582 (1953). 48. H. Simon: On a class of skew distribution functions. Biometrika 42, 425-440 (1955). DISCUSSION Platt: Simon (48) has shown that skewed distributions (Yule distributions), such as those in Fig. 2, can be obtained from models based on probabiHty assumptions much weaker than those we were looking for. Thus our inability to determine constraints from a study of the distribution of amino acid and letter frequencies in proteins and words is not surprising. However (in agreement with our summarizing statement for that section), Simon points out that the occurrence of a Yule distribution does not obviate more stringent constraints as the underlying probability mechanism. SPECIFIC MECHANISMS OF PROTEIN SYNTHESIS AND INFORMATION TRANSFER IN THE DEVELOPING CHICK EMBRYO* H. R. Mahler, H. Walter, A. Bulbenko and D. W. Allmann Department of Chemistry, Indiana University, Bloomington, Indiana Abstract — Some preliminary data on precursors and pathways of protein biosynthesis in chick embryos have been presented. The tentative conclusions stated are: 1. Egg white proteins are not utilized for the synthesis of embryonic proteins up to and including the ninth day. Soluble proteins added to the yolk are incorporated effectively, and preferentially to some of the yolk proteins proper. 2. Proteins, peptides and amino acids injected into the yolk sac are incorporated at approximately equal rates. Considering the relative available pool sizes of the various pre- cursors present in the egg, added proteins have to be regarded as the preferred amino acid source of embryonic proteins. 3. A common precursor formed efficiently from proteins and relatively slowly from added amino acids and peptides is considered a likely intermediate in the process. 4. Homogenates of adult organs injected into embryos can be used to elicit a response previously reported for organ transplants, i.e. the apparently specific transfer of labeled material from donor organs to the corresponding organ in the embryonic host. The super- natant fraction of the cytoplasm appears to be, at least in part, responsible for the results observed. I. INTRODUCTION It is the purpose of this contribution to describe, in brief, some preHminary experiments on a controlled biosynthetic activity, namely, the precursors and pathways of protein formation. It differs from most of the papers in this symposium in dealing with phenomena rather than with concepts and in the absence of any attempt to establish a functional correlation between these biological phenomena and information-theoretical abstractions. It shares with other papers in this volume the properties of being highly tentative, and in presenting data and comments on a subject to which it is felt information theory should eventually make significant contributions. With the hope that arrival of that time might be hastened and that thought and discussion might be stimulated, our data are presented for consideration. Some of the results are derived from single experiments only and thus lack further con- firmation. All of the approaches and conclusions reported are still under active investigation and thus subject to revision and modification. Embryos were chosen for the experiments since their cells exhibit two fundamental and related properties, both apparently controlled by the nuclear * The investigations reported have been supported by grants-in-aid of the National Heart Institute, National Institutes of Health, U.S. Public Health Service (Grant No. H 2177) and of the National Science Foundation. This article is contribution No. 746 from the Department of Chemistry, Indiana University. 124 Specific Mechanisms of Protein Synthesis in the Developing Chick Embryo 125 machinery, which set them apart from otlier cells of higher organisms. These are: the capacity for replication, that is, rapid yet controlled growth; the capacity for differentiation, that is, continuous yet controlled change and evolution (1). Therefore, one might consider this the system of choice for attempts at discovering how the information content of the hereditary material, the genetic potentialities, are translated into progressive biochemical capabilities and thus into physiological and morphological realities (2). The experiments were done with chick embryos in ovo because of the ease of handling and the essentially closed and self-contained nature of the experimental system. Further- more, there is a relative paucity of reliable, modern information available about their metabolism and that of embryos of higher vertebrates in general, as contrasted to the large body of knowledge derived from experimental embryology. Our eventual aim is to study the initiation, the mode, and the control of synthesis of highly specific, respiratory enzymes as an indicator of controlled biosynthetic events; however, our initial investigations deal with the more modest one of a definition of parameters for embryonic protein synthesis (3). For any protein formed de novo, as has been pointed out by Spiegelman (4) essentially three different mechanisms may be envisaged: 1. The rearrangement of pre-existing protein molecules; namely, the urprotein hypothesis of Northrop (5), with suitable modifications. 2. The accretion of amino acids on to pre-existing proteins or peptides. 3. De novo synthesis from amino acids. In the special case of the formation of induced enzymes in rapidly dividing bacterial cells and cell-free systems derived therefrom, the evidence is over- whelmingly in favor of the third alternative (4, 6). The situation is not nearly as straightforward in the vertebrate systems studied. On the one hand, for example, Work and collaborators investigated the synthesis of milk proteins (7), Velick, Simpson and co-workers the synthesis of several specific enzyme proteins for muscle (8, 9), and Loftfield and Harris the synthesis of liver ferritin (10). All this work was in vivo and by different experimental techniques, but all these authors presented strong evidence for the last alternative and against the first two. On the other hand Anfinsen and his co-workers, working with hen's oviduct in vitro, have demonstrated that in short term incubations incorporation of amino acids into freshly formed ovalbumin is non-uniform, which is suggestive of the second alternative, but that after longer periods there is a redistribution towards unifonnity (11). Similar results have also been obtained for ribonuclease and insulin synthesis by pancreas sHces. In the case of the proteins of the chick embryo proper, Francis and Winnick have presented data on the incorporation of labeled amino acids in free and protein-bound form as possible precursors of cardiac muscle protein grown in tissue culture (12). The amino acids of the proteins did not exchange with large pools of the corresponding unlabeled acid in the medium, and from this and from experiments with doubly-labeled proteins it was concluded that proteins could be transferred from a nutrient embryo extract medium to heart muscle protein without release of free amino acids. Tracer experiments of this sort, as will be discussed later, do not, however, prove the direct transfer 126 H. R. Mahler, H. Walter, A. Bulbenko and D. W. Allmann of protein, but solely suggest that there may not be free equilibration between the free added amino acid pool and amino acids formed and utilized metaboli- cally during precursor protein breakdown and product protein formation respectively. Another potentially very fruitful line of investigation is provided by some experiments of Ebert's, the results of which tentatively suggest the incorporation of organ specific adult proteins into those of embryos subsequent to chorio- allantoic grafts of the donor organs (3, 13). These researches were the out- growth of findings by Murphy (14) and by Danchakoff (15), made some forty years ago, that such transplants of adult chicken spleen lead to a specific enlargement of the host organs. A systematic re-investigation of the phenomenon by Weiss led to the conclusion that transplants of kidney and liver, as well as injections of organ breis of six-day old chick embryos into four-day old hosts, could lead to similar effects (2). Weiss correctly pointed out that experi- ments of this sort did not permit a choice between a 'template' or a 'specific precursor' type of mechanism. Ebert's investigations are designed to shed some light on this question as well as on the more general ones of protein synthesis and organ specific growth control in embryonic development. In our own investigations we have made use of S^^-labeled organ homo- genates, isolated proteins, peptides, and amino acids to gain some insight into the pattern of embryonic protein biosynthesis. In this work we have been interested not only in the immediate but also in the original precursors, which in this case must consist of all or part of the egg white and yolk proteins. Preliminary accounts of some aspects of this work have appeared (16). II. METHODS AND RESULTS 1 . Preparation oj Labeled Precursors In the experiments to be reported in tliis and subsequent sections S^^- labeled proteins, peptides, and a mixture of amino acids were prepared bio- synthetically as follows: Torulopsis utilis was grown on S^^-sulfate (obtained from Oak Ridge National Laboratory), according to Wood and Perkinson. (17) After extraction with organic solvents (18) the yeast protein was hydrolysed with a 1 :1 mixture of 6N HCl and 90 per cent fomiic acid. Humin was removed by centrifugation and a portion of the neutralized hydrolysate, which also served as source of amino acids in the experiments to be reported, corresponding to 50 mc of the original S^^, was injected intraperitoneally into a laying White Leghorn hen in two doses, about five hours apart. Eight hours after the second injection the blood was withdrawn by heart puncture, allowed to clot, and serum albumin and serum globulin prepared (19). The oviduct was removed from the hen, and ovalbumin prepared essentially as described by Steinberg and Anfinsen (11). All proteins were treated with cysteine at a pH of 8.0 to 8.5 to assure removal of exchangeable S^^, and then dialysed. Peptides were prepared by peptic hydrolysis of the proteins. Aliquots of the radioactive amino acids, peptides, and proteins were prepared by standard methods and counted. In the tracer experiments, 0.05 to 0.1 ml aliquots of the radioactive precursor solutions, containing 0.3 to 1.8 mg and 6000 to 25,000 counts per minute each, were injected into the yolk or the albuminous portion of some two Specific Mechanisms of Protein Synthesis in the Developing Chick Embryo 127 to three dozen unincubatcd, embryonated White Rock eggs. The punctures were sealed with paralTin wax and the eggs then incubated at 38° C under conditions of controlled humidity. Starting with the fifth and ending with the ninth day after the injection, embryos were harvested and a number pooled. The mixture was homogenized for about three minutes in a Potter-Elvehjem homogenizer in Ringer's isotonic saline solution, made up to 10 ml (fifth and sixth days) or 20 ml (seventh through ninth days), and precipitated with tri- chloracetic acid (final concentration, 8 per cent). Dry protein powders were then prepared and counted (20). 2. Is There Evidence for Selective Utilization of Egg-white or Yolk Proteins'} In the first set of experiments, chicken serum albumin injected into yolk or egg-white was used as a protein tracer. Table I shows the results of two Table L Injection of Chicken Serum Albumin into Embryonated Eggs njection Egg white Egg yolk Day after i % of injected activity found per embryo Protein wt of embryo in mg % of injected activity found per embryo Protein wt of embryo in mg 5 .006 .008 5.5 7 0.79 1.12 5 6.5 6 .012 .100 13 16 1.34 0.31 11 19 7 .015 .029 28 29 2.84 1.58 17 27 8 .016 45 4.04 3.35 43 48 9 .088 .133 72 79 2.86 7.28 53 87 series of experiments. The spread of the data is indicative of the precision, reliability, and reproducibility usually obtained in experiments of this sort. Let us now make the following assumptions: (a) that the injected protein is a true tracer for egg-white and yolk protein respectively, i.e. that no permea- bility or other pool barriers exist for its equilibration with the corresponding unlabeled egg proteins; and (b) that there is no selectivity in the uptake mechan- ism of the embryo either for or against a serum albumin tracer as a typical precursor protein. Now we can calculate data shown in Table II and compare the observed mean of the amount of protein actually formed, with that expected on the basis of the above assumptions. The latter value is calculated by multiplying the weight of total yolk or egg-white protein, about 3000 mg each, by the per cent of the injected activity incorporated per embryo (from Table I). There are profound discrepancies between the calculated and the observed 128 H. R. Mahler, H. Walter, A. Bulbenko and D. W. Allmann values. Those for the egg white are only a small fraction of those expected, while those for the yolk are uniformly about two-fold greater. It is thus apparent that at least one of the assumptions cited cannot be valid. The simplest modification would be to postulate that assumption (b) is not true, and that over the time-period studied egg white proteins are not precursors Table II. Amounts of Embryonic Protein Formed Compared to that Calculated from Tracer Data Protein (mg/embryo) Day after injection Observed Calculatec Egg-white 1* Yolk 5 6 0.21 28.8 6 15 1.68 24.9 7 29 0.66 66.3 8 45 0.48 111.0 9 76 3.30 152.0 * From injected albumin tracer. of embryonic proteins. Soluble proteins injected into the yolk can be utilized for this purpose, and may be more efficient than some of the yolk proteins proper. 3. Is There Evidence for Selective Utilization of Amino Acids, Peptides or Proteins ? In the next series of experiments we compared serum albumin, albumin peptides and amino acids all injected into the yolk, with the same precursors injected into egg white. The design of the experiment was the same as before and the results of one run are summarized in Table III, Table III. Incorporation of Protein Precursors into Chick Embryos* Day after Precursors injected into YOLK Precursors injected into egg-white injection albumin albumin amino albumin albumin amino peptides acids peptides acids 5 0.75 0.44 0.34 0.0063 0.35 1.19 6 1.30 0.90 1.53 0.013 0.56 3.03 7 2.80 1.70 3.86 0.015 1.59 3.48 8 4.05 4.72 5.15 0.016 2.32 4.94 9 2.85 8.52 9.18 0.088 5.94 5.65 * Expressed as per cent of injected activity recovered per embryo. We see that except for albumin injected into egg-white, which has already been discussed, all the precursors tested appear to be utilized with approxi- mately equal efficiency regardless of whether they are injected into the yolk or the egg white. This is not limited to serum albumin, but holds true equally well for serum globulin and ovalbumin and their peptides as is shown in Table IV. Specific Mechanisms of Protein Synthesis in the Developing Chick Embryo 129 Table IV. Incorporation into Embryos of Proteins and Peptides Injected into the Yolk* Day after injection S. albumin S. globu- lin Ovalbu- min S. albumin peptides S. globulin peptides Ovalbumin peptides 5 0.75 1.10 0.45 0.44 0.20 0.95 6 1.30 1.75 0.80 0.90 0.55 1.65 7 2.80 2.35 0.40 1.70 1.15 2.20 8 4.05 2.55 1.45 4.72 2.20 — 9 2.85 4.50 2.95 8.52 4.50 6.60 Expressed as per cent of injected activity recovered per embryo. 4. Is There Evidence for Organ-specific Transfer? In order to test the hypothesis of organ-specific transfer advanced by Ebert we have attempted to extend investigations of this sort to the use of S^^-labeled aduh chicken Hver and heart homogenates. These were prepared from deep-frozen organs of a White Leghorn hen injected with a mixture of S^^-amino acids, and treated as described above. After several months the tissues were thawed and homogenized in a tris- (hydroxymethyl)-aminomethane buflfer solution at pH 7.4 containing 0.9 per cent KCl, first in a Waring blender and then in a Potter-Elvehjem homogenizer. The liver and heart homogenates, made up to 10 per cent (weight/volume) with the same buffer solution, were then treated with cysteine at a pH of 8.0 to 8.5 to assure removal of all exchangeable S^^. After dialysis, some undis- solved material was removed by low-speed centrifugation, and the relatively clear supernatant fluid was used for intravenous injection into 9-day-old chick embryos. Embryonated White Rock eggs were incubated at 38° C under controlled humidity conditions for a period of 9 days. They were then candled, and the location of the blood vessels was marked on the shell of each egg. An area of about 1 cm^ of the shell above the vessel was carefully cut out by means of a dental drill and burr without injuring the membrane, and the small square was removed with a razor blade. A drop of mineral oil was placed on the membrane to render it transparent, and 0.1 ml of the liver or heart homogenate was intravenously injected in the direction of blood flow. The eggs were reincubated for 24 hours and the embryos were excised. Hearts and livers were removed, the organs were pooled, and homogenized; dry protein powders were prepared for counting as described before. Similarly aliquots of the homogenates used for injection were prepared and counted. The results of these experiments are given in Table V. In all, two series of experiments make up the Table. In the first series, twenty-four embryos each were injected with heart and liver homogenates; of these, twenty-two and eleven respectively survived. In the second series, forty-four out of forty-seven embryos injected with the heart preparation survived, while the number of survivors was twenty-two out of twenty-eight for the liver homogenate. Thus the table summarizes data obtained on 99 survivors out of 123 embryos that were injected: 66/71 for heart; 33/52 for liver. 130 H. R. Mahler, H. Walter, A. Bulbenko and D. W. Allmann It can be seen that the relative specific activity of hearts is higher than that of hvers when chicken heart homogenate is injected, whereas the relative specific activity of the livers is higher than that of hearts when chicken-liver homogenate is injected. Table V. Incorporation of Activity from Adult-Tissue Homogenates into Nine- Day Embryos after Twenty-four-hour Incubation Injection Item Chicken heart homogenat Chicken liver homogenate CoLint/min per embryo injected 398 398 2780 2780 mg injected per egg 0.1 0.1 0.1 0.1 Organs investigated Hearts Livers Hearts Livers Livers Hearts Livers Hearts No. of organs cut out 22 11 22 11 11 11 11 11 Dry protein wt of organs obtained (mg) 38.2 72.0 38.8 70.0 84.7 20.9 77.6 22.4 Wt counted (mg) 18.3 29.8 23.4 30.0 30.1 11.6 30.2 12.6 Count/min observed* 21 24 22 19 366 173 389 214 Corrected count/min per 30 mg 28 24 25 19 365 286 386 340 Relative specific activity 1.00 0.86 1.00 0.76 1.00 0.78 1.00 0.87 Counts per minute are within 5 per cent standard deviation. III. CONCLUSIONS The experiments on soluble protein tracers added to yolk and egg-white demonstrate quite clearly that proteins added to the egg-white or, probably, egg-white proteins themselves are incorporated with such low efficiency as to rule out any important contribution from this source to the protein of the developing embryo, at least up to and including the ninth day. Incorporation of protein from the yolk is rapid, and soluble proteins injected into this source may be utilized preferentially to some of the yolk proteins themselves. This utilization of yolk rather than egg-white proteins as a source of embryonic protein during this period is in accord with other investigations, notably the quantitative protein depletion studies of Rupe and Farmer (21). For the intervals studied, amino acids, peptides and proteins, even those of relatively 'foreign' origin such as the serum proteins, all apparently provide an equally acceptable source of S^^ for embryonic protein synthesis (within an order of magnitude or so), provided they are injected into the yolk. Now the protein tracer must be diluted by at least a portion of the 3.0 g or so of yolk protein — an estimate of approxim.ately 50 per cent would appear reasonable in view of the results reported above. On the other hand, amino acids or peptides cannot be diluted to any appreciable extent since the pools of these substances in the egg are vanishingly small (22). From this one might conclude that proteins themselves or substances easily formed from them must be the preferred precur- sors of embryonic proteins. Since the egg protein ovalbumin is used no more efficiently than the more "foreign" serum proteins, the pathways of assimilation for these precursors, available to the embryo, must have at least some inter- mediates in common. The data on peptides may find a similar interpretation. Specific Mechanisms of Protein Synthesis in the Developing Chick Embryo 131 These intermediates are not free amino acids, as evidenced by their relatively low incorporation rates. They may be small peptides or activated forms of amino acids, formed readily and reversibly from protein precursors, but not identical and not in equilibrium with the pool of added low-molecular weight precursors. This view would be in accord with the findings of Francis and WiNNiCK (12), although not with their interpretation. The occurrence of pools of modified amino acids, incapable of equilibrating with those in the medium, has been demonstrated in micro-organisms. Thus Gale, working with Staphylococcus aureus, found that added glutamic acid could be so trans- formed, and the modified fonn used for protein synthesis (24). Similarly CowiE and Walton (25) have presented evidence that the pools of amino acids formed metabolically in Torulopsis utilis and utilized as effective precur- sors in protein synthesis, are present in some modified form, possibly as com- plexes adsorbed onto macromolecules, and do not equilibrate freely with added amino acids in the medium. In all the cases presented, this metaboli- cally active form of the amino acids may be formed by a variety of pathways as indicated below. Proteins 1 [Peptide Intermediates] y >' 1 Free peptides ^'Amino Acids'-^ Free amino acids (modified) Recent investigations, especially by Zamecnik and his collaborators, (26) have disclosed that free amino acids are first 'activated' by enzymes in the soluble portion of the cytoplasm (27), probably through mixed anhydride formation with adenylic acid (27, 29, 30) prior to their incorporation into a protein-bound form (30, 31), which takes place in RNA-rich granules associated with the microsomal fraction of homogenates (32, 33, 34). Whether or not the metabolically active form of amino acids alluded to above can be equated with these aminoacyl adenylates has not yet been established. An alternative explanation, which has been invoked to account for apparent preferential utilization of proteins over amino acid precursors in the formation of specific proteins, postulates proteolysis and protein synthesis sites in such close spatial juxtaposition as to permit ready transfer of intermediates from breakdown to synthesis site at the expense of penetration of the latter by added amino acids. This has been suggested by Loftfield and Harris (10) as the mechanism operative in ferritin synthesis, and by Walter et al. (20) in the transformation of serum into organ proteins. Purely spatial factors of this sort are probably not the determining ones in the present instance, since it can be demonstrated that the bulk of the proteolytic activity is centred in the yolk (23), and thus remote from the synthetic activity which is, presum- ably, occurring in the embryo itself. It is hoped that critical experiments now in progress will permit a choice to be made between the various alter- natives suggested. 132 H. R. Mahler, H. Walter, A. Bulbenko and D. W. Allmann We have shown that the organ-specific locahzation phenomenon, previously observed with chorio-allantoic transplants, can be dupHcated by the injection of homogenates of aduh tissue. Similarly Tumanishvili et al. (35) found almost simultaneously that host organ enlargement could also be elicited by the same technique. This demonstration of the essential similarity of two approaches clears the way for an investigation of the problem by means of relatively straightforward biochemical and enzymological techniques rather than the more demanding ones of experimental embryology. Obviously only a bare beginning has been made. The findings will have to be confiiTned and extended and several relatively trivial explanations excluded. Among such explanations are, for instance, the transfer of whole cells on the one hand, and differential composition and/or incorporation rates with respect to cystine and methionine in the two tissues studied, on the other. Ebert claims to have eliminated both these alternatives in his transplantation experiments; in the light of the available information, they are not very likely in the present case. Nevertheless they will have to be rigorously excluded. Our tentative interpretation of the prelimi- nary results described is identical with that advanced by Ebert: that we are dealing with a specific transfer of rather large units from the donor preparation to the embryonic organ. Preliminary experiments indicate that the injection of either heart or liver (donor) homogenates leads to an increase in specific activity in the liver as compared to the heart. The effect in this case is therefore non-specific and possibly related to the higher mitotic and synthetic activity of liver relative to heart, i.e. to fuller differentiation. Another line of approach which promises to be of some interest is to determine the cell fraction or fractions, if any, responsible for eliciting the effect both with respect to the donor and the acceptor organ. Impetus is added to this approach by the recent experiments which have focussed attention on the soluble and microsomal fractions as being involved in the initial phases of protein synthesis. In preliminary experiments with fractionated, dialysed heart homogenates the data of Table VI were Table VI. Transfer of Label from Donor Heart Fractions into Organs of Recipient Embryos Fraction Relative specific activity of embryonic organs (heart/liver) Homogenate Nuclei Mitochondria Microsomes Soluble 1.17, 1.32, 1.23 0.65, 0.74 0.22 (?) 2.56 1.85, 2.50, 1.49 obtained. The number of data in each row corresponds to the number of experiments actually performed. Thus the results for the microsomal and mitochondrial fractions must be regarded as exceedingly tentative. With this proviso, components of the soluble fraction of the cytoplasm might be regarded as responsible for the phenomenon observed with whole heart homogenates. Specific Mechanisms of Protein Synthesis in the Developing Chick Embryo 133 A similar observation has been reported by Kutsky who found the supernatant fraction of embryo extract to be most active in stimulating the growth of heart fibroblasts in vitro (36). REFERENCES 1. For a very stimulating and up to date review the reader is referred to: B. Ephrussi: Enzymes in Cellular Differentiation, in: O. H. Gaebler: Enzymes: Units of Biological Structure and Function, Academic Press, New York City, 29-40 (1956). 2. For an intriguing hypothesis involving control of organ development by diffusible com- ponents see the following articles: P. Weiss, in A. K. Parpart: The Chemistry and Physiology of Growth, Princeton University Press, 135-186 (1949). P. Weiss: Self regulation of organ growth by its own products. Science 115, 487-488 (1952). P. Weiss: Some introductory remarks on the cellular basis of differentiation. /. Embryol. Exp. Morph. 1, 181-211 (1953). 3. The present state of knowledge, with special emphasis on protein biosynthesis, is admirably reviewed in the following: J. D. Ebert: Some aspects of protein biosynthesis in develop- ment, in: D. Rudnick: Aspects of Synthesis and Order in Growth, Princeton University Press, 69-112(1956). 4. S.Spiegelman: On the nature of the enzyme-forming system, />;; O. H. Gaebler: Enzymes: Units of Biological Structure and Function, Academic Press, New York City, 67-89 (1956). 5. R. S. Alcock: The synthesis of proteins in vivo. Physiol. Rev. 16, 1-18 (1936). J. H. Northrop, M. Kunitz, and R. M. Herriott: Crystalline Enzymes, 2nd ed., Columbia University Press, New York (1948). S. B. KoRiK and H. Chantrenne: The relationship of ribonucleic acid to the in vitro incorporation of radioactive glycine into the proteins of reticulocytes. Biochim. Biophys. /^cfa 13,209-215 (1954). 6. D. S. HoGNESS, M. CoHN, and J. Monod: Studies on the induced synthesis of /3-D- galactosidase in Escherichia coli. The kinetics and mechanism of sulfur incorporation. Biochim. Biophys. Acta 16, 99-116 (1955). 7. P. N. CAiMPBELL and T. S. Work: The biosynthesis of protein. Uptake of glycine, serine, valine and lysine by the mammary gland of the rabbit. Biochem. J. 52, 217-227 (1952). B. A. AsKONAS, P. N. Campbell, and T. S. Work: The biosynthesis of proteins. Synthesis of milk proteins by the goat. Biochem. J. 58, 326-331 (1954). B. A. Askonas, p. N. Campbell, C. Godin, and T. S. Work: Biosynthesis of Proteins. Precursors in the synthesis of casein and )3-lactogIobulin. Biochem. J. 61, 105-115 (1955). C. Godin, and T. S. Work: Biosynthesis of proteins. The effect of intravenous peptides on casein synthesis in a lactating goat. Biochem J. 63, 69-71 (1956). 8. M. V. Simpson and S. F. Velick: The synthesis of aldolase and glyceraldehyde-3- phosphate dehydrogenase in the rabbit. /. Biol. Chem. 208, 61-71 (1954). M. V. Simpson: Further studies on the biosynthesis of aldolase and glyceraldehyde-3- phosphate dehydrogenase. J. Biol. Chem. 216, 179-183 (1955). 9. M. Heimberg and S. F. Velick: The synthesis of aldolase and phosphorylase in rabbits. J. Biol. Chem. 208, 725-730 (1954). S. F. Velick: The metabolism of myosin, the meromyosins, actin and tropomyosin in the rabbit. Biochim. Biophys. Acta 20, 228-236 (1956). 10. R. B. Loftfield and A. Harris: Participation of free amino acids in protein synthesis. /. Biol. Chem. 219, 151-159 (1956). 11. D. Steinberg and C. B. Anfinsen: Evidence for intermediates inovalb umin synthesis. /. Biol. Chem. 199, 25-42 (1952). C. B. Anfinsen and D. Steinberg: Studies on the biosynthesis of ovalbumin. /. Diol. Chem. 189,739-744(1951). 10 134 H. R. Mahler, H. Walter, A. Bulbenko and D. W. Allmann M. Vaughan and C. B. Anfinsen: Nonuniform labeling of insulin and ribonuclease synthesized in vitro. J. Biol. Chem. Ill, 367-374 (1954). M. Flavin and C. B. Anfinsen: The isolation and characterization of cysteic acid peptides in studies of ovalbumin synthesis. J. Biol. Chem. Ill, 375-390 (1954). D. Steinberg, M. Vaughan, and C. B. Anfinsen: Kinetic aspects of assembly and degradation of proteins. Science 124, 389-395 (1956). 12. M. D. Francis and T. Winnick: Studies on the pathway of protein synthesis in tissue culture. /. Biol. Chem. 202, 273-289 (1953). 13. J. D. Ebert: The effects of chorioallanteic transplants of adult chicken tissues on homo- logous tissues of the host chick embryo. Proc. Nat. Acad. Sci., Wash. 40, 337-347 (1954). 14. J. B. Murphy: The effect of adult chicken organ grafts on the chick embryo. /. Exp. Med. 24, 1-6 (1916). 15. V. Danchakoff: Equivalence of different hematopoietic anlagen (by method of stimula- tion of their stem cells). II. Grafts of spleen on the allantois and response of allantoic tissues. Amer. J. Anat. 24, 127-189 (1918). 16. H. Walter, A. Bulbenko, and H. R. Mahler: Precursors of embryonic chick proteins. Nature, Lond. 178, 1176-1177 (1956). H. Walter, D. W. Allmann, and H. R. Mahler: Influence of adult tissue homogenates on formation of similar embryonic proteins. Science 124, 1251-1252 (1956). 17. J. L. Wood and J. D. Perkinson Jr. : Yeast biosynthesis of radioactive sulfur compounds. J. Amer. Chem. Soc. 74, 2444-2445 (1952). 18. R. B. Williams and R. M. C. Dawson: The biosynthesis of L-cystin and L-methionine labeled with radioactive sulphur S^\ Biochem. J. 52, 314-317 (1952). 19. W. Friedberg, H. Walter, and F. Haurowitz: The fate in rats of internally and exter- nally labeled heterologous proteins. /. Immunol. 75, 315-320 (1955). 20. H. Walter, F. Haurowitz, S. Fleischer, A. Lietze, H. F. Cheng, J. E. Turner, and W. Friedberg : Metabolic fate of injected homologous serum proteins in rabbits. /. Biol. Chem. 224, 107-119 (1957). 21. C. O. RuPE and C. J. Farmer: Amino acid studies in the transformation of proteins of the hen's egg to tissue proteins during incubation. /. Biol. Chem. 213, 899-906 (1955). 22. A. L. Romanoff and A. J. Romanoff: The Avian Egg. J. Wiley and Sons, New York City. (1949). 23. A. L. Romanoff: Membrane growth and function. Ann. N.Y. Acad. Sci. 55, 288-301 (1952). 24. E. F. Gale: Assimilation of amino acids by gram-positive bacteria and some actions of antibiotics thereon. Advances in Protein Chemistry 8, 285-391 (1953). 25. D. B. CowiE and B. P. Walton: Kinetics of formation and utilization of metabolic pools in the biosynthesis of protein and nucleic acid. Biochim. Biophys. Acta 21, 21 1-226 (1956). 26. M. D. Hoagland, E. B. Keller, and P. C. Zamecnik: Enzymatic carboxyl activation of amino acids. /. Biol. Chem. 218, 345-358 (1956). 27. P. SiEKEViTz: Uptake of radioactive alanine in vitro into the proteins of rat liver fractions. /. Biol. Chem. 195, 549-565 (1952). 28. J. A. DeMoss and G. D. Novelli: An amino acid dependent exchange between inorganic pyrophosphate and ATP in microbial extracts. Biochim. Biophys. Acta 18, 592-593 (1955). 29. P. Berg and G. Newton: Acyl adenylates : the interaction of adenosinetriphosphate and L-methionine. /. Biol. Chem. 222, 1025-1034 (1956). 30. E. B. Keller and P. C. Zamecnik: The effect of guanosine diphosphate and triphosphate on the incorporation of labeled amino acids into proteins. /. Biol. Chem. 221, 45-59 (1956). 31. H. Sachs and H. Waelsch: The effect of pyrophosphate on amino acid incorporation into rat liver microsomes. Biochim. Biophys. Acta 21, 188-189 (1956). Specific Mechanisms of Protein Synthesis in the Developing Chick Embryo 135 32. J. W. LiTTLEFiELD, E. B. Keller, J. Gross, and P. C. Zamecnik: Studies on cytoplasmic ribonucleoprotein particles from the liver of the rat. J. Biol. Chem. 217, 111-123 (1955). 33. M. V. Simpson and J. R. McLean: The incorporation of labeled amino acids into the cytoplasmic particles of rat muscle. Biocliim. Biophys. Acta 18, 573-575 (1955). 34. G. C. Webster and M. P. Johnson: Effect of ribonucleic acid on amino acid incorpora- tion by a particulate preparation from pea seedlings. /. Biol. Chem. 217, 641-649 (1955). 35. S. TuMANisHviLi, K. M. Djandier, and I. K. Skanidze: Specific stimulation of the growth of the chicken embryo by the effects of tissue extracts. C.R. Acad. Sci., U.R.S.S. 106, 1107-1109(1956). 36. R. J. Kutsky: Growth stimulating effects by nucleoprotein and cell fractions on chick heart fibroblasts in vitro. U.S. Atom. Energ. Comm., U.C.R.L. No. 2270 (1953). DISCUSSION Quastler: It is useful to compare the informational requirements of various alternative methods of protein synthesis. If the whole protein is synthesized directly from amino acids, then each locus on the template must carry sufficient information to specify a single amino acid, or approximately four bits; this is well within the informational capacities of chemical reactions. If the incor- poration occurs in two steps, as has been suggested, then each step might have to specify no more than two bits. If the protein is synthesized from peptide chains, then the informational requirements are much more stringent. Consider the linking of two peptide chains of, say, five amino acids each. If each of the ten amino acids can be any one of the whole set of amino acids, then the linking operation must, in some way, identify ten amino acids, for a total of about forty bits — which is a very large amount of information to be processed in a single act. The requirements are greater — in fact, almost certainly too great — if two chains of ten amino acids are to be linked. The following possibilities exist which allow the use of large fragments without imposing high informational requirements : (a) the terminal amino acid in a chain identifies automatically the other members — this would imply very strong sequential dependencies within peptide chains, and consequently a low informational capacity of the whole amino acid sequence; (b) linkages are formed without reference to the nature of residues remote from the locus of linkage, and the resulting proteins are torn down again if not functional — in this case, the probability of producing functional sequences by chance is small, and the efficiency of protein synthesis is low; or (c) the protein studied is such that the exact sequence of residues is irrelevant. THE MECHANISM OF ACTION OF METHYL XANTHINES IN MUTAGENESIS Arthur L. Koch Department of Biochemistry, College of Medicine, J. Hillis Miller Health Center, University of Florida, Gainesville, Florida Abstract — The biochemical findings relating to the action of methyl xanthines on bacteria and bacterial extracts have been reviewed. These observations, together with those of Novick and SziLARD on the mutagenic activity of these substances, have suggested that the biological action results from an inhibition of enzymes of nucleic acid biosynthesis. Consequences of this hypothesis have been discussed relative to the regulation of growth of cell constituents. Alternative hypotheses are enumerated. I. INTRODUCTION A NUMBER of agents, both chemical substances and radiations, cause mutations. One particular class appears to be potentially most fruitful in an attempt to understand the genetic replication process. This class includes purines and related compounds. Particularly important are the plant alkaloids responsible for the pharmacological effects of coffee, tea and cocoa. If these substances are added to a continuously growing culture of bacteria, the mutation rate is caused to increase markedly (1, 2). If we compare the structures (Fig. 1) of the alkaloids caffeine, theobromine and theophylline, with the purine bases normally present in nucleic acids of Me N- = C C- ,/ Me Me N C- CAFFEINE H N- Me N- Me > THEOBROMINE Me N C-= 0 I I SI Me- -N C- THEOPHYLLINE XANTHINE -NH, -OH H2N C / ADENINE Fig. 1 . The structure of purine derivatives GUANINE all species, adenine and guanine, the similarity is readily apparent. The former are methyl derivatives of xanthine, the latter amino and deoxy derivatives of xanthine. It is tacitly assumed that these agents are mutagens because of this similarity. 136 The Mechanism of Action of Methyl Xanthines in Mutagenesis 137 II. TRACER STUDIES The first possibility to test was that these compounds, or products derived from them, are utihzed for the synthesis of the nucleic acid of the host (3). To do this we prepared these substances as well as some others, labeled with carbon 14 in the 8-position of the heterocyclic nucleus. These were then added to growing cultures o^ Escherichia co// under conditions similar to those employed by NoviCK and Szilard (I, 2) in their studies. In Table I the data so obtained are presented. Adenine and guanine as well as the deaminated derivatives are very well incorporated into the nucleic Table I. Incorporation and Mutagenicity oj Various Purines RSA* of DNA purines Mutagenicity Adenine 0.3 + Guanine 0.20 ± Hypoxanthine Xanthine 0.30 0.20 — Theobromine 0.00002 + + + Caffeine 0.00001 + + + Theophylline 0.00001 + + + *RSA = relative specific activity = ratio of the specific activity of the purine isolated from the bacteria to that of the growth medium. acids of both the RNA and DNA type, whereas all methylated substances are incorporated only to a very small extent, if at all. On the other hand, the correlation of mutagenesis is the reverse. A mutation is a very rare event, and though these agents, when present in quite high concentration, may raise the mutation rate by a factor of fifteen or so, this still only corresponds to one event in 10'^ duplications. The small amount of radioactivity that is found associated with the DNA from cells grown in the presence of radioactive mutagens is probably experi- mental contamination. However, although these experiments are technically excellent, they cannot begin to exclude the possibility that a methylxanthine molecule is incorporated into the DNA molecule in the process of the rare mutational event itself, since the resultant incorporation for one locus would be many orders of magnitude below the trace amount observed here. Considera- tion of the structures of these substances, however, makes this possibility rather unlikely. In the formation of the normal 9-A^-riboside or 9-A^-deoxyriboside linkages, the single replaceable hydrogen which may be in either the 7- or 9-position is replaced by the glycosyl residue. In the case of caffeine or theobromine, which are 7-methyl derivatives, this is not possible because of the prior replacement of the hydrogen by the methyl group. Thus even though the methyl group is 138 Arthur L. Koch attached to the 7-position it prevents bond formation at the 9-position. Con- sequently, the methyl group must be removed if the molecule is to be incor- porated into the nucleic acids. The isotopic data, as well as other information, are adequate to demonstrate that there is not a single molecule of enzyme present in these bacteria capable of removing this methyl group (3). Therefore it would appear that certain of the mutagenic materials are not and cannot be converted into a form in which they can be linked covalently to cell materials, not at least by the 9-A'^-glycosyl bond which has been universally found in biological materials. III. PURINE METABOLISM IN ESCHERICHIA COLI The next possibility we investigated was that the mutagens act by inter- fering with nucleic acid biosynthesis. First, however, it is necessary to discuss the metabolism of the organism under study. Fig. 2 summarizes, from the Adenine Hypoxanlhine and derivatives Glycine CO, ''"'C '" ^^r Serine _- / \ 0" DR Glucose Formate Ammonia J ) _ "GR - '- G « ' 6D Guonine Xanthine and derivatives CELL WALL AD ♦ONA Fig. 2. The purine metabolism of Escherichia coli available tracer data, the pathways of purine synthesis in growing cultures of the test organism (4, 5, 6, 7, 8). C^*-labeled COg (4), glycine (8), and serine or formate (unpublished) lead to the formation of RNA adenine, DNA adenine, RNA guanine, and DNA guanine, all of equal specific activity. The activity in the purines derived from CO2 and glycine is such as to indicate that the well-accepted scheme for purine biosynthesis is the major pathway in tliis organism (4). C^Mabeled adenine and hypoxanthine and their derivatives yield adenine samples of equal, but lower, specific activity in both RNA and DNA. From these facts it is inferred that there are three pools at which purine metabolism branches, namely, a 'purine' pool which is common to all cellular purines, and an 'adenine' and a 'guanine' pool which are precursors of the corresponding purine in both types of nucleic acid. So far, attempts to find a precursor which enters purine metabolism at some point beyond the 'adenine' or 'guanine' pool have failed. Even when the intracellular adenine-C^"* ribo- nucleotides were specifically labelled (5), the incorporation into the purines of the ribose nucleic acid was equal to that in the deoxyribose nucleic acids. It should be mentioned that in organisms under conditions of rapid growth, the soluble intermediate pool concentrations relevant to this scheme are small (5). It was impossible to demonstrate guanosine, adenine deoxyriboside, guanine deoxyriboside or phosphorylated derivatives. The Mechanism of Action of Methyl Xanthines in Mutagenesis 139 Although the tracer data dehneate the pathways, they do not define the intermediates. It is, however, possible to conclude from available enzyme data that 'adenine' and 'guanine' pools are made up at least in part of the free bases themselves. This follows from the fact that the known enzymes of purine metabolism which might be involved in the conversion of the hypothetical 'purine' precursor to the two types of nucleic acids catalyze reactions involving the free purine base. The purine nucleoside hydrolases, purine nucleoside phosphorylases, purine #-trans-glycosidases, and purine nucleotide pyro- phosphorylases yield the free purine base. These enzymes and the postulated pathway of direct reduction of the riboside to the deoxyriboside constitute the only pathways of interconversion of ribose and deoxyribose purine com- pounds that can be imagined at present. Since the reductive pathway is known not to occur in E. coli (9) (although the interesting work from Volkin's labora- tory may be relevant (10)), it appears quite likely that the free purine base is involved in the 'adenine' and 'guanine' pools. In addition to these general considerations, the specific observation of Lampen and Manson (11) that purine deoxyriboside phosphorylase is inhibited by adenine led us to investigate the inhibition of phosphorylases of E. coli by methyl xanthines. IV. ENZYMATIC INHIBITION STUDIES The main conclusion from these studies (12, 13) was that the organism possesses enzymes, particularly nucleoside phosphorylases of both types (ribose and deoxyribose), that are inhibited by purines generally but specifically by the mutagenic substances. It was also found that even in the presence of 10 20 CAFFEINE CONC. (mM) 30 5 10 20 CAFFEINE CONC. (mM) Fig. 3. The inhibition of purine nucleoside phosphorylase. The effect of caffeine concentration on the arsenolysis of adenine riboside is shown at the left, and on adenine deoxyriboside on the right. The systems contain arsenate to prevent the complication of back reaction. large amounts of inhibitors enzyme action was not completely repressed (Fig. 3). In all cases this suggested the presence of more than one enzyme catalyzing the reaction under study. Studies of the effect of pH and the separa- tion of the bacteria into several chemical fractions supported this notion. 140 Arthur L. Koch The activity in various fractions was differently affected by caffeine and this effect was different in acid and at neutrahty and at alkaline reaction (see Table II). This finding explains the relatively low toxicity and bacteriostatic power of the plant alkaloids. Table II. Inhibition of Inosine Arsenolysis by Caffeine Enzyme preparation* No. Inhibition produced by 10 [J. moles caffeine per ml Distribution of activity (measured at pH 7) pH5.0 pH 7.0 pH 9.0 6-1 (soluble) 6-2 (particulate) 6-3 (phosphate extract) per cent 29 64 78 per cent 59 97 78 per cent 35 46 6 per cent 67 17 16 TOTAL 100 * Enzyme Preparation 6-1 was most active at pH 5, preparation 6-2 at pH 9, and preparation 6-3 at pH 7. In more recent work (13) three new enzymes have been demonstrated in extracts of this organism: an inosine hydrolase, a purine-pyrimidine trans- ribosidase, and a purine-purine transribosidase. All are inhibited to some degree by various purines. The results of the enzymatic studies are summarized in Table III. Table III. Enzymes of Nucleic Acid Metabolism Type Specificity Inhibition by methyl purines Adenosine deaminase Ribose 0 Cytidine deaminase Deoxyribose Ribose 0 0 Purine phosphorylases Pyrimidine phosphorylase Deoxyribose Ribose Deoxyribose Ribose 0 some some 0 Inosine hydrolase Purine-pyrimidine trans- Deoxyribose Ribose Ribose 0 + glycosidase Purine-purine trans- Ribose -f glycosidase V. WORKING HYPOTHESIS The mutagenic agents do inhibit enzymes that appear to be directly linked to the path of nucleic acid synthesis, but how can such an interference affect the The Mechanism of Action of Methyl Xanthines in Mutagenesis 141 mutation probability? We have proposed (12) that this may result from a change in the steady slate concentrations of the intennediates that are to be assembled together to form the macromolecular DNA. This must happen without any change in the flow of intermediates, in accord with the experimental fact that the growth rate of the bacteria is not affected significantly by the mutagens when present at concentrations that give rise to large changes in the mutation rate (1). Let us first consider the consequences of lowering of the concentration of whatever adenine deoxyriboside or guanine deoxyriboside derivative is involved in the polymerization reaction leading to macromolecular DNA. The Watson-Crick model for DNA assumes that the specificity lies in the forma- tion of two or three hydrogen bonds between specific pairs of nucleotides: adenine and thymine, and guanine and cytosine. It has been suggested by Watson and Crick (14) that the mutational event is the entry of a heterocylic base which is not complementary. This would yield a double helix which is energetically less stable. Upon subsequent duplication this yields two stable molecules, one of the parental type and one of a new mutant type. guanine thymine Fig. 4. It is to be recognized that the mutational event is an improbable one, and therefore quite improbable structures may be involved. Two options for the unfavorable pairing are available. First, two pyrimidines or two purines may become situated opposite each other. This gives structures that should be capable of forming hydrogen bonds, but are either too long or too short. Alternatively, a purine and a pyrimidine may pair, but the purine may occur in the uncommon tautomeric form and consequently pairing will occur abnor- mally. Watson and Crick (14) suggested adenine in the lactim form binding with cytosine, more probable is the pairing of guanine with thymine (Fig. 4). This pair has the proper dimensions; there are no steric difliculties. In this structure guanine is written with the oxygen in the 6- position in an enol form. X-ray-diffraction workers have concluded that guanine is ordinarily found in the keto form, but the evidence is not strong that the keto form is even dominant (15), and considerations of the resonance possibilities indicate a considerable stabilization of the enol fonn because the latter allows aromaticity of the heterocyclic ring. Thus, guanine-thymine pairing might well be of likely occurrence. With this in mind, we have attempted in our enzyme studies to find differences of the effects of mutagens on the inhibition of reactions of the adenine compounds, as opposed to the guanine ones, that would be implied if this structure were to 142 Arthur L. Koch account for the mutational activity of these methylated purines. So far we have been unable to detect any such differences. We may have been examining the wrong systems. For the present we shall tentatively suggest the pair thymine-cytosine (Fig. 5) as the culprit. This pair is shorter than the conventional structures. In the very interesting paper by Donohue (16) a large number of possible pairings are suggested. For our purposes most of these are unsatisfactory because they give rise to helices possessing a two-fold axis parallel to the hehcal axis, whereas thymine cytosine Fig. 5. in the Watson-Crick structure this two-fold axis is perpendicular to the helical axis, and thus consistent hehces formed by substitution between the two types can not occur. One structure (Donohue's no. 22) would fit into the symmetry of the Watson-Crick model and it is the pairing suggested in Fig. 5. VI. STEADY-STATE CONSIDERATIONS Whatever may be the critical or quantitatively most significant substitution in this type of mutational change, the hypothesis we have proposed requires that the concentration of terminal pools be altered. The experimental data that we have obtained have been primarily with purine ribonucleoside phos- phorylase which catalyzes a step which is clearly non-terminal in DNA synthesis, and very likely the reaction catalyzed by purine deoxyriboside phosphorylase is also not the transformation of the last small-molecular-weight intermediate into DNA. Although it may be that the terminal processes are inhibited, let us examine some possible situations that might lead to an alteration of the steady-state concentration of the penultimate substance without influencing the steady-state flux of DNA synthesis. To do this, the question of bacterial growth itself must be raised. Bacteria grow autocatalytically. Hinshelwood (17) as well as others have pointed out that this results from an interaction of catalytic units. Thus, if the amount of one component, P (protein), controls the rate of synthesis of another component, N (nucleic acid), then dP dt dN ~di k,P (1) The Mechanism of Action of Methyl Xanthines in Mutagenesis 143 where k^ and k^ are characteristic constants. The steady-state solution of this pair of equations is P = p (,\Vki^k^l \ (2) where P^) and A^,, depend on the initial conditions and the constants Aj and ko. Thus both P and A^ increase exponentially at the same rate and each therefore appears to be 'autocatalytic'. Clearly, processes of this kind are responsible for the maintenance of constant growth rates and constant composition of cells during the exponential growth of bacteria. However, the control of the system by this type of interaction cannot explain the regulation of synthesis of intermediates for the biosynthesis of either P or A'^. Additional regulatory processes must be considered. From equation (2) it is evident that for any constituent of the cell (intermediate or enzymatic catalyst) the steady concentration increases autocatalytically. If expressed as amount per unit number of bacteria or per unit bacterial mass, any cell constituent may be considered constant. Thus, if such a transformation is made, we can consider a system with time-invariant concentrations of inter- mediates and catalysts and also time-invariant fluxes. Thus, the steady-state treatment of reaction rates is immediately applicable to our problem. The most general formulation is that of Christiansen and has been well described by Hearon (18, 19). In essence the rate expression for each step of a concatenated reaction scheme, in which a substance is produced in one step and utilized in the next, is written down. Each of the terms in these expressions is the product of the intermediate with a rate constant and also with either unity or with the concentra- tion(s) of the other chemical reactant(s). If the product of the two latter factors is set equal to a quantity W, bearing suitable subscripts to identify the term, and if the usual steady-state assumptions are made, then the solutions for both the flux of the system or the over-all reaction rate v and the concentration of each intermediate [A'J may be computed. If the very last reaction is irreversible, equations (3) and (4) are obtained. W.W^W^"- W, [X,] = V (3) Wi-,1 '"W^ (4) The assumption of the irreversibility of the last step is made necessary by the well-known metabolic stability of DNA. Recent experiments (20) demon- strate the extreme irreversibility in the normal adult rat. The evidence for growing cultures of E. coU is less stringent (21, 22) but does permit this assumption in comparison with the tremendous synthetic rate in these organisms. 144 Arthur L. Koch Now if in addition we assume that some step is either rapid in the direction of synthesis or irreversible, then it may easily be seen that the reaction velocity V, is completely independent of subsequent steps. Thus, the synthetic rate can be made to depend on the level of a few catalysts or other reactants involved earlier in the sequence. Consequently, increased protein synthesis would cause increased synthesis of a very few enzymes critical for nucleic acid biosynthesis, and this would lead smoothly to increased DNA synthesis without requiring exact synchronization in the increase of each enzyme on the biosynthetic path- way. The concentration of the last intermediate i'j_i can be seen from equation (4) to be vlW^, and thus is completely independent of any step that has no effect on the reaction velocity, v. This case does not therefore satisfy the requirements suggested above to explain the mutagenic effects of the plant alkaloids. The independence of growth rate in the presence of caffeine could be explained simply by assuming that the inhibition occurs after some fast or irreversible reaction; but the action of the inhibitor on any but the final step has no effect on the concentration of the immediate precursor of the macromolecule, and thus cannot affect the probability of mutation. The scheme considered above has two desirable features: it permits a reciprocal control of nucleic acid by the level of protein synthesis, and it prevents the accumulation of large amounts of intermediates. Let us now turn to a possible mechanism that will do these two things but also will fulfill the conditions imposed by our ideas of the mutation event. Such a mechanism occurs in systems showing product inhibition. Here the rate of production of the final product will depend on the level of some enzyme catalyzing a step late in the reaction sequence, but at the same time, the inhibition prevents the unlimited synthesis of earlier intermediates. Product inhibition is of common occurrence. It has been suggested as having metabolic significance in two cases (23, 24) in which the product of a reaction sequence inhibits some earlier reaction than its own formation. In the present case it has been shown that adenine deoxyriboside is an inhibitor of the phos- phorylase (12) as well as purine bases. Let us assume that all of these agents are competitive inhibitors of enzyme action, although this remains to be demon- strated conclusively. Under such conditions the reaction velocity is given by the well-known expression for competitive inhibition (see, for example, (25)) K, K, + Ul) + US) where Kis the maximal velocity obtainable, K^ is the Michaelis-Menten constant for the substrate S, and Ki is the constant for the binding of the enzyme with the inhibitor, 1. If Kj{I) is the dominant term in the denominator, this expression simplifies to give: K.V(S) ... In the present case, adenine deoxyriboside is the inhibitor which is formed from the substrate adenine and deoxyribose-l-POj. Now, if the net rate of The Mechanism of Action of Methyl Xanthines in Mutagenesis 145 removal of adenine deoxyriboside is to be maintained constant and determined solely by the process of removal, then a steady-state will quickly ensue in which (S) oc (/), and in which the rate of formation of / is dependent only on the rate of utilization. The concentration of / will become adjusted to estabHsh such a condition. In the presence of the mutagen, the total inhibitor is effectively derived from three sources; deoxyribosides, free normal bases, and the mutagen. While maintaining constant synthesis of DNA, the effect of the mutagen will then be to decrease the level of the normal reaction product, adenine. Similar relations will hold for guanine deoxyriboside. It should be noted that in this case, although not in the case considered above, any number of intermediates may occur between the step under considera- tion and the polymerization step, if these reactions are rapidly reversible. Then a change in adenine deoxyribose concentration will lead to a proportional change in the precursor immediately used for the formation of the macro- molecule. This model can then utilize the enzymatic finding, and the biological facts. There is, however, one additional fact that should be introduced, viz. certain specific substances, the purine ribosides (26), are anti-mutagens. That is, these substances will prevent the action of caffeine and related compounds in causing mutations. Moreover, they will decrease the so-called 'spontaneous mutation' rate. This can be tentatively explained on the basis that these substances are substrates or immediate precursors of the substrates of the key step, and that their increase simply affects the system so as to cause an increase in the concen- tration of purine deoxyribosides and thus a decrease in the mutation rate. VII. ALTERNATIVE HYPOTHESES In concluding, I should like to list various hypotheses that one should consider in this type of chemical mutagenesis. They will be considered in order of the intimacy of the mutagen with the duplication process. 1. The mutagen is incorporated into the nucleic acid. This is tentatively rejected as indicated above, from the tracer evidence, and the argument that methylation in the imidazole ring prevents A^-glycoside formation. It should be noted that production of a self-duplicating 'methylated gene' can be rejected because the mutants cannot metabolize methyl purines and certainly do not require them (3). 2. The mutagens inhibit enzymes of nucleic acid biosynthesis, and this causes a change in the concentration of intermediates. This latter effect changes the probability of mutation. This is the hypothesis we favor, but it is clear that a great deal of work will be required to establish it or some variant thereof. It is also clear from what has been said above that special circumstances must occur in order that the proposed mechanism can work. 3. The mutagen causes some change in the general metabolism of the organism and this leads to a change in the mutation probability. It is certainly true that the mutation probability is dependent on a great many factors. Kihlman (27, 146 Arthur L. Koch 28), working with plants, has suggested such a mechanism to explain chromo- some breakage induced with caffeine derivatives. He proposes that ATP is necessary for the aberrations produced by the compound 8-ethoxy caffeine. However, there appear to be considerable differences between the two systems; with the bacteria one thinks the process involved is one of 'point mutation', but certain clearcut differences are evident in the two types of material with regard to the interaction ofoxygen tension and ionizing radiations. (Compare (2) and (27), 4. The mutagen causes the organism to 'adapt' to its presence, and thus causes widespread alterations in the amount of enzymes and intermediates. This could lead to a change in mutation rate. This may be in fact the explanation of the effect of adenine (12). This substance inhibits the growth of bacteria which have previously been grown in its absence. Growth resumes when the organism has 'adaptively' produced an 'adenine deaminase' activity which is not de- monstrable in bacteria grown in its absence. This shift in metabolism can then be envisioned to lead to changes in the mutation rate. This list is probably sufficiently inclusive to include the right answer if there is only one, but at least the necessary research, both with test tubes and with pencil and paper, to test these possibilities is feasible. REFERENCES 1. A. NoviCK and L. Szilard: Experiments on spontaneous and chemically induced mutations of bacteria growing in the chemostat. Cold Spr. Harb. Symp. Quant. Biol. 16, 337-343 (1951). 2. A. Novick: Mutagens and anti-mutagens. Brookhaven Symp. Biol. No. 8, 201-214 (1956). 3. A. L. Koch: The metabolism of methyl purines by Escherichia coli. I. Tracer studies. J. Biol. Chem. 219, 181-188 (1956). 4. A. L. Koch, F. W. Putnam, and E. A. Evans Jr.: The purine metabolism of Escherichia coli. J. Biol. Chem. 197, 105-112 (1952). 5. A. L. Koch: Biochemical studies of virus reproduction. XI. Acid soluble purine meta- bolism. /. Biol. Chem. 203, 227-37 (1953). 6. M. E. Balis, C. T. Lark, and D. Luzzati: Nucleotide utilization by Escherichia coli. J. Biol. Chem. 212, 641-645 (1955). 7. E. Bolton: Biosynthesis of nucleic acid in E. coli. Proc. Nat. Acad. Sci., Wash. 40, 764^772 (1954). 8. A. L. Koch: The kinetics of glycine incorporation by Escherichia coli. J. Biol. Chem. 217, 931-945 (1955). 9. I. A. Rose and B. S. Schweigert: Incorporation of C^* totally labeled nucleosides into nucleic acids. /. Biol. Chem. 202, 635-644 (1953). 10. E. Volkin and L. Astrachan: Phosphorus incorporation in Escherichia coli ribonucleic acid after infection with bacteriophage T2. Virology 2, 149-161 (1956). 11. J. O. Lampen: Symposium on Piwsphorus Metabolism, vol. II, ed. by W. D. McElroy and B. Glass, Johns Hopkins Press, Baltimore, 363-380 (1952). 12. A. L. Koch and W. A. Lamont: The metabohsm of methyl purines by Escherichia coli. II. Enzymatic studies. /. Biol. Cliem. 219, 189-201 (1956). 13. A. L. Koch: Some enzymes of nucleoside metabolism oi Escherichia coli. J. Biol. Cfiem. 223, 535-549 (1956). 14. J. D. Watson and F. H. C. Crick: The structure of DNA. Cold Spr. Harb. Symp. Quant. Biol. 18, 123-131 (1953). The Mechanism of Action of Methyl Xanthines in Mutagenesis 147 15. D.O.Jordan: The physical properties of nucleic acids. In: The Nucleic Acids, Cii. hy E. Chargaff and J. N. Davidson. Vol. I, 447-492, Academic Press, New York (1955). 16. J. Donohue: Hydrogen-bonded helical configurations of polynucleotides. Proc. Nat. Acad. Sci., Wash., 42, 60-65 (1956). 17. C. N. HiNSHELWOOD : The Chemical Kinetics of the Bacterial Cell, Clarendon Press, Oxford (1946). 18. J. Z. Hearon: The steady-state kinetics of some biological systems. I. Bull. Math. Biophys. 11, 29-50 (1949). 19. J. Z. Hearon: Rate behavior of metabolic systems. Physiol. Rev. 32, 499-523 (1952). 20. R. W. SwiCK, A. L. Koch, and D. T. Handa: The measurement of nucleic acid turnover in rat liver. Arch. Biochem. Biophys. 63, 226-242 (1956). 21. A. D. Hershey: Conservation of nucleic acids during bacterial growth. J. Gen. Physiol. 38, 145-148 (1954). 22. A. L. Koch and H. R. Levy: Protein turnover in growing cultures of Escherichia coli. J. Biol. Chem. Ill, 947-957 (1955). 23. R. A. Yates and A. B. Pardee: Control of pyrimidine biosynthesis in Escherichia coli by a feedback mechanism. /. Biol. Chem. Ill, 757-770 (1956). 24. H. F. Umbarger: Evidence for a negative-feedback mechanism in the biosynthesis of isoleucine. Science 123, 848 (1956). 25. P. W. Wilson: Kinetics and mechanism of enzyme reactions, in Respiratory Enzymes, ed. by H. A. Lardy, 22-67, Burgess Publishing Co., Minn. (1949). 26. A. NoviCK and L. Szilard: Anti-mutagens. Nature, Lond. 170, 926-927 (1952). 27. B. Kihlman: Chromosome breakage in Allium by 8-ethoxy-caffeine and X-rays. Exp. Cell Res. 8, 345-368 (1955). 28. B. A. Kihlman: Oxygen and the production of chromosome aberrations by chemicals and X-rays. Hereditas 41, 384-404 (1955). EVIDENCE FOR A NEGATIVE FEEDBACK SYSTEM CONTROLLING LIVER REGENERATION Andre D. Glinos Growth Physiology Laboratory, Walter Reed Army Institute of Research, Washington, D.C. Abstract — Cell division was induced in the resting liver of the rat by lowering the concentration of serum constituents through plasmapheresis, and was inhibited in the regenerating liver by increasing the concentration of the serum by fluid intake restriction. Electrophoretic analysis of serum proteins and histochemical investigation of the organiza- tion of cytoplasmic ribonucleoprotein of the liver cells during regeneration suggest that plasma proteins may participate as information-carrying agents in a negative feedback system controlling the growth of liver cells. Liver is an excellent tissue for investigating mechanisms of growth control because it regenerates very rapidly. In the rat, removal of up to two-thirds of the total mass of the liver is followed by active cell division leading to complete restoration of the organ within two weeks. As early as 1923 Akamatsu (1) reported that tissue cultures of rabbit hver grew better in plasma from partially hepatectomized animals than in normal control plasma, and more recently it was shown that cell division can be induced in the resting liver of a parabiotic rat by a partial hepatectomy performed on its partner (2, 3, 4). These findings were considered to indicate the presence or the increase of growth-stimulating factors in the plasma of partially hepatec- tomized animals. In our own studies on the possible participation of the humoral system of communication in the control of this growth, blood serum from animals undergoing liver regeneration was assayed in tissue culture (5). These cultures showed a comparable outgrowth in a high concentration of serum of partially hepatectomized rats and in a low concentration of normal serum. A high concentration of normal serum showed inhibitory effects. Based on these findings a hypothesis was formulated with regard to the induction of the regenerative process in the liver which follows partial hepatectomy. According to this hypothesis, certain constituents of normal blood serum exert a growth-inhibitory action at their normal concentration. Partial hepatec- tomy would be expected to result in a decrease of the serum concentration of these constituents. Thus in turn regenerative growth is initiated. During regeneration, as the number of liver cells increases, the concentration of these constituents will also increase. When the original equilibrium between a given number of liver cells and a given concentration of the serum constituents is restored, further growth is expected to cease. The evidence for a negative feedback system of this type should satisfy the following two conditions: 1. Induction of growth in the resting tissue by plasma dilution. 2. Inhibition of growth in the regenerating tissue by plasma concentration. 148 Evidence for a Negative Feedback System Controlling Liver Regeneration 149 Figure 1 illustrates the application of the classical method for plasma dilution, plasmapheresis, and the results obtained. Normal adult male rats were used. Blood was withdrawn every twelve hours corresponding to 31 to 38 per cent of the initial total blood volume of the animals in the first group RATE TIME IN HOURS 36-48 60 72-96 MEAN 0% <.002 .003 .002 <.002 .004 .002 .002 31-38% <.002 .022 .002 .048 .009 .002 .014 39-46% .168 .002 .019 .054 .441 .197 .147 MEAN .048 .031 .163 Fig. 1. Induction of cell division in the resting liver by plasmapheresis. A total of eighteen adult male rats was used. The rate of plasmapheresis is expressed as the percentage of the initial total blood volume of the animal replaced by saline per 12 hours. In the control group 0 rate refers to the fact that blood was merely withdrawn and re-injected, with the animals submitted to the same stressful conditions of restriction, anesthesia, venipuncture etc. as the experimental groups. The mitotic activity was obtained by counting 50,000 cells, and expressed as the per cent mitotic index. When no mitosis was found, the mitotic index was recorded as <0.002. and 39 to 46 per cent in the second. The bleedings were followed by re-injections of the blood cells suspended in an equal volume of saline. Under such conditions cell division was induced in the resting liver of adult rats and was intensified with increasing dilution of the plasma. In this experiment, then, the evidence obtained satisfies the first condition for a negative feedback system. With respect to the second condition, the method used to achieve plasma concentration was restriction of fluid intake as illustrated in Fig. 2. Two experi- mental groups were used, differing with regard to the weight of the animals and the extent of the partial hepatectomy. All animals were partially hepatec- tomized and tube-fed an identical isocaloric fluid diet containing 3 per cent water. The controls were given drinking water ad libitum but the experimental animals were deprived of water for the duration of the experiment, which was sixty-four hours, starting sixteen hours prior to the operation and continuing for forty-eight hours postoperatively, at which time the animals were sacrificed. A measure of total body-water loss obtained by this regimen is given by the difference in weight change between experimental and control animals in each group. A measure of the plasma concentration achieved is given by the difference in total protein change. In both the experimental groups an effective inhibition of cell division in the liver was obtained; this inhibition became greater with increasing concentration of the serum. On the other hand mitosis in the intestinal epithelium was not affected. The evidence obtained in this experiment, then, satisfies the second condition for a negative feedback system. The smaller extent of total body-water loss and plasma concentration in the first group can be ascribed to the greater initial weight of the animals in 11 150 Andre D. Glinos this group. It is well known that when dehydration proceeds slowly the main- tenance of plasma volume at the expense of extravascular fluid may be quite successful. This is significant since the extravascular fluid of the liver must parti- cipate in the transmission of information to the liver cells. The serum albumin fraction in this experiment was found to be low when liver cell division was present and nonnal or slightly increased when liver cell division was absent. In the framework of the present discussion this feature is somewhat suggestive DEGREE OF HEPATECTOMY TREATMENT NO. OF RATS ANO WEIGHT WEIGHT % CHANGE SERUM PROTEIN % CHANGE SERUM ALBUMIN % CHANGE MITOSES /50000 CELLS 30 % (MEDIAN LOBE) CONTROLS iO 457 - 7.9 - 4.7 - 16.0 81.4 FLUID RESTRICTION 9 458 - 12.9 + 5.0 - 14.0 18.0 10 % ( CAUDATE LOBE) CONTROLS 8 331 - 9.1 + 1.8 - 26.2 16.5 FLUID RESTRICTION 8 313 - 17.6 + 28.9 + 8.1 0.1 Fig. 2. Inhibition of cell division in the regenerating liver by fluid restriction. The experimental variables are defined in the text. The percentage changes refer to differences in weight, total serum protein, and serum albumin between the values obtained before treatment and those obtained at sacrifice. when it is considered that albumin is synthesized in the liver. In view of these facts it was thought that an investigation of changes in protein metabolism of the liver cells, early after partial hepatectomy, could help in elucidating the possible role of the serum proteins and the extravascular fluid of the liver in the transmission of information to the liver cells. In this, we took advantage of the many observations showing histochemically detectable changes in the organization of cytoplasmic ribonucleoprotein with increasing demands on the protein synthetic mechanism of the cells (6, 7). Briefly stated, these changes consist of the disappearance from the cytoplasm of discrete basophilic bodies which are associated with ribonucleoprotein; the cytoplasm then stains unifonnly with basic dyes. Rats were sacrificed at frequent intervals after partial hepatectomy and their livers fixed and stained with gallocyanin chrome alum. Within thirty minutes after partial hepatectomy the ribonucleoprotein-associated basophilic bodies started disappearing from the cells in the periportal area. This change proceeded gradually toward the center, so that eight hours after the operation all cells, even those in the centrolobular area, were affected. After this time reconstruction of the basophilic bodies proceeded in the opposite direction from the center of the lobule towards the periphery. At 24 hours cells in the centrolobular area had completed the cycle Fig. 3. Regenerating liver twenty-four hours after partial hepatectomy. Central vein at lower left corner. Adjacent centrolobular zone with cells con- taining ribonucleoprotein-associated basophilic bodies in their cytoplasm. Middle and periportal zones with mostly altered cells having a uniformly basophilic cytoplasm. Two mitotic figures in the middle zone among altered cells. to face p. 150 Evidence for a Negative Feedback System Controlling Liver Regeneration 151 showing well organized basophilic bodies, whereas cells in the middle and the periphery of the lobules remained altered (Fig. 3). Confirming the earlier data of Harkness (8), we found that cell division begins between 16 and 24 hours postoperatively in the periportal area. This is significant because cells in this area remained altered for the longest time. The changes in cytoplasmic ribonucleoprotein organization indicate an activa- tion of the protein synthesizing mechanism of the liver cells after partial hepa- tectomy, proceeding in a topographical pattern related to the direction of the intralobular blood flow. According to the Law of Mass Action these changes would be expected to appear with decreased protein concentration in the immediate environment of these protein-secreting cells. The cells in the periphery of the lobules would be expected to react faster and longer since the ones more centrally located are in an environment richer in protein produced by the more peripheral cells. This interpretation was, in part, verified experimentally by TREATMENT FLUID SERUM PROTEIN CHANGE LIVER RIBONUCLEOPROTEIN CHANGE ADDITION SALINE - 11.8 0 DEXTRAN - 31.2 + SERUM + 7.9 0 REPLACEMENT SALINE - 19.2 + DEXTRAN - 37.8 + SERUM - 8.7 0 Fig. 4. Induction of cytoplasmic ribonucleoprotein changes in the liver by plasma dilution. A total of six male adult rats was used. Addition refers to a single intravenous injection of 5.5 ml of fluid. Replacement refers to a 5.5 ml single plasmapheresis treatment. All animals were sacrificed two hours after treatment. Serum protein change refers to the percentage difference between the values obtained before treatment and those obtained at sacrifice. Liver ribonucleoprotein change refers to the disappearance of the basophilic bodies from the cytoplasm of the cells in the periportal area. showing that changes in the cytoplasmic ribonucleoprotein of the cells in the periportal area appear rapidly after a sudden decrease of the serum protein concentration (Fig. 4). After partial hepatectomy, however, these histochemical changes occur as we have seen within thirty minutes before any appreciable changes in the plasma proteins. The relationships between increased pressure in the portal system following 1 52 Andr£ D. Glinos partial hepatectomy and regeneration have been demonstrated by Grindlay and BoLLMAN (9). It is conceivable that, under conditions of increased pressure immediately following partial hepatectomy, the transfer of protein and water from the intravascular to the extravascular space is altered and results in a rapid lowering of the protein concentration of the interstitial fluid of the liver. This leads within a short period to increased protein production in the liver cells and sometime later to cell division. REFERENCES 1. N. Akamatsu: Ober Gewebskulturen von Lebergewebe. Virc/i. Arch. 240, 308-311 (1923). 2. B. G. Christensen and E. Jacobsen: Studies on liver regeneration. Acta Med. Scand. 234, Suppl, 103-108 (1949). 3. N. L. R. Bucher, J. F. Scott, and J. C. Aub: Regeneration of the liver in parabiotic rats. Cancer Res. 11, 457-465 (1951). 4. A. S. Wenneker and N. Susman: Regeneration of liver tissue following partial hepa- tectomy in parabiotic rats. Proc. Soc. Exp. Biol. Med. 76, 683-686 (1951). 5. A. D. Glinos and G. O. Gey: Humoral factors involved in the induction of liver regenera- tion in the rat. Proc. Soc. Exp. Biol. Med. 80, 421^25 (1952). 6. S. Lagerstedt: Cytological studies on the protein metabolism of the liver in the rat. Acta Anat., VII, suppl. 9, 1-116 (1949). 7. A. F. HowATSON and A. W. Ham: Electron microscope study of sections of two rat liver tumors. Cancer Res. 15, 62-69 (1955). 8. R. D. Harkness: The spatial distribution of dividing cells in the liver of the rat after partial hepatectomy. J. Physiol. 116, 373-379 (1952). 9. J. H. Grindlay and J. L. Bollman: Regeneration of the liver in the dog after partial hepatectomy. Role of the venous circulation. Surg. Gynec. Obst. 94, A9\-A96 {\952). FLUCTUATIONS IN NEURAL THRESHOLDS* Lawrence S. Frishkopf and Walter A. Rosenblithj Research Laboratory of Electronics, Massachusetts Institute of Technology, Cambridge, Massachusetts Abstract — Over the past twenty-five years several independent investigations of the responsivity of nerve tissue have led to the conclusion that the threshold of a resting neuron fluctuates in time. The conclusion is based on the study of sensory and motor fibers, of monosynaptic arcs and neuromuscular junctions. A number of these studies have been reviewed and com- pared. The degree of threshold correlation among neurons of a given 'pool' or population has been considered for several systems. A number of possible sources of threshold fluctuation, giving rise to correlated and uncorrected threshold variations, have been distinguished. A mathematical model based on the concept of fluctuating thresholds has been described and applied to the problem of ensemble response from the peripheral auditory nervous system. The results of three experiments have been described and compared with the predictions of the model. I. THE CONCEPT OF A FLUCTUATING THRESHOLD The threshold of a nerve fiber is defined as the minimum stimulus intensity that will cause an action potential to propagate. If the threshold of a nerve fiber were a fixed parameter — not changing in time — its value could be deter- mined by presenting stimuli of increasing intensity. The fiber would fail to respond to all stimuli less than some value Srp, and would respond to all stimuli greater than Srp\ Sj, would then be the threshold of the fiber. However, careful experiments on a number of specific neural systems — sensory and motor, peripheral and central — have shown that such a unique value Sj, does not exist; instead, there is a range of stimulus values, 5^ to ^'2, such that a stimulus S lying within that range, when repeatedly presented at a rate well below that which would involve the refractory period of the fiber, sometimes evokes and sometimes fails to evoke a response. We find that the fiber responds in a fraction p of all trials and that p{S) is a monotonic function that rises from zero to one as the stimulus increases from S-^ to 5^2. Stimuli less than S^ never evoke a response; stimuli greater than So always evoke a response. We conclude that the threshold of a neuron which exhibits this behavior is a time-varying para- meter. The value p approximates the fraction of the time that the threshold is somewhere below the stimulus value S. An equivalent statement is that p approximates (and for large sample size, approaches) the probability of finding the threshold of a fiber below the value 5". * This work was supported in part by the U.S. Army (Signal Corps), the U.S. Air Force (Ofiice of Scientific Research, Air Research and Development Command) and the U.S. Navy (Office of Naval Research). t Also in the Department of Electrical Engineering, M.LT. 153 154 Lawrence S. Frishkopf and Walter A. Rosenblith II. SUMMARY OF STUDIES OF OTHER WORKERS The class of phenomena that we have been discussing was first observed by Blair and Erlanger (1). They reported that an electric stimulus, repeatedly presented to a single sciatic nerve fiber of the frog, will for most stimulus values either always produce or always fail to produce a response. The transition between these two situations, however, is not sharp. Upon raising the shock intensity, a value is reached at which the fiber sometimes responds and some- times fails to respond to repeated stimulation. In order to obtain a response every time it is necessary to raise the shock intensity an additional two per cent, far in excess of the uncontrollable variation in the stimulus. Moreover, Blair and Erlanger were able, on occasion, to record simultaneously from two fibers whose potentials could be distinguished by their difference in latencies. On repeated testing with a near-threshold stimulus, sometimes both would respond, sometimes one, sometimes the other, and sometimes neither. Such a result cannot be accounted for on the basis of stimulus instability alone. The most complete study of this kind that has been published to date was made by Charles Pecher (2) in 1939. Using a technique similar to that of A- A- J^ .^ ^ J\^ Fig. 1 . Left : ink tracings of recordings from single units of frog sciatic nerve, showing occurrence and failure of response to repeated presentations of identical shock stimuli. Right: same, with amplitude of pulse producing the shock raised 4 per cent. Each series shown is part of a longer sequence of 100 presentations. Thirty-five responses were obtained with the weaker stimulus (left); 85 responses were obtained with the stronger stimulus (right). After Pecher (2). Blair and Erlanger, he also found a stimulus range in which a fiber sometimes responded and sometimes failed to respond to a constant stimulus. Some of his data appear in Fig. 1. In the column on the left we see the responses to successive identical stimuh, of which some produce a response and some fail to do so. In the second column the intensity was raised four per cent. In Fluctuations in Neural Thresholds 155 Fig. 2 the percentage of responses of a fiber is plotted as a function of stimulus intensity. Again each point is based on 100 stimulus presentations. The total range of thre'shold variation is, on the basis of these data, about seven per cent. The function shown in Fig. 2 approximates the threshold probability function p(S) that was discussed earlier. 100 7b 50 25 O '/]': -.y^: 97 98 99 X)0 101 102 103 Fig. 2. Relation between stimulus intensity (abscissa) and the number of respon- ses obtained in 100 presentations at a fixed intensity from a single unit of frog sciatic nerve (see Fig. 1). The interpolated solid line approximates the threshold probability function of a unit. From Pecher (2). Fig. 3. Left: ink tracings of simultaneous recordings from two units of frog sciatic nerve to repeated presentations of identical shock stimuli. Units A and B are identified by their latencies. Right: same, but recording from two other units, identified by their amplitudes. After Pecher (2). In the left column of Fig. 3 the responses of two different fibers were simul- taneously recorded from a single electrode; the responses arc distinguishable by their latencies. At a fixed level of stimulation all possible combinations of response occur: fiber A responds alone, fiber B responds alone, both respond, neither responds. On the right we see the responses from two other fibers; here the responses are distinguished by their amplitudes. Again, all possible 156 Lawrence S. Fpushkopf and Walter A. Rosenblith combinations occur. Such a result can only be explained as a result of spon- taneous variation in fiber threshold. If threshold were fixed and the stimulus unstable, then only three of the four combinations could occur. That combina- tion would be excluded in which the fiber with higher threshold fires alone. When responses from two fibers can be distinguished, an opportunity is offered to test the degree of correlation of threshold fluctuation among different fibers. If fluctuations occur independently in two fibers, the probability of both firing to a single stimulus would be the product of their probabilities of firing separately. Any correlation in threshold variations would alter the probability of joint firing. These probabilities can be approximated by counting the number of times that fiber A fires, that fiber B fires, and that both fire, and dividing each by A'^. In the table below the results of such measurements by Pecher Table I Calculated Number of stimuli Number of responses of fiber A Number of responses of fiber B number of simultaneous responses (independence assumed) Observed number of simultaneous responses 100 78 25 19.5 19 188 129 26 17.8 18 285 205 33 23.7 18 222 150 79 53.4 56 370 214 93 53.8 50 194 113 34 19.8 19 155 110 62 44.0 40 218 168 87 67.0 59 236 152 24 15.5 17 are given for nine different fiber-pairs. In all of these instances the computed and observed frequencies of joint occurrence are in good agreement. The hypothesis of independent fluctuations is thus supported by this experiment. Pecher tried to determine whether or not for a single fiber the 'response no-response' pattern to a sequence of periodic stimuli can be accounted for by the hypothesis that successive responses occur with equal and independent probability p. He chose a criterion of independence that relates the variables r and n^, where n^ is the number of times that a sequence of r successive responses (bounded at each end by the absence of a response) occurs in a sample of length A^(r u z UJ D o hi cc 20 40 60 80 FIRING INDICES 100 Fig. 5. Histogram showing the number of spinal motoneurons (triceps surae) within each firing index interval ; responses were obtained by delivering repeated shocks to the gastrocnemius nerve. The firing index of a unit is the percentage of total stimulus presentations to which the unit responds. Units with firing indices of zero and 100 are not included in this diagram. From Lloyd and McIntyre (5). same amount; thus some units with a firing index of zero will be shifted into the intermediate range ; some with intennediate firing indices will be shifted into the range of firing index 100. But because the units are uniformly distributed the same number will move into the intermediate range as move out of it, and the distribution of intermediate firing indices will remain unchanged. Fig. 6. Idealized relation between the threshold probability distribution of a motoneuron and the levels of synaptic drive to diff"erent motoneurons of a population (see text). The particular choice of a bell-shaped probability distribution will lead to the U-shaped histogram of Fig. 5. For it is clear that if we divide the abscissa in such a way that equal areas under the distribution are subtended, those intervals will be largest near the tails of the distribution (firing indices near 0 and 100) and smallest at the center of the distribution (firing index near 50). Since the density of units along the abscissa is uniform, this means that many more motoneurons will have firing indices between 0 and 10 than between 45 and 55. Fluctuations in Neural Thresholds 159 As in the study by Pi chlr, the degree of correlation of thresJiold variation for members of the same pool of motoneurons was investigated. The extent of correlated and uncorrelated fluctuations is a measure of the relative impor- tance in producing fluctuations of events extrinsic and intrinsic to the fiber. In the spinal cord there is reason to believe that threshold fluctuation is, at least in part, the eff'ect of background activity in other fibers. Such activity would presumably aff'ect many fibers in a neighborhood; the threshold fluctua- tions of these fibers would therefore show definite correlations. To determine the extent of correlated variation Rall and Hunt (6) recorded the response of a ventral root together with the response of a single moto- neuron belonging to an adjacent root; an example of such a recording is shown in Fig. 7. Fig. 8 shows the results of an experiment based on a thousand -n nHHi « 1 Fig. 7. Simultaneous recording of the responses of a single motoneuron (hori- zontal deflection) and of an adjacent ventral root (vertical deflection) upon repeated stimulation of the gastrocnemius nerve with identical shock stimuli. From Rall and Hunt (6). such responses. The population response amplitudes were divided into class intervals, and the number of responses within each class interval was plotted. For each population response within a class interval, the occurrence or failure of a unit response was noted and the number of unit responses plotted (shaded area). The unit responded a total of 697 times out of 1000. If the population response and the unit response were not correlated, the firing index of the unit would be about the same in each class interval. This is clearly not the case. Instead, firing occurs infrequently when the population response is small, and more often as the population response grows. The probability of unit firing when the population response amplitude is in a given class interval — that is, the ratio of shaded to unshaded amplitude — is plotted in the lower part of the figure. If unit response and population amplitudes were uncorrelated this function would be a horizontal line at about 0.7. However, it is also clear that correlation of unit and population response is not complete. In other words, the thresholds of the units within the population vary with respect to one another, in addition to their collective (that is, correlated) fluctuation. If this were not so, a particular unit would respond only after all units of lower threshold had responded; therefore its probability of response would be zero if the population response were smaller than a certain value, and would be one if the population response were larger than that value. The lower curve would therefore be a step function. III. POSSIBLE SOURCES OF THRESHOLD VARIATIONS Fatt and Katz (7) have found that at motor endplates miniature end-plate potentials occur more or less randomly even though no stimulus is present. 160 Lawrence S. Frishkopf and Walter A. Rosenblith They regard these potentials as being the result of spontaneous firings in the fine terminal branches of a motor nerve. The occurrence of an impulse in the nerve causes simultaneous firing in about a hundred such teitninals, giving rise to the normal end-plate potential. Spontaneous firing implies the existence of a local source of varying excitation. Fatt and Katz compute that for fibers with a diameter of 0.1 /,( thermal fluctuations in ionic concentrations >- u z u o UJ cr 1«U 1 1 ifin 140 - r^'l I2U 100 - — j 80 - f — 60 40 _ ^ 20 r-H -'■> 1 .t, i. -i I n^ {/ ^T / / O " 2 4 6 8 10 12 14 16 18 20 22 24 26 28 POPULATION RESPONSE AMPLITUDE 1.0 0,9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 O / Fig. 8. Top: the upper curve is a histogram of population response amplitudes obtained as in Fig. 7 from triceps surae motoneurons by delivering repeated identical shock stimuli to the gastrocnemius nerve. The lower curve (shaded) was obtained from single-unit recordings like those shown in Fig. 7 ; the number of single-unit responses associated with population responses in each amplitude interval is plotted. Bottom : for a given population amplitude interval the number of single unit responses is divided by the total number of trials in that interval, and the ratio plotted as a function of population amplitude. The interpolated solid curve is a sigmoid fit to the data points and approximates the probability of unit response as a function of population amplitude. Note that when the popula- tion amplitude is large, the probability of unit response is large, and when the population response is small, the single unit probability is small, thus signifying a high degree of correlation among the thresholds of different units of the popula- tion. From Rall and Hunt (6). could cause variations of resting potential of 1 mV to 2 mV. Though probably insufficient to produce excitation, such a variation would cause threshold fluctuations and contribute to spontaneous firing. Both Pecher (2) and Hunt (8) have discussed possible sources of threshold fluctuation. Pecher considers in detail the apparent threshold variation that would result from statistical variations in the number of ions traversing the axon membrane when a constant potential is applied across it. Assuming that the excitatory current that he uses is uniformly distributed over a cross section of the nerve trunk, he concludes that at threshold about a million ions traverse a single nerve fiber. The statistical variation in this number of ions is Fluctuations in Neural Thresholds 161 given by its square root, leading to a variation of about 0.1 per cent. This is several orders of magnitude below the range of threshold variation that he observed. However, he points out that the number of ions actually effecting excitation is probably considerably less than the value mentioned above and the resultant variability correspondingly greater. Pecher also considers as a possible source of threshold fluctuations local statistical variations of mem- brane potential, of the sort discussed by Fatt and Katz. Hunt discusses two classes of possible sources of threshold fluctuation for spinal motoneurons : (a) sources with a local origin such as we have mentioned above, which give rise to an independent component of threshold variation and (b) sources whose effect is felt by many fibers and which therefore produce at least partially correlated variations in threshold. In the latter category are included the effects of activity of spinal interneurons. By using a drug (myanesin), in doses that block transmission through polysynaptic paths without reducing monosynaptic reflex responses, a considerable reduction in the range of variation of population response amplitudes was obtained. On the basis of this result it appears likely that internuncial activity is important in producing correlated threshold changes in spinal motoneurons. IV. A MATHEMATICAL MODEL Let us consider a mathematical model which is based on the concept of fluctuating thresholds, and which attempts to derive the ensemble behavior of large numbers of neural elements from assumed properties of neural units in a specific area of the nervous system (9, 10, II). This model is based on data obtained from the peripheral auditory system of the cat. When an electrode is placed near the round window of the cochlea, responses to clicks can be detected; such responses contain a component that represents the summated activity of peripheral auditory neurons. Fig. 9 shows such population responses at a number of intensities. In Fig. 10 the average peak-to-peak amplitude of such responses has been plotted as a function of stimulus intensity. The resultant 'intensity function' relates the number of units firing and the intensity of the stimulus. The present version of the model (11) postulates the existence of several independent populations of neural units; within a population all units are identical. The threshold of a unit is a fluctuating parameter which can be described by a probability distribution; threshold variations in different units occur independently. At a rate of stimulation slower than one per second the 'response no-response' sequence obtained from a single unit is assumed to consist of a series of independent events. Thus we postulate units whose statistical properties resemble those found by Pecher in the frog's sciatic nerve. The experiments used to test the model fall into three classes: two-click experiments (9, 10), measurements of variability of response amplitude (II), and studies of masking of click responses by noise. When two clicks are delivered at an interval of less than approximately 100 msec the population response to the second click is smaller than it would be if the first click had not occurred. This effect is more pronounced the stronger the first click and the smaller the interclick interval, as illustrated 162 Lawrence S. Frishkopf and Walter A. Rosenblith in Fig. 11. Consider the ratio of the response amplitude ^R^ to a second click and the response amphtude R^ to the same click presented alone. In Fig. 12 this ratio is plotted, for a fixed second-click intensity, as a function of the inten- sity of the first click. The parameter is the interval between clicks, At. If STIMULUS INTENSITY IN DB (RE 1.29 VOLTS) •90 •40 -80 -30 ■70 •20 -60 ■10 ■60 3 msec 3 msec 0 DB RELATIVE GAIN -12 DB Fig. 9. Ink tracings of responses obtained from an anesthetized cat to clicks over a 90-dB range. The electrode was located near the round window. Note that the voltage gain of the recording equipment was reduced by 12 dB (factor of i) for stimulus intensities above —40 dB. The first peak represents the summated activity of first-order auditory neurons. With this calibration, click threshold for humans (verbal report) is about —95 dB. we assume a one-population model, we obtain the result that the ratio 1R2IR2 is Hnearly related to the intensity function for the first click, provided that the second-click intensity (S^) and At are held constant. Specifically, we obtain rR. , ^^^^\l - giS„ Ar)] (1) R, 1 - Rr Determination of a single intensity function therefore permits us to predict the dependence of this ratio on S^ for any value of 5*2 and of At. We may Fluctuations in Neural Tiiresholds 163 u E 3.0 AMPLITUDE (-c^) AND LATENCY (-•-)OF N, AS A FUNCTION OF 2-5\-|NTENSITY OF CLICK STIMULUS " 150 125^ -100 -80 -60 -40 -20 0 CLICK INTENSITY (dB RE 1.29V ACROSS PHONE) Fig. 10. Intensity function (open circles). A''i is the first diphasic response com- ponent seen in the traces of Fig. 9. The amplitude measurement is made between the positive and negative peaks of N^. Each plotted point is the median of about ten such measurements. m o o o in tr 50 DB 30DB RESTING RESPONSE tl -lODB 12.2 MSC 30MSC 6IMSC TIME INTERVAL BETWEEN CLICKS Fig. 11. Two-click paradigm: the responses shown are to a constant intensity (—45 dB) second click. The vertical set shows the effect of varying the intensity of the first click; the horizontal set shows the effect of varying the interval between clicks. Upper right: response to a — 45 dB click presented alone. From McGiLL (10). in each case choose one constant, ^(5*2, At). Fig. 12 shows a number of fits to the data points which were obtained in this way; 5*2 is constant and each curve corresponds to a different value of At. In a second group of experiments the standard deviation of a hundred response ampHtudes was computed at each stimulus intensity, and the result was plotted as a function of stimulus intensity. It is readily shown that N 164 Lawrence S. Frishkopf and Walter A. Rosenblith independent units, each with a probability p of firing, will have a standard deviation of total response proportional to \/Np{\ — p). As a function of /; this quantity has minima at zero and one and has a maximum at/? = ^. The value of p at any stimulus intensity can be obtained from the intensity function. u Ll -I OU "□ c CM ^2 > CO I- UJ' < cr _i UJ tr 1.00 0.75 0.50 0.25 1.00 0.75 0.50 0.25 0 |\ V , c X • E G IX \ o - - • "•n N* N. 1 ' D "^•.^ m F S2=-45 dB 1 (m sec} g(Al A 6.4 0.05- B 9.1 0.15 C 12.2 0.27 D 21.5 0.46 E 30 0.54 F 61 0.70 G 102 0.83 60 40 20 0 60 40 20 0 60 40 20 0 INTENSITY OF FIRST CLICK IN dB BELOW REFERENCE LEVEL Fig. 12. 1R2IR2 (see text) as a function of first click intensity. In each block tiiis ratio is plotted for a different interclick interval, as indicated at the lower right. The intensity of the second click was —45 dB throughout. The curves are obtained from the first click intensity function and eq. (1); the parameter ^(At), whose values are given at the lower right, is chosen in each case to give the best fit to the data. After McGill (10). Fig. 13. Intensity function (upper) and the corresponding amplitude variance function predicted by the model: (a) for one population; (b) for two disjoint populations. Oq was chosen arbitrarily. Note that a peak of the variability function occurs at the stimulus value at which an intensity function component reaches half its maximum amplitude. Fig. 13 shows the kind of variability function obtained by assuming one and two disjoint populations; Cq is the stimulus-independent component of variability arising from biological and non-biological sources. We have shown (II) that instability in stimulus intensity, which would also lead to a peaked variability function, can account for at most three per cent of the observed variability. A detailed study of the shape of the intensity function led us to postulate Fluctuations in Neural Thresholds 165 two populations of neural units, one consisting of 'sensitive' units and one of 'insensitive' units. In the three animals tested, variability measurements over the sensitive range are in good agreement with the theory stated above. One case is shown in Fig. 14. The intensity function and the probabilities obtained from it are shown with the derived standard deviation function. Here, Oq is determined from measurements of baseline variability in the absence of a stimulus; TV is chosen to give the best fit to the data. Over the sensitive range CLICK INTENSITY (DB RE I.29VACR0SS PHONE) Fig. 14. Comparison of the theoretical variability function (with 70 per cent con- fidence limits) and the measured values of cr, over the range of initial growth of the intensity function. Each point represented by a solid circle is based on 100 responses; the open circles are based on the first fifty of these responses. The corresponding intensity function, and the probabilities obtained from it, are also shown. (— lOOdB to — 60 dB) the data fall within the indicated confidence interval approximately seventy per cent of the time, as they should if the model is correct. Over the insensitive range of the intensity function (—60 dB to 0 dB), the standard deviation shows a complex behavior which cannot be simply reconciled with the idea of a single population over that interval. The third aspect of this study concerns the masking of the neural responses to clicks by a background noise. Fig. 15 shows the effect of a constant noise level on response amplitude at several stimulus values. In Fig. 16 we have plotted these masked and unmasked intensity functions. The observation was made that a very weak level of continuous noise was sufficient to reduce almost to zero the N^ response to a fairly intense click. A fixed threshold model would predict masking of only the units whose thresholds are below the noise level. If the threshold fluctuates, however, and does so rapidly, nearly all units of a given population will drop below the noise level and fire in a short 12 166 Lawrence S. Frishkopf and Walter A. Rosenblith RESPONSE TO CLICK ALONE CLICK INTENSITY IN DB (RE 1.29 VOLTS) CLICK RESPONSE IN RELATIVE PRESENCE OF NOISE GAIN (-82 DB RE 1.29 VOLTS) -60 ODB -50 -40 ^x^ v/-' -30 •12 03 -20 -10 -3 m sec - -3 msec- Fig. 15. Ink tracings of responses obtained from an anesthetized cat to clicks over a 60-dB range, with and without background noise; noise level, —82 dB. Note that the voltage gain of the recording equipment was reduced by 12 dB (factor of i) at a click intensity of -30 dB. Fluctuations in Neural Thresholds 167 interval preceding the click; this assumes, of course, that the noise level lies within the range of threshold fluctuations of the unit. By a quantitative treatment based on these qualitative notions we have been able to show (a) that the hypothesis of a fixed threshold does not account for the observed data and (b) that over the sensitive range of the intensity i25r 100 75 50 I 25 ^^^_ NO NOISE BACKGROUND (COMPOSITE OF 3 FUNCTIONS) NOISE BACKGROUN0(OB RE 1.29V ACROSS PHONE lUNFILTEF -92 -82 -67 -80 -70 -60 -50 -40 -30 CLICK INTENSITY (06 RE 1.29V ACROSS PHONE) Fig. 1 6. Intensity functions for clicks, with an(d without noise backgroun(i ; noise levels —92, —82 and —67 dB. Each point of the masked functions represents the average A''i amplitude of ten responses to identical stimuli. The upper curve was obtained by averaging the three unmasked functions which correspond to the masked functions shown ; thus each point represents the average A''i amplitude of thirty responses to identical stimuli. Typical data on which these curves are based are shown in Fig. 15. function a single population of units making threshold 'jumps' at a rate of about 2000 times per second can account for the data. In addition, it is observed that low level noise has little effect on the intensity function over the insensitive range, except to reduce it by the constant contribution of the sensitive popu- lation. The need for a division of units into at least two populations is thus confinned. When the noise level is raised into the insensitive range the observed effect is not nearly so marked, implying either that more than one population is involved in that interval or that the rate of threshold fluctuation is considerably slower than for the sensitive units. It is noteworthy that population analyses based on two very diff'erent experiments, variability and masking, have a great deal in common. REFERENCES 1. E. A. Blair and J. Erlanger: A comparison of the characteristics of axons through their individual electrical responses. Anier. J. Physiol. 106, 524-564 (1933). 168 Lawrence S. Frishkopf and Walter A. Rosenblith 2. C. Pecher: La fluctuation d'excitabilite de la fibre nerveuse. Arch. Int. Physiol. 49, 129-152 (1939). 3. A. Hald: Statistical Theory with Engineering Applications, p. 344 (eq. 13.3.6), J. Wiley and Sons, New York (1952). 4. W. A. Rosenblith: Some electrical responses from the auditory nervous system. Pro- ceedings of the Symposium on Information Networks, Polytechnic Institute of Brooklyn, 223-247 (1954). 5. D. P. C. Lloyd and A. K. McIntyre: Monosynaptic reflex responses of individual motoneurons. J. Gen. Physiol. 38, 771-787 (1955). 6. W. Rall and C. C. Hunt: Analysis of reflex variability in terms of partially correlated excitability fluctuation in a population of motoneurons. /. Gen. Physiol. 39, 397-422 (1956). 7. P. Fatt and B. Katz: Spontaneous subthreshold activity at motor nerve endings. /. Physiol. 117, 109-128 (1952). 8. C. C. Hunt: Temporal fluctuation in excitability of spinal motoneurons and its influence on monosynaptic reflex response. J. Gen. Physiol. 38, 801-811 (1955). 9. W. J. McGiLL and W. A. Rosenblith: Electrical responses to two clicks: a simple statistical interpretation. Bull. Math. Biophys. 13, 69-77 (1951). 10. W. J. McGill: a statistical description of neural responses to clicks recorded at the round window of the cat. Ph.D. Thesis, Harvard University (1952). 11. L. S. Frishkopf: A probability approach to certain neuroelectric phenomena. Research Laboratory of Electronics, Massachusetts Institute of Technology, Tech. Rep. No. 307 (1956). PART III DETERMINATION OF INFORMATION MEASURES It is possible (as shown by several papers in this volume) to apply information theory to biology without introducing any actual information measures. Indeed, if one considers that it is very difficult to estimate information measures for living systems, and that the resulting measures are of an irreducibly relative nature, one might wonder whether it is worth-while to take such measures at all. However, it is difficult if not impossible to validate firmly the application of information theory without critical tests based on quantitative measurements; moreover, one hopes to discover lawful relations in the results of the measure- ments themselves. So, attempts are being made to estimate information contents associated with various biological structures and functions. All the papers in this part are chiefly concerned with such estimations; some from a general point of view, some with regard to particular systems, ranging in complexity all the way from simple molecules to whole men. H. Q. 169 CHEMISTRY AND BIOCHEMISTRY AT LOW TEMPERATURES AND DISCRIMINATION OF STATES AND REACTIVITIES* Simon Freed Chemistry Department, Brookhaven National Laboratory, Upton, New York Abstract — In order to apply information theory to biochemistry and biology at the molecular level it is advantageous to reduce the number of classifications and specifications involved by reducing the temperature of the system. In this way the number of species and states with their reactivities is reduced. At the same time the chemical noise level falls and in consequence a resolution may be obtained between components whose properties are practically indis- tinguishable at ordinary temperatures. Weakly bonded systems and intermediates become more easily detectable not only because of an increase in their concentration, that is, an increase in their signal, but in addition because the noise level is weaker at the lower temperatuie. Illustrations are given from chemistry where reactions in solutions proceed at the tempera- tures approaching that of liquid nitrogen. The information content of irreversible reactions at room temperature may be thought of as being stored in intermediates that participate in reversible reactions at the low temperatures. Once the properties of the more stable states have been understood, the way is clear for investigating the system in its thermally active states since allowance can be made for the presence of the former. In this way, an ordering of experimentation according to temperature will bring into activity successive components of the system. Examples have been selected mainly from work on the preservation of biological systems at low temperatures which indicate that biochemical and biological processes may likewise be investigated and that the finer discriminations and specificities associated with lower temperatures may be brought to light in these fields also. If we wish to measure a physical property, such as electrical conductivity or viscosity, with an instrument which we have no intention of modifying, there is little point in seeking the information content of the instrument. On the other hand, if we wish to employ chemical substances as probes for uncovering structures of enzymes by means of enzyme-substrate reactions, we are at once confronted by the need of the structural and functional information of our probes. In fact we are discussing properties at the molecular level. Pure substances at this level are mixtures composed of molecules in various energy states with their characteristic configurations, motions, and reactivities. The application of information theory to biology at the molecular level requires therefore a great expansion in the number of categories and specifications. It is to reduce this number in a systematic manner and make these categories more precise that I wish to draw upon the relation that has been recognized between information and entropy which asserts that the amount of information * Research performed under the auspices of the U.S. Atomic Energy Commission. 171 172 Simon Freed required to specify the system will be less at lower temperatures. The system will redistribute itself from higher to lower energy levels so that only the more basic ones remain appreciably occupied. Fewer chemical species are now present and also active. There has been, in a sense, a reduction in chemical noise differing in its frequency spectrum from the continuum characteristic of an electrical conductor. Chemical noise reflects the structural properties of molecules and may consist of dominant discrete frequencies associated with virtual continua of modulations. Usually these represent couphng of the electronic system of the molecule in a given atomic configuration with its 300 ■ II B 77 II 1 1 II 4 1 Fel III 1 l|l 1 II 1 |1 If 1 Fig. 1 . The variation of absorption spectrum of praseodymium chloride with temperature. Line drawings of visible absorption spectra of crystals of anhydrous praseodymium chloride (PrClg) at room temperature, at that of liquid nitrogen, and that of liquid helium. Sharper spectra, improved resolution, and fewer hnes are evident at lower temperatures. The fewer hnes correspond to fewer energy states which are occupied by the praseodymium ions. At room temperature the blocks of diffuse spectra are actually not uniform in intensity but are more intense as a rule in those regions where the spectrum of the crystal at 77°K possessed its most intense line spectrum. The greater diffuseness of the lines and their increased numbers at the higher temperature may be regarded as chemical noise associated with the spectroscopic signals from the more stable states at the lowest temperature. own vibrations, restricted rotations, etc. If the molecules are complex, fluctua- tions between difl'erent atomic configurations may contribute to the noise. In addition, coupling of the molecule in each of its states with the molecules of its environment in different configurations leads to more and more densely spaced energy levels which I referred to as the continua. A reduction in temperature removes thennal energy required to activate some motions and effect changes in configurations, and reduces the number of perturbations of a given configuration. Not only are fewer species present but each species is more sharply defined; thus, less infonnation is required for specifying the system than at higher temperature. Clearly, the system is now more specific in its reactions than at higher temperature and its specificity can be related to more sharply defined geometric configurations. The chemical system has become a more precise probe. The following illustrations have been selected for the simplicity of their phenomena rather than for their direct relevance to biology. The sharp absorption spectrum of a crystal of a rare earth salt (Fig. 1) shows very clearly that at the lower temperature fewer lines are present; they are sharper and more clearly resolved and the general diffuse background prominent at the higher temperature (not shown in the line drawing of the figure) becomes decidedly weaker. There are then fewer kinds of absorption centers at the lower temperature and, because the stable states are exposed to Chemistry and Biochemistry at Low Temperatures 173 more sharply defined environmental fields, there are fewer kinds of pertur- bations. An especially vivid example of a solution showing somewhat similar pheno- mena is given by the fluorescence spectrum of solutions of europium chloride in ethanol at various temperatures (1), The spectra were taken to discover the discrete number of lines in the three separate sets which may furnish the point 4000 4500 5000 Fig. 2. Absorption spectra of carotene (90% alpha and 10% beta). A — In hep- tane at room temperature; B — In equal volumes of liquid propane and propene at 77°K. group symmetry of the electrical fields about europium ion in the solution. It is clear that at room temperature the continuous noise is so great as to make enumeration impossible. As the temperature is lowered a few discrete Hnes can be resolved with such definiteness that they serve to eliminate some of the possible point group symmetries. At the temperature of liquid nitrogen and even at the temperature of dry ice adequate resolution is clearly achieved and the number of possible symmetries of the environmental fields is reduced to one only. Figure 2 gives the absorption spectrum of a substance of some biological interest, /9-carotene, and illustrates the increased contrast between absorption and transmission at the lower temperature, that is, the increased signal to noise ratio. Figure 3 is presented to illustrate the resolution into components of what 174 Simon Freed is apparently a single species at room temperature. The figure reproduces the absorption spectra of chlorophyll b in ethyl ether and methanol (2). Our first inclination is to ascribe the differences in the spectra to the perturbations produced on the structure of the chlorophyll molecules by the two types of solvent molecules. Figure 3b is a magnification of the Soret band in the blue 400 500 600 700 WAVELENGTH IN mjJi Fig. 3a. Absorption spectra of chloro- phyll b at room temperature. The thin- lined curve with maxima at shorter wave- lengths represents a solution of chloro- phyll in ethyl ether; the thick-lined curve gives the spectrum when the solvent is methanol. 4100 5000A Wavelength Fig. 3b. The dependence of the absorp- tion spectra of chlorophyll b on tem.pera- ture. Only the Soret band in the blue is shown. Enlarged scale of wave-lengths. At 300"K the solvent is 20% propyl ether, 80% hexane. At the lower tem- perature it is 20% propyl ether, 40% propane, and 40 % propene. The hexane was substituted at 300°K for the hydro- carbons propane-propene since they are normally gases at room temperature. region and shows that a solution of chlorophyll b in ether is really a mixture of two species (etherates) in equilibrium with each other in roughly equal amounts and clearly resolved at 180°K. A study of the dependence on tempera- ture of the absorption spectrum of chlorophyll b in methanol reveals that in this solvent, chlorophyll b also exists as a mixture of solvates which are about equal in concentration at room temperature and together they yield the com- posite spectrum. However the spectrum of each alcoholate differs very little in shape from that of each etherate. Fig. 4 illustrates a form stable at a lower Chemistry and Biochemistry at Low Temperatures 175 temperature reacting to produce reversibly a stable intermediate but at still higher temperature ending in an irreversible reaction. The following specific observations may prove worthwhile in illustrating what is probably a rather common phenomenon. Chlorophyll b dissolved in CHLOROPHYLL B' IN 15% MONO-i- PROPYL AMINE AND 1: 1 PROPANE -PROPENE 230°K ! : : : I93°K • i ; 1 z > o i^ '• H ^t^^^^ 1 i cr X^- \ ; o CD y^ 1 > <£ tv -^"'"^ ^^V '^ 1 \ 1 — 1 4000 5000 WAVELENGTH % 6000 Fig. 4. Chlorophyll 6' in 15% mono-/-propyl amine in 1 : 1 propane-propene. To show the presence of the red-brown intermediate stable at 193''K which is in equilibrium with the original chlorophyll. At temperatures higher than about 235''K, an irreversible reaction occurs. ether is deposited as a green powder by pumping off the ether at room tempera- ture. When the temperature of the powder is reduced to that of dry ice (about 193°K) and propylamine is condensed upon it at this temperature, it dissolves quickly, forming a red solution. Note in Fig. 4 the new absorption between 5000 A and 6000 A. A rise in temperature transforms the color into the green of chlorophyll with its characteristic spectrum which reverts back reversibly to the red substance when the temperature is reduced. However, if the tem- perature is kept any length of time at about 235°K or higher, an irreversible reaction sets in. For example, at room temperature the red color lasts only a fraction of a second. This evanescent red color is produced in the well known phase test for chlorophyll. Figure 5 represents a chemical reaction which appears rapid even between 167°K and 75°K. Chlorophyll h dissolved in di-/.so-propylamine is undergoing 176 Simon Freed transformation probably in an acid-base reaction. The quick readjustment to equilibrium is shown by the interchange in relative intensities of the bands in the red region. The band furthest towards the red grows in as the temperature is reduced, at the expense of the band near it toward shorter wavelengths. That these reactions occur rapidly at such temperatures is not very sur- prising since little heat of activation is required for this type of reaction. Figure 6 depicts a type of oxidation-reduction at low temperatures. When iodine 3000 4000 5000 6000 WAVELENGTH ANGSTROMS 7000 Fig. 5. Chlorophyll b in 15% dipropylamine diluted with equal proportions of propane and propene. A chemical readjustment toward equilibrium occurs between 170°K and 75°K. is finely divided it rapidly dissolves in isoprene at the temperature of dry-ice, 193°K. A brown solution forms at the solid-liquid interface but it decolorizes very quickly, becoming colorless a little distance from the iodine surface. In the light of other investigations it was surmised that the solution is brown because of the presence of a 1:1 (molecular iodine-hydrocarbon molecule) addition compound which possesses a characteristic absorption band in the ultraviolet region. To build up any appreciable concentration of this compound it would evidently be necessary to make solutions of iodine in isoprene below 193°K. When a solution of isoprene in propane (to which propene had been added to increase the solubility of isoprene) at the temperature of liquid nitrogen (77°K) is mixed with a solution of iodine in propane and propene. Chemistry and Biochemistry at Low Temperatures 177 the new band anticipated in the ultraviolet does not appear within a day or two. Figure 6 indicates what happens when such a solution is warmed. At 146°K the absorption band shown is due to the iodine-propene molecular addition compound which has been identified in a previous experiment. At 150°K appears the anticipated new band arising from the compound iodine- isoprene. At 154°K, this band quickly disappears irreversibly and at the same time decoloration of the solution occurs. The molecular iodine has been removed, presumably by the halogenation of the double-bond system of isoprene, just 0. q: o (J) CD WAVELENGTH Fig. 6. Isoprene dissolved in 1 : 1 propane-propene to which iodine dissolved in 1 : 1 propane-propene has been added. The new absorption band which appears at 1 50°K is due to a 1 : 1 molecule addition compound of the iodine to isoprene. Its disappearance at 154°K is due to an irreversible reaction, probably halogenation across the double bond. as had occurred when solid iodine reacted with isoprene at the temperature of dry ice. This oxidation appears to require the prior formation of the inter- mediate molecular addition compound stable at about 150°K at the concen- trations employed. By investigating the properties and reactions from the lowest practicable temperature upward we would observe the appearance of new thermally activated states and their subsequent reactions. 1 78 Simon Freed In analogy with the phenomena illustrated we would expect that a knowledge of biochemical and even biological processes of considerable value may be gained by investigations at low temperature. Support for these expectations comes mainly from recent investigations directed toward the preservation of cells, tissues, and entire organisms. Even more cogent for our purposes are the instances of partial preservation at low temperatures which becomes more effective at still lower temperatures. Unless explicit references are given, the following examples are drawn from the excellent review by Audrey U. Smith (3). For example, H. F. Smart found that twenty-one species of bacteria, yeasts, and molds continued to multiply in frozen media at 264. 1°K. Sizer and Joseph- son found that lipase was active at 248. 5°K, tryptic digestion proceeded at 258°K, and that invertase continued to hydrolyze sucrose at 255°K. At 203°K, however, they could detect no hydrolysis during several weeks. In the preser- vation of red blood cells, about ten per cent deterioration occurs per year at dry ice temperature, 193°K, but scarcely any loss is incurred when they are kept at the temperature of liquid air, 80°K. Ovarian tissue failed to survive nine days at 193°K but survived more than a year at 80°K under otherwise similar conditions (4). Revival of rats after cooling to 273. 5°K was reported by Andjus (5, 6). Irreversible reactions are then clearly progressing at low temperatures, in red blood cells and ovarian tissue at 193''K and at somewhat higher tempera- tures in the enzymatic reactions. If the simple reactions such as those of isoprene and iodine, chlorophyll and propylamine serve as models, the irreversible reactions are preceded in their first and intermediate stages by reversible reactions at still lower temperatures.* Becquerel found that rotifers, spores of bacteria, non-sporing bacteria, algae lichens, mosses, and seeds of higher plants, after having been dried in a vacuum of 10^^ mm Hg over barium oxide, could be successfully kept at the temperature of liquid helium (4°K). Parkes showed that human spermatozoa survived exposure and storage at 80°K. Ovarian, testicular, pituitary, and adrenal tissue have given functional grafts after storage at 80°K, especially if glycerine was added. Luyet established that vinegar eels, spermatozoa muscle fibres of frogs, and hearts of embryonic chicks could be revived after sudden cooling to the temperature of liquid air (80°K). It is then not surprising that enzymes have been cooled to such temperatures without loss of subsequent potency. It would seem then that a number of biochemical and biological processes are available for study at low temperatures. I shall consider both homogeneous and heterogeneous solutions. The first implies that solvents must maintain all the reactants in solutions fluid at low temperatures. It would seem well worthwhile to employ conventional solutions at as low temperatures as possible, and aqueous systems near zero degrees or under supercooled conditions. It has been shown (8) that proteins * Lovelock (7) ascribes the deterioration of red cells to a physical mechanism rather than to a chemical process, namely, that the dissolution of lipoprotein and other components of the cell membrane proceeds more rapidly than the biochemical processes can repair them at the low temperature. Since the lipoprotein etc. is presumably bound as an integral part of molecules composing the membrane material, the physical process may also be initiated by reversible chemical transformations. Chemistry and Biochemistry at Low Temperatures 179 such as enzymes are soluble in some non-aqueous solvents and that a few enzymes can be recovered with virtually their full potency. Since some of the solvents have melting points below that of water they can be utilized for investi- gations of solutions of proteins at relatively low temperatures. It appears entirely possible that had the solution process been carried out at lower tempera- ture a larger fraction of the enzymes would have been recovered without deterioration. Indeed it may prove fruitful to undertake studies at low tempera- tures of the first stages of reactions which are toxic at ordinary temperatures since the toxic substances may be removed at temperatures so low that httle permanent injury is done to the enzyme or organism. In analogy with the dissolution of finely divided chlorophyll and iodine by solvents at low temperatures it is to be expected that at low temperatures heterogeneous reactions are also possible between substances in solution and biological materials having high specific areas. Ready-made for such reactions with solutions seem sections of tissue with water removed by freeze-drying. Likewise Becquerel's procedure of removing water by pumping at room temperature would prepare material for reaction at low temperature. Some of the reactions with the surfaces constitute a generalized staining. Many staining processes are acid-base reactions and would be expected to be rather rapid at low temperatures. As has been remarked, molecular steric factors are as a rule more specific at the lower temperatures in general; hence finer dis- criminations between structures within the surfaces are to be anticipated. REFERENCES 1. E. V. Sayre, D. G. Miller, and S. Freed: Symmetries of electric fields about ions in solutions. Absorption and fluorescence spectra of europic chloride in water, methanol, and ethanol. /. Cfiem. Phys. 26, 109-113 (1957). 2. D. G. Harris and F. P. Zcheile: Effects of solvent upon absorption spectra of chloro- phylls A and B. Bot. Gaz. 104, 515-527 (1943). 3. A.U.Smith: Eifects of low temperatures on living cells and tissues. In: Biological Applica- tions of Freezing and Drying, tA. by R. J. Harris, 1-62, Academic Press, New York (1954). 4. A. S. Parkes and A. U. Smith: Regeneration of rat ovarian tissue grafted after exposure to low temperatures. Proc. Roy. Soc, (B) 140, 455-470 London (1953). 5. R. K. Andjus: Sur la possibilite de ranimer le rat adulte refroidi jusqu'a proximite du point de congelation. C.R. Acad. Sci., Paris 232, 1591-1593 (1951). 6. R. K. Andjus and A. U. Smith: Revival of hypothermic rats after arrest of circulation and respiration. J. Physiol. 123, 66-67 (1954). 7. J. E. Lovelock: Physical instability and thermal shock in red cells. Nature, Lond. 173, 659-666 (1954). 8. M. J. Loiseleur: Sur quelques proprietes des proteides en solutions organique. Bull. Soc. Chim. Biol. 14, 1088-1100 (1932). J. J. Katz: Anhydrous hydrogen fluoride as a solvent for proteins and some other bio- logically important substances. Arch. Biochem. Biophys. 51, 293-305 (1954). E. D. Rees and S. J. Singer: A preliminary study of the properties of proteins in some nonaqueous solvents. Arch. Biochem. Biophys. 63, 144-159 (1956). 1 80 Simon Freed DISCUSSION Mahler : I can see where this might be useful in the study of the rate of formation of enzyme-substrate complexes. This is a reaction which proceeds much too rapidly to be measured by most ordinary techniques. It is only with very rare and very stable enzyme complexes and by using very interesting and very sensitive experimental devices that Chance*, for instance, has been able to study this at ordinary temperatures. But if one can find the right kind of solvent for both substrate and enzyme — -there is no reason to assume that some of these solvents might not work — one might be able spectroscopically to study the rate of formation of enzyme-substrate complexes at low temperatures. * B. Chance and G. R. Williams: The respiratory chain and oxidative phosphorylation. In: Advances in Enzymology, ed. by F. F. Nord 17, 65-134. Interscience, New York. (1956). INFORMATION CONTENT OF TRACER DATA WITH RESPECT TO STEADY-STATE SYSTEMS* MoNES Berman and Robert L. Schoenfeld Division of Biophysics, Sloan-Kettering Institute, New York Abstract — A method for the quantification of information in data from tracer experiments on steady-state systems is presented. It is shown that if the system is represented by n com- partments a point in an n^ dimensional space can serve to represent a specific model. Further- more, uncertainty about the system due to statistical fluctuations and incomplete data can be represented by regions in the n"^ dimensional hyperspace. A unit of information for such a system is defined and serves as a measure of the amount of information necessary to determine the system to within a desired accuracy. In order to express the data in terms of the generalized n^ dimensional space, a set of invariants is defined for the data. A concise matrix relation is shown to exist between the invariants of the data and the parameters that characterize the compartmental system. The matrix relation allows mappings between the data and the system. The method presented is applicable to any compartmentalized system that shows linear kinetics. I. INTRODUCTION This paper is concerned with the quantification of information contained in data from tracer experiments performed on steady-state biological systems. In general, the same set of data may be analysed in terms of different systems of various degrees of complexity. To define the information content of the data, therefore, it is necessary to specify the system in terms of which the data are to be analysed. It can be assumed for many tracer experiments that the system! consists of a discrete number of compartments (or pools) each representing a locali- zation or chemical state of the labeled material, with exchange of molecules between compartments. The rate of exchange of the unlabeled molecules between compartments is in general a non-linear function of the amounts of material in the compartments. If, however, the system is in a steady state and the amount of the tracer is sufficiently small compared to its unlabeled isotope, the rate of exchange of the tracer may be treated as a linear function of the amounts of labeled material in the compartments (1). The problems that arise in treating the data of tracer experiments are: first, to define the information content in the data, and second, to translate the information in the data into values of the system parameters (the turn-over rates of the compartments). In addition, it is desirable to have a measure of * This work was supported in part by the U.S. Atomic Energy Commission Grant AT(30-1)-910. t For this paper, the word 'system' will be used to mean a specific number of compartments independently of how they are interconnected. The word 'model' will refer to a specific configuration of the system. 181 13 182 MoNES Berman and Robert L. Schoenfeld uncertainty in the values determined for the system parameters. The uncer- tainty in these values arises from the fact that the collected data may not be sufficient to define the system completely and that the collected data have associated fluctuations. A method for the quantification of the information in data and the systematic formulation of models consistent with it is presented here. The information content in the data is expressed by a set of invariants, and a concise matrix relation is shown to exist between the invariants of the data and the system parameters. Uncertainties in the data due to incompleteness or fluctuations are mapped into a generalized co-ordinate space which also represents the degrees of freedom of the system parameters and their uncertainty. The uncertainties in the data are expressed in terms of regions in the generalized co-ordinate space in such a way as to suggest a criterion for their quantification with respect to the system. II. DATA INVARIANTS AND SYSTEM PARAMETERS The response of the system to a tracer injected into any one compartment can be expressed in terms of the amounts of tracer in the various compartments as a function of time. If we define the probability per unit time for a transition from any compartment / to compartment j as A^^, then the kinetics of the tracer in the /th compartment of an n compartmental system can be represented by the following set of differential equations : ^^ = -K^iit) +lh^qlt) (/ = 1, 2, • • •, n) (1) where ^^(0 is the amount of tracer material in the ;th compartment at time t and hi ^ i hi (2) is the probability per unit time that any molecule in compartment / will leave that compartment. The inequality sign expresses the possibility that a molecule may leave the entire system from compartment / as in the case for open systems. The solution of the set of differential equations (1) is: n q,{t) = I A,, e-^' (3) i=i In a recent paper (2) we have pointed out that data expressed in the form of equation (3) have the following properties: (a) There are at most n a^- in the data and these are invariants of the system and independent of the initial conditions or site of measurements. (b) The Ay.^ represent n^ independent variables in the data. Specification of the initial conditions reduces the Aj.^ to {n^ — n) independent variables which are a function of the system parameters only. The Aj^^ thus represent {n^ — n) invariants of the system parameters. Information Content of Tracer Data With Respect to Steady-state Systems 183 (c) The n a,- and rr" Aj.j comprise a necessary and sufficient set of data to define uniquely the parameters of the system. (d) A simple matrix relation (3) exists between the Aj^^ and a,- of the data and the A,y of the system. This relation can be written: Ml = \A l«l or where I3I — Aj2 — Aj3 /122 — "^ 23 1 32 a = h 33 ai (-11 (4) (5) A, 11 1 '31 All -^22 ^32 '13 '23 ^33 0 0 0 0 iy.2 0 0" a. Equation (5) expresses the system parameters in terms of the invariants in the data. If these invariants are known, the fractional turnover rates, Aj-;, can all be determined. However, in most cases the experimental data are incomplete in that certain of the A^j and a^ are not known. For these cases, an infinity of models mathematically consistent with the data can be obtained from equation (5) by inserting arbitrary values for the unknown Aj.j and a^, preserving the initial conditions and other constraints in the data. Most of these arbitrary models, however, will be physically meaningless because some of the fractional turnover rates will be negative. Consequently, it is necessary to investigate what range of values of the unknown A^^ and a, correspond to physically meaningful models. This can be done by relating variations in A^^j and a.; to variations in the X^j. One may define (2) a matrix |P| in such a way that the product \PA\ will preserve the known A^j. The number of variables in \P\ will be equal to the degrees of freedom in the Aj^j. If both sides of equation (4) are premultiplied by the matrix \P\ this equation can be rewritten: (6) (7) a \PXP-'^\ \PA\ = \PA\ which is of the form [A'l l^'l = |/1'| |a| where M'l = l^^l (8) \l'\ = \PKP-^\ (9) Equation (9) expresses a mapping of the matrix \X\ corresponding to varia- tions in the unknown Aj,j only. It also represents a general solution of all models mathematically consistent with the data in terms of a minimum number of variables. This solution is expressed in terms of an arbitrary model represented by the matrix \X\. Similarly, we can define a matrix \D\ so that the product |aZ)| will preserve all the known a^. Incorporating this into equation (4), we get |;.^Z)^-i||^| = |y4||aZ)| (10) 1 84 MoNES Herman and Robert L. Schoenfeld which is of the fonn mMI = MIH (11) where |a I = |ax^| |A'| = \UDA-^ (12 Equation (12) represents a mapping of the matrix |A| in terms of the variations in the unknown a^ only. By applying the restriction that every fractional turnover rate must be positive, r,, ^ 0 A',,^iA',i (13) i = \ equations (9) and (12) limit the range of values of the variables in the matrices \P\ and \D\. Since these variables are all independent, they represent a co-ordinate space of dimension equal to their number. Every point in this space specifies a set of values for the variables in the matrices |P| and \D\ and, thus, defines a model through equations (9) and (12). The restrictions on the range of values of the variables as expressed by equation (13) correspond to a region in the co-ordinate space in which all physically meaningful models must lie. The choice of the starting point for the transformations indicated above is completely arbitrary and does not affect the final result. Any mathematically consistent model leads to a region in the mapping space corresponding to proper physical models. III. UNCERTAINTY MAPPINGS IN GENERALIZED SPACE We now wish to examine the problem from a somewhat different point of view. The system is represented by n^ X^^, generally independent of each other. We can, therefore, consider the X^^ to represent an n^ dimensional space, and any point in that space as a specific model of the system. It was also indicated earlier that the data could be represented by a set of invariants composed of n oij and {n^ — n) A^j or a total of n^ invariants. Hence, the transformation from the data space to the X^^ space is dimensionally consistent and unique. This means that a complete set of A^j and a^ corresponds to a point in the \X\ space, and vice versa. By definition, however, the values of the A,,- must all be positive. Consequently all the models must lie in a restricted region of the \X\ hyperspace. This restriction carries over to the data space, limiting the region in which the Aj^j and a^ may lie. Any specified A^j or a_, implies a one dimensional constraint in the data space. This carries over as a one dimensional constraint in the \X\ space, and restricts all models to a surface in the hyperspace. If, however, the value of A^j or oij is known only within a certain range, the surface has correspondingly a certain thickness. When several A^^ or a^ are known, the dimensions of the space in which all models must he is reduced by a corresponding number. Statistical uncertainties Information Content of Tracer Data With Respect to Steady-state Systems 185 for any of the known values correspond to similar uncertainties along the appropriate co-ordinates in the hyperspace. Thus, if all Aj,j and a^ are known exactly, a point in the hyperspace of n^ dimensions specifies the model. If all the data are known to within a certain statistical precision, the most likely model is estimated as a point in the n~ dimensional space surrounded by a region that corresponds to the statistical uncertainty. If some Aj^j or a_, are unknown, the corresponding dimensions in the n^ dimensional hyperspace extend to the limits imposed by the relation that all Xjj are positive. IV. UNIT OF UNCERTAINTY Based on the point of view presented, we can define a unit of uncertainty to be a certain volume of the hyperspace. The size of the volume so defined is arbitrary; it may correspond to a volume that is equivalent to the actual standard deviation in the data, or to some convenient standard deviation that may serve as a reference. The information necessary to define the system can then be expressed as the number of binary choices, or bits of information, necessary to reduce the total uncertainty space to the size of a defined unit. V. CONCLUSION The treatment presented provides a framework in which information in data from tracer experiments on steady-state systems can be quantified in terms of a compartmental system and its parameters. Before the information can be quantified, however, a number of compartments has to be chosen for the system. Unless this is known from independent sources, the method in choosing the number of compartments is based on the minimum number of exponential terms that 'reasonably' describe the data. This, at present, is by no means a unique procedure. It was shown in this treatment that a model representing the system can be expressed as a point in a generalized co-ordinate space, and that any uncertainty in the system can be represented by a certain region in that space. The nature of the uncertainty (whether incomplete data or statistical fluctuations in the data) did not matter in the treatment. There is, however, one difference in the regions of the hyperspace corre- sponding to these two sources of uncertainty. The difference is in the probability that any model in the region represents the true system. In the case of incomplete data, the probability density over the entire region is assumed constant; that is, every model in the region is considered equally probable. In the case of statistical fluctuations, however, a certain point or unit volume represents the most likely model, and the rest of the points or unit volumes decrease in probability in a manner governed by the statistics of the data. The region in the |A| hyperspace can serve to define the information content in the data of the system as a whole or of each parameter of the system, namely the turn-over rates, separately. The latter can be obtained by investigating their values over the bounded region. One need not necessarily deal with all the dimensions of the hyperspace. One can express the uncertainties in terms of a subspace whose dimensions are equal ] 86 MoNES Berman and Robert L. Schoenfeld to the degrees of freedom of the system, as imphed by equations (9) and (12). In this case, however, the statistical variations of the collected data cannot be represented since their dimensions are omitted. Any new data to be collected, however, can be represented in this subspace. The significance of any new data can also be evaluated by the relative reduction in the size of the region in the subspace. A unit of uncertainty may be defined for this subspace as was done for the hyperspace. In references (1) and (2) it was shown how information about the system from steady-state measurements and thermodynamic considerations can be combined with tracer data to form a unified methodology in reducing the uncertainty about the system. The treatment presented here can be extended to include such additional information. Whereas the concepts presented here are relatively simple, the application to specific problems involves considerable work. One can handle two or three compartmental systems with few degrees of freedom fairly easily using a desk calculator. The handling of more complex systems becomes quite time con- suming. It is hoped that a programming of this on digital computers can be worked out for routine applications. REFERENCES 1. M. Berman: The formulation of biological models from tracer and steady-state data. Ph.D. Thesis, Polytechnic Institute of Brooklyn (unpubUshed) (1957). 2. M. Berman and R. Schoenfeld: Invariants in experimental data on linear kinetics and the formulation of models. /. Appl. Phys., 27, 1361-1370 (1956). 3. H. Margenau and G. M. Murphy: The Mathematics of Physics and Chemistry, chap. 10, Van Nostrand, New York (1943). THE DOMAIN OF INFORMATION THEORY IN BIOLOGY* Henry Quastler Brookhaven National Laboratory, Upton, New York In the proper course of events, a theory is introduced to account for a specific body of facts ; then nobody will presume to expatiate upon the domain of the theory. With information theory and biology, the situation is less simple. The modern development of the theory stems largely from C. E. Shannon's concern with certain problems of communication engineering (1). I have heard Shannon say that he was somewhat dubious about the extension of his results to remote fields, and that he felt that people working in other disciplines might do better to develop their own theories. This is not what happened. Shannon's theory has been taken up with enthusiasm by psychologists, linguists, historians, planners, librarians, sociologists, and by biologists with a wide variety of interests. Motives for such generalizations were supplied by Wiener, who pointed out that all control (in the animal and in the machine) depended on communication, and that all communication involved measurable quantities of information (2) ; and by Weaver, who emphasized the great generality of the information concepts in a searching study (1). It appeared then that information theory was a tool made to order to deal with a vast variety of problems. This variety, however, is not limitless. There- fore, a discourse on the domain of information theory is indicated. One part of this discourse will deal with the negative domain, or with some of the limita- tions of the theory. The other part will be concerned with positive applications ; it is largely an attempt to give clearer definition to the somewhat vague hopes most people have when proposing to apply information theory. It is curious that applied information theory produces rather violent reactions, some of them negative. Certainly, it is entirely possible that every biologist who works with information theory, or any other systems theory, is wasting his time. But this, of course, applies to anybody who works with a new theory. It is difficult to see how applying information theory should irritate people — unless the cause should be the very pleasure of gently playing with the theory. Every scientist is aware that there is a 'difference between the labor of thought, and the sport of musing', and knows well the danger inherent in the latter. To go on with Dr Johnson: 'There is nothing more fatal to a man whose business is to think, than to have learned the art of regaling his mind with those airy gratifications .... This is a formidable and obstinate disease of the intellect, of which, when it has once become radicated in time, the remedy is one of the hardest tasks of reason and of virtue. Its slightest attacks, therefore, should be * Research carried out at Brookhaven National Laboratory under the auspices of the U.S. Atomic Energy Commission. 187 188 Henry Quastler watchfully opposed' (from The Rambler). Is this why so many scientists do not mind too much having collected a lot of useless data but dread to be found working with a useless theory ? I. APPLICATIONS Every kind of structure and every kind of process has its informational aspect and can be associated with information functions. In this sense, the domain of information theory is universal — that is, information analysis can be applied to absolutely anything. The question is only what applications are useful. 1 . Use of Basic Concepts The basic concepts of information theory — measures of information, of noise, of constraint, of redundancy — establish the possibility of associating precise (although relative!) measures with things like form, specificity, lawful- ness, structure, degree of organization. This alluring promise has introduced the information concepts into the thinking of many biologists. The results of conceptual applications range from harmless modernisms of language to very serious reasoning. In particular, the information concepts seem to lend them- selves readily to dealing with the problems of emergence and destruction of order in complicated systems. The problem of emergence of order is usually treated in terms of Darwinian machines, large more or less random assemblies of parts which can both function and, in some manner, register the results of their functioning. The resulting feedback loop produces some order amazingly fast (3, 4). The theory of random networks is a very active field, and some very competent men expect that the main contribution of information theory to biology (and to other fields concerned with very complicated systems) will come from this endeavour. Closely related is the problem of destruction of orderhness. In biology, this is the problem of aging and decay; it is the topic of a major fraction of this conference (5, 6, 7). 2. The Representation Theorem The use of the basic concepts of information theory becomes more powerful if one considers that the behavior of information measures follows certain rules; these rules are the theorems of information theory. There are two basic theorems which I like to call the 'representation theorem' and the 'noise-and-redundancy theorem'. The first has to do with the possibility of representing one kind of information by another kind of information. There are absolutely no quahtative limitations as to how information can be represented ; but, there is a quantita- tive limitation: any physical entity can assume only a limited number of distinguishable states, and this limits the degree to which it can represent information. This degree is further modified by the rules of selecting successive states. The applicability of the representation theorem depends to a high degree on knowing the process by which states are selected. The representation theorem applies every time information is transferred — because the transfer does involve representation of the information existing The Domain of Information Theory in Biology 189 in the transmitter, in the medium and, finally, in the receiver. It can thus be stated as follows: A source cannot transmit more information than it has, a receiver cannot register more information than it can display. This sounds trivial, but the point is that information contents can be precisely estimated in ways which are not trivial. The representation theorem implies that it is possible to establish an upper bound of the flow of information simply by investigating the terminals. It is, thus, a one-sided conservation principle; being one-sided, it is not as strong as the two-sided conservation principles which are so commonly used in physics. It becomes stronger in situations where one may assume that the inequality approaches an equality. There are two conditions which are conducive to the establishment of full conservation of information: one, that information is a valuable and critical commodity, and two, that noise can be minimized. The concept that informa- tion is the most precious commodity for living things has been formulated strikingly by Schroedinger in his assertion that 'living things feed on orderli- ness'— that they feed because they need fresh supplies of orderliness, not of energy or matter (8). The need for fresh supplies of orderliness presupposes that orderliness is somewhere lost, that is, that noise is present. This, however, does not mean that noise is present everywhere. Some processes may occur in 'clockwork fashion', without loss of information. That is the case which Schroedinger classifies as 'generation of order from order'. He suspects that each individual act of transmission of genetic information from parent to offspring occurs without serious loss of information. This idea agrees with the current (Watson-Crick) model of DNA duplication; it recurs in Gamow's and YcAs' models of information transmission from genetic to somatic material (9). 3. The Noise-and- Redundancy Theorem Infonnation transfer from one body of information to another is not often with clockwork regularity. As a rule, interferences occur which will more or less affect the process of information interaction. Interference can be of many kinds: the worst kind of interference is one the results of which are not pre- dictable in detail. In this case, some information will be irretrievably lost. However, in general some but not all order is lost. It is one of the most significant results of information theory to have shown that order and disorder can be measured by a common yardstick. Hence, it is possible to investigate the quantitative relations between total information, noise, and remaining order- liness. The second basic theorem of information theory states that the amount of information effectively transmitted is exactly the amount of information transmitted minus the amount of information lost because of noise. This implies that a source can transmit a certain amount of information reliably in the presence of noise provided it transmits more than the desired amount of information. This surplus must be distributed over the whole activity because it is never known which portions of the total activity will be interfered with by noise; necessarily, the surplus takes the form of redundant information. Thus, the second fundamental theorem states precisely the relation between amount of information to be transmitted, amount of information which will be lost through noise, and amount of redundant information needed to make up the loss. Like the first fundamental theorem, it is a one-sided conservation principle; it limits 190 Henry Quastler the amount of order which can prevail in an 'order-from-disorder' situation. Again, the one-sided conservation principle will become more powerful if it can be assumed to approximate a two-sided conservation. However, very stringent conditions must be fulfilled if one expects to use the second theorem. There is some reason to believe that these conditions are at least approximated in some biological situations; this is stated in Dancoff's principle (10). Dancoff' s principle deals with the economics of information. In 'noisy' situations, information is lost and errors will occur unless they are checked by redundant information. Now, errors may be costly, but so is redundant information; accordingly, the optimum amount of redundant information will be not that which makes all errors vanish, but that which minimizes the sum of the cost of errors plus the cost of redundant information, plus the cost — in information units — of error checking. Dancoff's principle asserts that any organism or organization which has gone through competitive evolution has approximated such an optimum; that is, it will commit as many errors as it can get away with, and use the minimum of redundant information needed to hold errors to this level. It follows from Dancoff's principle that the amount of redundant information in a system is bound to be limited, even if it is a system of enormous information content like a living thing. This is of great interest particularly in radiobiology, because what radiation does very effectively is to destroy information. 4. The Estimation of Information Measures and the Search for Invariants It may well turn out that the qualitative and semi-qualitative applications of information concepts are going to be the most important contribution of information theory to biology. But, even successful qualitative applications have very little power in excluding the possibility that other sets of concepts could have been used just as successfully; besides, all scientists like to take measures. Thus, the problem arises of estimating information measures associated with biological structures and functions. One fundamental diflficulty appears immediately: information measures are relative and not absolute ; hence, any information measure associated with a given set of biological objects will depend on the set itself and on the scientist who does the estimating. To be sure, one can establish objective bounds. Thus, if a certain genetic locus is known to be capable of having thirty-two distinct allelic states, which are transmitted to the offspring with equal prob- ability given the proper conditions, then the information stored in this locus cannot be less than five bits. If it is also known that the region containing the locus under consideration comprises no more than, say, 20,000 atoms, then the total information stored cannot be more than about 60,000 bits (10). These brackets are safe, but they are too wide to be of interest. They can be very much reduced if one introduces specific assumptions. For instance, if the locus is known to contain no more than, say, 2 X 50 nucleic acid residues, and if one assumes that the genetic information is completely coded in the sequence of the residues on one strand of a double helix, with the information carried by each residue corresponding to unconstrained selection from four possibilities, then the upper bound is reduced to 100 bits — but its validity is less absolute. The Domain of Information Theory in Biology 191 Because of the relative nature of information measures, it will always be up to the ingenuity of the biologist to find ensembles which result in useful measures. In many cases, even the estimation of a limit is of interest: as in Ehret's demonstration that a few bits could be sufficient to specify the nature of cytoplasmic structures (11), or the result easily derived from D'Arcy Thompson's work (12) that apparently considerable differences in fonn could be coded in, say, a few nucleic acid residues. The relativism of information measures is a basic difficulty in estimation ; besides, the biologist will encounter a number of technical difficulties arising from the fact that 'message sets' and 'selection rules' are not perfectly known. A number of approximation methods for such situations have been worked out (13). The relative nature of information measures and the technical difficulties of their estimation, cast some doubts on the usefulness of actual information measures in biology. Only experience will show whether these doubts are justified or not. Measures will be valuable if they lead to the discovery of invariants. In psychology, some invariants seem to be crystaUizing out of a number of measurements: there seem to be invariant upper limits for the channel capacity for single activities; for the range of classes distinguishable in a single act, etc. (14). In biology, independent estimates of information transfer associated with three elementary biological functions (allelic, anti- genic, enzymatic specificity) have yielded closely similar values (15). Much more material will be needed before we can draw definite conclusions. The analysis which underlies the estimation of information measures presents certain novel features. Consider, for instance, the informational analysis of a honnonal control system. The traditional approach consists in isolating one hormonal function and one hormone after the other. In principle, this quest never ends — although physiologists might hope that some day they will run out of undiscovered hormones. The information theorist attacks the problem from the opposite end. He will argue that each hormone molecule constitutes a message from a control organ to a target organ, a message which is diffusely broadcast through the blood stream. In general each message must contain two parts, an address and an order. Actually, one or the other part can be omitted. We can imagine a hormonal control system in wliich only the addresses are specified — the 'order' may be completely determined in the target organ, and be executed automatically upon receipt of the only kind of hormone molecule with the proper address; or, the address may be unspecific, but the order such that only the right target organ can execute it. One would expect that the natural systems be somewhere between these two extremes. For the sake of simplicity we will consider a system in which only addresses are specified — the foiTnal results have complete generality. Thus, each hormone will be represented only by the address of the target organ. In the interest of detailed and accurate control, it is desirable to have a maximum number of different addresses. Any duplication of addresses will lead to concomitant responses in other organs. On the other hand, the 'reading' of every single address involves distinguishing it from all other addresses; the greater the variety of addresses, the greater the labor in every single act of recognition. A compromise is indicated between the demand 192 Henry Quastler for a great variety of addresses and the contradictory demand to keep each address simple. For any kind of system, there will be an optimum number of different hormones; the actual number will depend on the relative strength of the two competing needs. By Dancoflf's principle, we expect that the actual number will not be too far from the optimum number. We can add another line of considerations on the number of possible addresses. In order to fulfill its function, the hormone molecule has to enter into some kind of relation with the target organ ; most likely, it has to form a complex. Now, the total surface area of any molecule that can enter into a specific complexing process is rather limited, and so is the number of molecular con- figurations available to living organisms; hence, a limited space accommodates only a limited number of significantly different configurations — and this limits the number of different hormones possible (and, incidentally, the number of distinct antigens and antibodies, enzymes and co-enzymes). The example illustrates the concern with the whole system which is charac- teristic of many applications of infomiation theory. It also illustrates a rather profound difference between the information theorist and many of his scientific colleagues. The information theorist will remain fairly cool at the news that another enzyme, or hormone, or vitamin has been isolated ; his basic question is: 'How many more are there to be discovered?' II. LIMITATIONS Information theory could not possibly apply to a wide variety of situations if it were sensitive to every detail in every situation. Like thermodynamics (to which information theory is related) it has a vast domain of application, and like in thermodynamics, the vastness of the domain is paid for by a limited scope of every single application (16). The following four limitations deserve emphasis : (i) information measures refer to ensembles and not to single instances, (ii) they are relative and not absolute, (iii) informational capabilities are often not fully utilized, (iv) information measures are related to other aspects of systems such as utilities and mechanisms but the relations are not simple. None of these observations is particularly profound, but each one has been overlooked by competent investigators. 1 . Information Measures are Functions of Ensembles Information measures are not defined for particular historical occurrences or existing individual things; rather, they are defined for whole ensembles of events that could happen, or things that could be. The information measures are descriptive of the operations by which a particular item is selected from the set of possible items, and are associated with the whole set and not with any particular item that happened to be selected in a particular instance. Ensembles are specified by their elements, by the classification to which these elements are subjected, and by the probability measures associated with the diverse classes. If these specifications are known, then the information functions can be derived — but not vice versa. For example: if it is known that a certain chemical system contains certain enzymes and certain substrates, if the probabilities of the various coUisions and the probabilities of all possible The Domain of Information Theory in Biology 193 outcomes of such collisions are known, then it is possible to derive a number of information functions for this system; on the other hand, a given set of information functions is compatible with any number of chemical systems. 2. Relativity of Information Measures In the early applications of information theory to problems of communi- cation, the ensembles to be used were virtually defined by the situation. Thus, in dealing with Morse code, the element is clearly the single symbol, the classes are dot, dash, letter space and word space, the probabilities are the large sample frequencies. Similarly, in dealing with printed English as an objective phenomenon, one natural unit (not the only one, though!) is again the single symbol, and classes and probabilities can be determined from any large sample. The situation is immediately more complicated if we deal with a particular person's concepts of printed English; the 'subjective probabihties' are not the same as the objective relative frequencies. Much confusion has come to psychologists from disregarding the fact that the probability measures upon which a subject bases his operations are not necessarily those known to be correct — in one sense — to the experimenter (17). There are situations where there is considerable leeway in defining the elements, classifications, and probability measures of an ensemble, and accord- ingly considerable variation in the infonnation measures which can be associated with the situation. This is strikingly illustrated by the attempts to measure the infonnation contents of molecules. Estimates have been based on con- siderations of structure (10, 18, 19, 20) or function (15, 22). Recently, Rashev- SKY and his associates (21) have shown that information measures can be associated with the topological representation of molecules. Each of these approaches yields some value of the information content of a molecule, and these values do not have to be identical. Yet, every one of them is a legitimate information measure. This may be disappointing, but hke all abstractions, information measures are not 'right' or 'wrong' — they are only more or less useful. In the case under discussion, we may legitimately ask how the various ways of estimating information measures are related to the actual processes of information storage and transmission by molecules, to reaction rates, to the activity of antimetabolites, etc. As a rule, the specifications of an ensemble do not result unequivocally from the given situation. Consequently, information measures are not properties but functions of a given situation — they are defined by the situation and the ensemble used in dealing with it. Information measures are irreducibly relative; they can be accurate and precise, but they cannot be absolute. The usefulness of a particular information measure in a particular context will depend on the way the defining ensemble is set up. Unfortunately, there exists no calculus, no set of hard and fast rules which tells one how to select the most appropriate elements, classifications, and probability measures. The choice must be made by guess, and its ultimate justification is only in the results it yields. 3. Informational Capabilities and Performance An informational capability represents an upper bound to some class of informational performances — but a particular performance does not have to 194 Henry Quastler approach this bound. One cannot transmit information through a channel at a rate higher than the channel capacity, but it is very easy to transmit at a lower rate. For instance: human capacity of transmitting information can be limited on the input side, on the output side, or centrally; if the limitation is central, then it can be due (a) to the limited channel capacity, but also to limitations of (b) the rate at which discrete acts of information-processing can be performed, of (c) the amount of information per single act, of (d) the number of information-carrying components considered in each act, of (e) the maximum amount of information per component, or finally, (f ) to inefficient coding (14). Parallel situations are likely to exist in molecular biology. For instance, Augenstine (18) discusses the fact that the amount of information which can be coded into an amino acid sequence is considerably greater than the amount of information needed to account for the functional specificity of a protein. This could mean that the channel capacity is only fractionally utilized, or that functional specificity is coded in an entirely different fashion. 4. Information Measures and Other Aspects of Systems If the mechanism of a reaction is known, then the probabilities of all input- output associations can be computed, and the information measures derived from them. On the other hand, an infomiation measure does not define a single mechanism — however, it imposes a condition with which input-output tables and, by implication, mechanisms have to comply. For instance, in the problem of the DNA-protein code studied by Gamow and Ycas (9), "" • the infoiTnational analysis furnishes conditions which the code must fulfill but does not yield the code itself. Accordingly, the informational analysis has served, repeatedly, to reject a proposed mechanism. It can, of course, never be used to prove a mechanism. Amount of information is in general related to the utility of being informed — but the relation is not necessarily one of simple proportionality; in fact, the utility of information is not always a monotonically increasing function of its amount. Similarly, the information content of a structure is in general related to the difficulty of construction, but the relation is not one of simple proportionality. The 'amount of information' in a statement is related to its capacity of carrying semantic information, but this capacity is rarely fully utilized (23). III. CONCLUSION I have tried to outline some of the applications and possible applications, and I hope to have shown that there is much promise in this field. I have tried to outHne some of the limitations of applying information theory— and I hope to have shown that they are not serious, provided one is always aware of them. To make more progress, we need much more mathematical work, and we need very much more experimental work. In looking over the past of information theory in biology, a very strong emphasis on theory— more or hss rigorous— is obvious; although more theory is needed, the most pressing need is now for a large body of good specific experiments. Also, it should be rewarding to examine closely other related possibilities in theoretical biology. The Domain of Information Theory in Biology 195 For some reason, our time has been (and still is) exceedingly fertile in producing theories dealing with systems, or in reviving and greatly expanding existing theories of systems. Information theory is one of several system theories — a very rewarding one, I believe, but one with very specific limitations; it should be useful to search specifically for theories with different limitations to supple- ment information theory. REFERENCES 1 . C. E. Shannon and W. Weaver: The Mathematical Theory of Communication, University of Illinois Press, Urbana (1949). 2. N. Wiener: Cybernetics, J. Wiley and Sons, New York (1948). 3. R. W. Ashby: Design for a Brain, J. Wiley and Sons, New York (1952). 4. M. Eden: A probabilistic model for morphogenesis. This volume. 5. H. P. Yockey: An application of information theory to the physics of tissue damage. Radiat. Res. 5, 146-155 (1956); and, A study of aging, thermal killing, and radiation damage by information theory. This volume. 6. H. B. Jones: Some notes on aging. This volume. 7. H. A. Blair: A quantitative description of latent injury from ionizing radiation. This volume. 8. E. Schroedinger : What is Life? University Press, Cambridge, England (1944). 9. G. Gamow and M. Ycas : The cryptographic approach to problems of protein synthesis. This volume. M. Ycas: The Protein Text. This volume. 10. S. M. Dancoff and H. Quastler: The information content and error rate of living things. In: Information Theory in Biology, ed. by H. Quastler, 263-273, University of Illinois Press, Urbana (1953). 11. C. Ehret: Information content and biotopology of the cell in terms of cell organelles. This volume. G. Karreman: Topological information content and chemical reactions. Bull. Math. Biophys. 17, 279-285 (1955). E. Trucco: a note on the information content of graphs. Bull. Math. Biophys. 18, 129-135, (1956). 12. D'A. W. Thompson: Growth and Form, Cambridge University Press, Cambridge, England (1952). 13. H. Quastler: Approximate estimation of information measures. \n: Information Theory in Psychology, ed. by H. Quastler, 124-139. The Free Press, Glencoe, 111. (1955). 14. H. Quastler: Studies on human channel capacity. In: Information Theory, ed. by C. Cherry, Academic Press, New York (1956). 15. H. Quastler: The specificity of elementary biological functions and the measure of specificity. In: Information Theory in Biology, ed. by H. Quastler, 170-188; 41-71, University of Illinois Press, Urbana (1953). 16. L. J. Cronbach: On the non-rational application of information measures in psychology. In: Information Theory in Psychology, ed. by H. Quastler, 14-30. The Free Press, Glencoe, 111. (1955). 17. H. Quastler (ed.): Information Theory in Psychology, The Free Press, Glencoe, 111. (1955). 18. L. G. Augenstine: Protein structure and information content. This volume. 19. H. J. MoROwiTz: Some disorder-order considerations in living systems. Bull. Math. Biophys. 17, 81-87 (1955). 20. H.Branson: Information theory and the structure of proteins. In: Information Theory in Biology, ed. by H. Quastler, 84-104, University of Illinois Press, Urbana (1953). 21. N. Rashevsky: Life, information theory and topology. Bull. Math. Biophys. 11,119-12)5 (1955). 196 Henry Quastler 22. P. D. Klein: Efficiency of information transmission by biochemical co-factors. This volume. 23. R. Carnap and Y. Bar-Hillel: Tech. Rep. Res. Lab. Electr., Mass. Inst. Tech., no. 247 (1952). DISCUSSION A. Rapoport: It is admirable of biologists to look 'up to' physicists and mathematicians, but it is somewhat embarrassing to physicists and mathematicians to be looked upon with such confidence as a source of methodological gifts which can be immediately applied in biological investigations. It is noteworthy that the greatest discoveries of the physicists are stated in 'pessimistic' terms. They are statements about what cannot be done. For example the First Law of Thermodynamics is essentially a definitive demolition of an age-old dream. The law says in effect that the perpetual motion machine cannot be constructed. But it also holds out a hope of a machine that will keep on working provided only that a large supply of heat is available — the so-called perpetual motion machine of the second kind. The Second Law of Thermodynamics puts an end to that dream. It says that such a machine cannot be constructed either and prophesies the 'heat death' of the Universe. Maxwell introduced his demon in the hopeful conjecture that the intervention of an intelligence might restore the order lost by the continual increase of entropy. Szilard in his now classical paper showed that this too is an illusion, that the demon must pay for the restored order by becoming 'disordered' (a biologist would say 'denatured') himself. Yet it would be a mistake to consider these discoveries as admissions of defeat only. Each has brought a broadened understanding; the First Law of Thermodynamics by revealing heat as a source of energy ; the Second Law by revealing the role of entropy. Szilard's investigation rests on quantum-theoretical principles and so provides an important juncture between thermodynamics, information theory, and quantum theory. It appears, therefore, that the grand discoveries of physics have a sobering effect. I think the principles of information iheory are of a similar kind. Typically they are statements of hmitations. Their constructive side is in defining the framework in which the search for new knowledge or for new means of prediction and control must be confined. SOME MEMBRANE PHENOMENA FROM THE POINT OF VIEW OF INFORMATION THEORY* Herman Branson Howard University, Washington 1, D.C., U.S.A. Abstract — The methods of information theory and of irreversible thermodynamics are applied to membranes. Equations are derived for the negentropy production of a membrane maintain- ing a concentration difference. The results are converted to bits. When applied to typical data for a nerve transporting Na+ against a concentration gradient, the equation gives for the negentropy or information production, H=7.3 X 10^^ bits/cm'' second. Enumeration based on Na+ : Cl~ : K+ = 1 : 1 : 10 gives a value of 4.3 X 10'^ bits/cm* second. In a classification of the significant problems of biophysics Danielli (1) listed four. Two of these relate intimately to membranes and their role in biological systems: cell permeability and cell electrophysiology. It is almost mandatory, then, to inquire into the behavior of membranes from the point of view of information theory. For if information theory is to have relevance to important biological problems, a coherent relation should be exhibited for membrane phenomena. This attitude was exhibited in the initial attempts to discuss biological problems within this discipline in Quastler's pioneering book (2). The formidable complexity of biological membranes is a recurring theme in the immense amount of experimental data which are being accumulated. Phenomena encountered in biological membranes may range from those explainable by the assumption of simple pores of various sizes, to those requiring charged pores, and on to those necessitating a picture of the surface as possessing pores, binding sites, permeability barriers, enzymes, and transport mechanisms. It is possible, however, to ignore the details of structure and specific mechanism — as is usually the case in thermodynamics — and formulate a limited model of membrane activity satisfactory to our analysis. Thus membranes may be classified by the manner in which they react to or treat a given substance. If we schematize the membrane as separating two media in each of which the reference substance is soluble, calling one region the 'inside' and the other the 'outside', we may introduce the following notation : Indifferent: A membrane is indifferent to a substance if the concentration of that substance is the same on both sides of the membrane at all times. Thus Q = Q, for all /. * This work has been supported in part by the U.S. Atomic Energy Commission, AT(30-l)-892. 197 14 198 Herman Branson Responsive : A membrane is responsive to a substance if tlie concentrations of that substance differ on each side at the same time. But neither concentration is zero. Thus, Q ^ Cq for some t. Exclusive: A membrane is exclusive with respect to a given substance if, for all time, the concentration on one side is finite but the concentration on the other is zero. Thus, Q = 0 and Q 7^ 0, or Q ^ 0 and Q = 0, for all t. The analysis of this paper will be limited to substances to which a given membrane is responsive. It must not be concluded, however, that indifferent and exclusive substances are of no biological significance. There are examples where seemingly the most important role of a membrane is its action to exclude a given substance from the internal medium or keep a component from diffusing out. The preliminary work (2) on a responsive membrane attempted to derive by the methods of irreversible thermodynamics an expression for the negen- tropy production of a membrane maintaining a concentration difference. The approach rested upon the concept that the entropy is an absolute maximum at equilibrium. Hence any deviation from equilibrium would mean a decrease in the entropy. Expanding the change in entropy, A^, in a Taylor's series about the equilibrium point yields as the first approximation, since the first derivative terms vanish at equilibrium, A5=-l/2 2gj.m«i«m • 0) Equation (1) was combined with some descriptive equations for the membrane and the final result ^ = fta (2) was deduced. In equation (2), H is the negentropy or the information (3), a is the rate constant governing the rapidity with which the membrane would approach uniform concentration on each side if it were not actively maintaining the concentration difference and k is Boltzmann's constant (1.380 X 10"^^ ergs per degree). Equation (2) had to be examined to determine if it is apphcable to a mem- brane maintaining a considerable concentration difference. Its significance with respect to the relation A5ir,. = A: In Q/Q (3) had to be clarified (4). Equation (3) gives the irreversible production of entropy for the passage of a single particle from concentration Q to the concentration Co. If Co is greater than Q-, we have the situation postulated for the membrane, thereupon A^'u-r. is negative and may be equated to information, H. Derivation of the Rate of Production of Information The methods of irreversible thermodynamics as presented by DeGroot (5) are applicable to effect the derivation. Following DeGroot's nomenclature, we have ^S = J^X, + J^nXm + h^, (4) Some Membrane Phenomena from the Point of View of Information Theory 199 where A^" is the rate of entropy production. The J's are 'fluxes' and the A"s are 'forces', u is energy, in matter, and // chemical potential. The J's are related to the A"s through Assuming an isothermal system, the A"s and y's are X, = -^m^T J,-^ dt dt Whereupon, equation (4) becomes AS = -J„, ^fl|T - J^ AmIT. (5) Substituting into equation (5), /* ^ //q + i^rin a where a is the activity, we get A^ = (R In ajao) ^ + 7? Am d/dt (In ajoo). (6) dt The a's refer to the activities in the two different regions. Equation (6) is the basic relation replacing both equations (2) and (3). If we assume that the activities may be replaced by the concentrations and that we are concerned with the passage of A^ particles from a lower to a higher concen- tration, Cq > Q, then equation (6) becomes ^^Ii=k-^ In Co/C, + Nk didt (In CJC,). If the outside is taken to be very large with no change in concentration by the addition of the particles from within, H=A:^lnC./C,--^. (7) Equation (7) gives the rate of decrease of entropy or the rate of production of negentropy or information by a membrane which is transferring material from a lower to a higher concentration where the particles leave at the rate dNjdt and the concentration within changes at the rate dCjjdt. Thus one may look upon equation (7) as the dynamic equation describing real transport. On the other hand equation (2) describes a situation where there is no macroscopically discernible change in concentration within or without. But inasmuch as the membrane maintains a concentration difference, it is producing information at the rate given to continue the imbalance. 200 Herman Branson If the concentration within is stabihzed in such a manner that for a sUght change the system returns to the resting value following the equation (1/Q) dCJdt = -a (8) where a is analogous to the rate constant appearing in equation (2), equation (7) may be written dN i^ = A: — in Co/C, + kN^- (9) Equations (8) and (9) seem harmonious with equation (2) interpreted as being applicable to the resting case when there is no net transport or to actual trans- port with the condition that Cq — Q. The application of equation (9) to a fluctuation should follow this sequence. Initially Cq = Q, and AA'^ particles jump from one solution to the other. The rate of negentropy production or the rate of production of information is /^ = A///A? = k^Noi. At the end of this fluctuation equation (9) becomes for the next fluctuation kANcc IC -1- AC\ n = A:(AA^/A/) In L I ^ J + k^N^. (10) Suppose that the next fluctuation is the movement of A A'' particles in the direction opposite to the first fluctuation in the same interval of time, A?. We should expect fi to be the negative of its original value. Expressing the logarithm 1 + AC/C in equation (10) as In -j . , and making use of the relation In 1^ = 2(A' + A-^/S + • • •), for X'' <\, the equation becomes ii^lk AMI At AC/C + kAN(x but AN I At AC/C = ANIC AC/ A/ = -AA^a from equation (8), and since AC is negative in the second fluctuation. Finally: AH = —kANoL At. Thus the system returns to the equilibrium position on the entropy surface with an increase in entropy exactly equal to the decrease of entropy experienced in the first fluctuation. Analysis for Charged Particles In the derivation of the basic equations (6) and (7), the chemical potential was employed, which limits the applicability of the analysis to uncharged particles. To derive the corresponding equations for an ion, the electrochemical potential /li' replaces ju, fx' = juq' + RT\na-^ZF(i> where Z is the valence, F is the Faraday, and 0 is the potential. Substituting Some Membrane Phenomena from the Point of View of Information Theory 201 and carrying through an identical analysis as followed in deriving equation (7), H becomes A --= k[{dNldt) In CJC, -f A^a] -]- {ZqlT)[cINldt(cf>, - - ,) ^ ^ where q is the numerical value of the charge on the electron, when the membrane transports ions from the lower concentration to the higher concentration, C, - Co. Application to a Nerve Data on the movement of ions across the nerve membrane may be substituted into equation (11) to obtain numerical values for ff. The knowledge of the transport of ions across the nerve membrane although quite extensive is still not complete enough to permit unequivocal choice of a model. There are two possibihties which suggest themselves for consideration. In the first, following the transport of an impulse, the nerve returns to its resting condition with respect to the concentration of Na+ or K+ before it can pass another impulse. The resting potential is reached at the beginning of this period. Calling this example model one, we have these data for a squid axon (6): [K+] Co =10 C, = 410 [Na+] Co = 460 Q = 49 [C1-] Co = 540 Q = 40 in units of millimoles/kg. At 300°K, (f)Q — (f)^ = 50 mV with the outside positive. If the length of the recovery period is taken as 1 millisecond, the equation (11) becomes* n = kidNjdt) [In Co/C, + (ZqlkT)(cf>, - cf>,)]. (12) The first term on the right yields for the concentration gradient above for K+, // = 5.3 bits/ion-unit time, for Na+, H — 3.3 bits/ion-unit time, for Cl~, li = 3.7 bits/ion-unit time. * In applying equation (1 1) the second and fourth terms contribute negligibly. Examining the two terms with the data that 3.7 /iju moles of Na+ enter per impulse per cm*, a cylindrical nerve of 100 ^i radius with unit surface area would have, Area = iTrrl = 1 cm^ Vol = 7Tr^l= 5 X 10-' cm'; and assuming that the nerve has the density of water, the section would weigh 5 x 10' gm. So 49 m mole/kg would correspond to 2.5 x 10^ jufi mole/cm', whereupon for sodium H = 2.25 k{dNldt) + 1.5 x 10"° kN per millisecond = 2.25 kidNjdt), since AAT = N and A/ is taken as 1 millisecond. 202 Herman Branson The term ^ (<^o -«/•.)= 1 -94 at 300°K. Using this value, the combined rates for both the concentration and electrical terms are K+: H = 2.5 bits/ion-unit time, Na+: //= 6. 1 bits/ion-unit time, (13) C1-: H = 0.9 bits/ion-unit time. log2 e These values have been arrived at by multiplying equation (11) by — j — to convert from k units to entropy units. [The conversion factor was derived in a previous paper by the author in considering proteins (2). It is easily shown that log,, x logj, y = 1, which means that this conversion factor is the same as that derived by Linshitz (!) and others (8).] On the basis of the first model the 3.7 /^/^ moles of Na+ entering during the impulse would be extruded during the millisecond following the return to the resting potential. This time interval requires that the nerve produce information at the rate 7^ = 9.3 X 10^5 ergs/°K cm^ sec, or (14) fl=\35 X 1016 bits/cm^sec. The alternative model is that the nerve does not extrude the Na+ in so short a time. Rather the nerve passes it from within to the outside at a rate of 20 i^fjL mole/cm^ second (9). However, the acceptance of this view does not alter equations (13). Inserting this value for Na+ in equation (12) results in if = 7.3 X 1013 bits/cm2 sec. (15) The experimental results seem to favor this model; thus equation (15) is the more reasonable result in comparison with equation (14).* The alternative mode of viewing the nerve according to information theory makes use of the concept that the sodium ions are chosen from the pool of ions within the nerve by some mechanism. Within the nerve the ratios are Na+ : CI" : K+ = 1 : 1 : 10. The mechanism chooses the Na+ from this group, requiring logg 12 bits of infomiation for each Na+ selected. This value, 3.57 bits/ion, leads to i^ = 4.3 X 1013 bits/cm2 sec, (16) for the nerve based on the second model which transports 20/^// moles cm^ sec. of Na+ against the concentration and electrical gradient. Assuming that the * Dr Leroy Augenstine made the astute observation with respect to equation (14) in the discussion that it is consistent with a value of 10^ A" for the area of a protein and that on 1 cm^ of nerve there could be 10^^ proteins transporting one ion per millisecond. It may weaken the argument to assume that the nerve surface has that many protein molecules. The lower value 1.20 X lO^ions/cm^ second is consistent with any combination of rates and numbers of protein molecules responsible for transport such that the product equals this numerical value. Thus 1 .20 X 10" protein molecules transporting an ion per miUisecond are suitable and require that only 0.001 of the nerve surface consists of such proteins, each of 10* A^ area. Some Membrane Phenomena from the Point of View of Information Theory 203 information units, the bits, appearing in equations (15) and (16) are identical with those used in discussing the information content of the printed page, the interpretation is that a cm^ of nerve has an enomious rate of production. The analogy of course is to a person who is called upon to separate red balls (Na+) from white (K+) and blue (CI ) balls in a box where he would reach in, pick up a ball, look at it, and if it is red it is taken out, and if not it is replaced. The nerve has some equivalent separating mechanism with the infomiation rate of equation (15). In terms of a familiar example, taking the information content of a single printed page as 10* bits (9), equation (15) requires that the cm^ of nerve surface produce information equivalent to that contained in a library of 7.3 million volumes of a thousand pages each second — this is over half the number of books in the Library of Congress! The value given by equation (15) is not inordinate, however, in comparison with the estimates of the information content of biological objects (9), where for man the value is of the order of 10^5 bits. The result in equation (16) may be viewed as the minimum information production necessary to effect the separation of sodium. The numerical value may be in error, for the choice could be from among more ions than the three employed. It is of interest to note that the nerve does not possess perfect coding inasmuch as it uses 1 .7 times as much information to effect the separation as is required. Alternatively the information efficiency may be expressed as 59 per cent. These comparisons may be without substance because of the inadequacy of the data. The only relevant comparison may be that the physio- chemical determination of information production as summarized in equation (15) is of the same order of magnitude as the value determined by enumeration in equation (16). REFERENCES 1. J. F. Danielli: Cytological research and biophysics. Nature, Loud. 162, 753 (1948). 2. H. R. Branson: A definition of information from the thermodynamics of irreversible processes. In: Information Theory in Biology, ed. by H. Quastler, 25-40. University of Illinois Press, Urbana (1953). 3. L. Brillouin: Science and Information Theory, Academic Press, New York (1956). 4. R. C. Tolman and P. C. Fine: On the irreversible production of entropy. Rev. Mod. Phys. 20, 51-77 (1948). 5. S. R. DeGroot: Thermodynamics of Irreversible Processes, Interscience, New York (1952). 6. R. W. Stacy, D. T. Williams, R. E. Worden, and R. O. McMorris: Essentials of Bio- logical and Medical Physics, McGraw-Hill, New York (1955). 7. H. LrNSCHiTz: Information and physical entropy. In: Information Theory in Biology. qA. by H. Quastler, 14-15, University of Illinois Press, Urbana (1953). 8. S. M. Dancoff and H. Quastler: The information content and error rate of living things. In: Information Theory in Biology, ed. by H. Quastler, 263-273, University of Illinois Press, Urbana (1953). 9. A. L. Hodgkin and R. D. Keynes: Active transport. Symp. Soc. Exp. Biol. 8, 423-437 (1954). EFFICIENCY OF INFORMATION TRANSMISSION BY BIOCHEMICAL CO-FACTORS* Peter D. Klein Division of Biological and Medical Research, Argonne National Laboratory, Lemont, Illinois Abstract — Amodelfor the transmission of information by biochemical co-factors is described. Two points of information transfer are apparent : formation of the holo-enzyme and formation of the holo-enzyme-substrate complex. The reduction in uncertainty taking place at these points is related to the sets of compounds existing before and after these points, and the values for artificial situations calculated. It is concluded that the artificial situation is an estimate of the minimum selection capabili- ties of the enzyme system. Co-factors are compounds of molecular weight 100 to 2000 which participate in a host of biochemical reactions. They are not metabolised per se, but serve as catalysts. The moiety with which co-factors cooperate may in this instance be limited to proteins. Co-factors for a particular protein may be either exogenous (vitamins) or endogenous (hormones). The flow diagram (Fig. 1) represents the fate of a co-factor in the organism. A particular apo-enzyme can operate on a substrate or class of substrates if provided with the suitable co-factor. In this case the co-factor is assumed to be a vitamin. Therefore, preliminary to its appearance in the cell, the com- pound must of necessity be ingested, absorbed and transported into the cell. Inside the cell, the compound may or may not be excreted again. Each time that it exists in a 'free' form, it has a finite possibility of leaving the site of action, including transformations which lead to the degradation of the molecule so that it cannot function. This possibility is represented by Probability Point 1 . This and subsequent probability points have the following characteristics: A molecule 'passing' through this point may undergo two or more transitions; each state resulting from these transitions has a certain probability but there is no control over the state into which the molecule passes. Next, it may be imagined that a collision between the compound and the apo-enzyme for which it may be destined takes place, leading to the formation of a complex. The formation of this complex, however, depends upon mutual exchange of information between the apo-enzyme and its co-enzyme and is indicated by Decision Point 1 . For example, if the co-enzyme for cocarboxylase collides with the apo-enzyme for riboflavin, no information exchange takes place and there is no complex formation. If, however, sufficient information is exchanged, there results the formation of the holo-enzyme, or in the case of the competitive inhibitor, a pseudo-holo-enzyme. Both forms of holo-enzyme have, of course, a dissociation constant, indicated by Probabihty Point 2. * This work was performed under the auspices of the U.S. Atomic Energy Commission. 204 Efficiency of Information Transmission by Biochemical Co-factors 205 f ' VLf I COMPOUND 1-7— • INGESTED ABSORBED PROBABILITY POINT DECISION POINT TRANSPORTED I [BOUND] REACTION Fig. 1. Pathway of utilization for a hypothetical co-factor. CELLULAR CONSTITUENTS APO-ENZYME ACCEPTORS SUBSTRATE ACCEPTORS Fig. 2. Topological considerations in evaluating information transmission by a co-factor. 206 Peter D. Klein The collision of the holo-enzyme with a potential substrate represents another decision point (Decision Point 2). Again, a complex will be formed if sufficient information is transferred. It is at this point that the pseudo-holo-enzyme must present the wrong information in order to be an effective inhibitor. This results in repeated cychng in the innermost loop diagrammed. The enzyme-substrate complex has a finite probability, designated Probabi- Hty Point 3, of decomposing unchanged before the reaction is catalysed to decompose the substrate into product and regenerate the holo-enzyme. There are, then, two points in tliis flow sheet at which information can be exchanged; between the co-factor and the apo-enzyme and between the holo- enzyme and the substrate. These are the two points to which attention will be devoted. At neither decision point is the decision unequivocal. There are several types of co-factor-substances which may form a complex with the apo-enzyme and several of these complexes are acceptable, though to differing degrees, to the substrate. The situation is graphically presented in the set diagram Fig. 2. The largest circle represents the class of all possible organic compounds. Let B designate these substances. A subset of 5, composed of the organic substances which normally occur in cells, is designated C. Another subset of B, designated A, is comprised of substances acceptable to the apo-enzyme for complex forma- tion. The set A.C'\ includes those compounds normally occurring in cells which are able to complex with the particular apo-enzyme considered ; the set (A~A.C) contains those substances which form a complex with the apo-enzyme but are not normally found in the cell. A subset of A, designated A.S, contains all compounds which complex with A and react with the substrate. (There may be other substances in B which would react with the substrate, if complexed with a proper apo-enzyme ; but these are not of concern here.) These substances which are contained in the set C.A.S are the natural co-factors for the apo-enzyme-substrate pair under consideration; the substances in the set A.S-C.A.S are artificial co-factors; the substances C.A-C.A.S are natural antimetabohtes ; those in the set A-A.S are artificial antimetabolites. The information measures associated with the two decision processes can be derived from the diagram. Let H(X) designate the uncertainty associated with the set X; then: H(C) is the uncertainty of substances in a normal cell. To give this quantity meaning, we shall consider it to be the uncertainty about the nature of an organic molecule which normally collides with the apo-enzyme. H{C.A) is the uncertainty concerning a substance which has formed a complex with the apo-enzyme. H(C.A.S) is the uncertainty concerning the complex which has reacted with the substrate. It should be noted that in dealing with a given apo- enzyme and a given substrate (or class of substrates) all uncertainties in question are due to the co-factor. t The set X. Y contains all substances which belong to both the set X and the set Y. The set JV- 7 contains those substances which are contained in A' but not in Y; alternatively this set is designated X-X. Y, all substances which belong to X but not to both X and Y. The latter notation is more explicit and will be used here. Efficiency of Information Transmission by Biochemical Co-factors 207 The informational performances at the decision point are related to the reduction of uncertainty, AH, at these points, i.e. the difference in uncertainty before and after: A//i = H{C) - H{C.A) AHii = H{C.A) - H{C.A.S) AHiji = H{C) - HiC.A.S) The comparable functions for artificial situations are given by: AHi* = H{B) - H{B.A) AHii* = H{B.A) - H{B.A.S) AH*iji = H(B) - H{B.A.S) There is a fundamental difference between natural and artificial functions. The quantities H{C) as well as H{C.A) and H(C.A.S) do not depend on the experimenter; furthermore, because of the constancy of the internal environ- ment, they can be considered to be numbers approximating natural constants, subject to relatively small fluctuations. The function AHjji represents the normal informational performance achieved in the particular metabolic process considered, whose average value has been placed at 9 bits (1). The quantity H{B), on the other hand, is completely or partially controlled by the experi- menter, who regulates the availability of substances in B; H(B.A) and H{B.A.S) depend on apo-enzyme, substrate and on the experimenter. Accordingly, the AH* functions have, in general, very little interest since it is easy to make AH* vanish by offering only a single co-factor which can be used, or to give it a very high value by introducing numerous compounds which are known not to react with the apo-enzyme or substrate. A great body of data is available, however, which lends itself to an examina- tion of the AH* functions as well as H{B.A) and H{B.A.S) from the standpoint of the systems' responses to a series of compounds closely resembling the natural co-factors. The values obtained may be regarded as a sort of minimum residual uncertainty associated with the various systems, and they form the subject of this report. We shall define the uncertainty functions H{B.A) and H(B.A.S) as: H(B.A) = -i:pJog,p, H(B.A.S)=-i:p,,log,p„, where p^ and p^^ are the normalized biological activities of the compounds tested in a particular system. Thus if four compounds were tested for their ability to combine with the apo-enzyme and all were found to be equally active, their /7^ would each be 0.25 and H{B.A) would be 2. This method of calculation takes into account the fact that with equal concentrations, equal activity may not be observed and that the information is of necessity related to the concentra- tion required to produce a complex. A word is in order as to the mechanics of calculation. The basic data were derived in large part from Williams and co-workers' treatise on the B vitamins (2) ; data on thyroxine are due mainly to the work of Bruice, Kharasch and 208 Peter D. Klein WiNZLER (3). H{B) was defined as the logarithm of the total number of compounds tested, both for ability to replace the natural co-factor and for those having antimetabolite activity. H(B.A) was calculated from the com- pounds active as antimetabolites plus those having substrate activity; in the former case, the inhibition index (number of molecules of inhibitor required to overcome the action of one molecule of the true compound) was considered as the reciprocal of biological activity and suitably transformed to agree in dimen- sion with the other data. H(B.A.S) was, of course, derived from the group which showed ability to replace the natural co-factor. RESULTS AND DISCUSSION The H functions and AHi*, Ai/u* and ^Hju* are fisted for a variety of compounds in Table 1. In addition, compounds for which partial data were Table I Compound Organism N H{B) H{B.A) H(B.A.S) A/Zj* AHu* AZ/fn Biotin MO 33 5.04 3.01 2.67 2.03 0.34 2.37 Riboflavin MO 40 5.32 2.91 2.61 2.41 0.33 2.71 Riboflavin R 18 4.16 2.93 2.61 1.24 0.33 1.57 Folic acid C 11 3.46 2.74 1.97 0.72 0.77 1.49 Folic acid MO 51 5.67 3.97 2.55 1.70 1.42 3.12 Thiamine R 8 3.00 1.39 0.99 1.61 0.40 2.01 Thyroxine F 34 5.09 2.55 1.47 2.54 1.08 3.62 Thyroxine R 43 5.43 4.28 3.22 1.15 1.06 2.21 /j-Amino benzoic acid MO 72 6.17 4.90 3.27 1.27 1.63 2.90 Biotin R 4 2.00 — 0.13 — — 1.87 Unsaturated fatty acids R 10 3.32 — 1.85 — — 1.75 Pantothenic acid C 4 2.00 — 1.86 — — 0.14 Vitamin D R 10 3.32 — 2.08 — — 1.24 Pantothenic acid R 10 3.32 — 2.19 — — 1.13 Nicotinamide C 5 2.32 — 2.23 — — 0.11 Ascorbic acid G 10 3.32 — 2.33 — — 0.99 Nicotinamide D 13 3.70 — 2.66 — — 1.04 Pyridoxine R 11 3.34 — 2.85 — — 0.49 Choline R 35 5.11 — 3.12 — — 1.99 Carotene R 15 3.90 — 3.68 — — 0.22 Estrogens R 18 4.16 — 3.88 — — 0.28 Average 3.96 2.39 1.57 Key: MO: Micro-organism R: Rat C: Chick F: Frog G: Guinea pig D: Dog available are included. Fig. 3 presents some values for the first group in graphic form. It can be seen that there is a range in H{B.A) of 1.39 to 4.90 and in H(B.A.S) of 0.13 to 3.88. There is also a marked tendency for A/Zn* to be smaller than A//^*, suggesting that the greater portion of the selection process is assumed by the apo-enzyme/co-enzyme complex formation, but sight must not be lost of Efficiency of Information Transmission by Biochemical Co-factors 209 the fact that the area H(B.A)-H(B.A.S) (representing antimetabolites) is dependent upon the number of successful inhibitors of a co-factor which have been devised. Over-all, the mean reduction in uncertainty in terms of actual compounds is such that when confronted with fifteen compounds, assumed to be equally Fig. 3. Residual uncertainties associated with two co-factors. effective, the system can weed out ten of these, leaving the equivalent of five, equally active, co-factors. Comparison of this to H(C) and H{C.A.S) is of course not plausible from these figures, but clearly indicates the relative chaos of the universe B, from the standpoint of the enzyme system : the assembly of letter-perfect molecules of protein or nucleic acids would be impossible under these circumstances. Nevertheless, these figures may have some interest as the minimum limits of discrimination ability by enzyme systems. REFERENCES 1. P. F. Fenton, cited by H. Quastler: The specificity of biological functions. In: Information Theory in Biology, ed. by H. Quastler, 170-188, University of Illinois Press, Urbana (1953). 2. R. J. Williams, R. E. Eakin, E. Beerstecher Jr., and W. Shive: The Biochemistry of B Vitamins. Reinhold Publishing Co., New York (1950). 3. T. C. Bruice, N. Kharasch, and R. J. Winzler: A correlation of thyroxine-like activity and chemical structure. Arch. Biochem. Biophys. 62, 305-317 (1956). DISCUSSION Quastler: The following interpretation of the 'residual uncertainty of co-factors' may be considered : Given a particular apo-enzyme-substrate system, and the set B of all substances bi which are or might be co-factors. Let c, be the reaction rate constant of the system in the presence of the (potential) co-factor b,. Suppose that all Ci's have been determined; it seems that there are two statistics of general interest: the average size of the c's, which characterizes the reactivity of the system in general, and the dispersion among the c's, which characterizes its specificity. 210 Peter D. Klein To study the specificity independently of general reactivity, we normalize the c/s by setting their sum equal to one. We consider a function called the tolerance of the apo-enzyme-substrate system ; the tolerance function shall have the following properties : it shall be zero if only one Ci is not zero; it shall increase with the number of substances b with non-zero c's; for a given number of substances the tolerance shall be called highest if all c ,'s are equal. These postulates are satisfied by the information function : Value of tolerance = — V ^r- log2 ^?^— Tentatively, one might assign a physical meaning to the tolerance as follows : the reaction rates are replaced by (suitably normalized) probabilities of a reaction following a collision ; it is assumed that the system can exist in a number of mutually exclusive states (configurations) and that a reaction will occur if the system is in the proper configuration; then, the c^'s are proportional to the fractions of the time the system is in a configuration compatible with functioning with the substances 6,. The value of the tolerance is the uncertainty concerning the actual state of the apo-enzyme-substrate system. XVvf i f / *^ /** UJ / , , •3r >'*■ ANTIGENIC SPECIFICITtS^ ''^z v: Bernard N. Jaroslow and Henry Quastu Division of Biological and Medical Research, Argonne National Laboratory, Lemont, Illinois and Department of Biology, Brookhaven National Laboratory, Upton, New York Abstract — The production of a specific antibody involves a transfer of information, and so does the specific reaction between antibody and antigen. This paper deals with the 'vocabulary' of this communication process. An antigenic determinant is considered as a 'word' of a certain number of 'letters', subject to certain constraints. It is shown that the number of 'words', the number of 'letters', and the degree of constraint can be estimated by methodical random sampling. Experimental methods suitable for this purpose are discussed. Preliminary results are given. I. INTRODUCTION The question, 'How many different antigens are there?' is one that has not been explored up to now, but which arises naturally if information theory is applied. Information theory interprets the process of antibody formation as the transmission of information from the antigen to the antibody-forming mechanism, with the information then utilized and again transmitted when the antibody reacts selectively with the appropriate antigen. It is then natural to ask how much information is transmitted from antigen to antibody and vice versa. More explicitly, one will ask certain questions about the kind of information traffic between antigen and antibody — the 'vocabulary' in which this information traffic is coded, the 'alphabet' that is used to make up the words of the vocabulary. Now, information theory is not concerned with specific features of 'alphabet' and 'vocabulary' but with general properties of both, such as their sizes. The problem, 'how large is the vocabulary of information transmission between antigen and antibody,' is closely related to the question posed above. A preliminary estimate by one of us (1) has led to a rough estimate of some 125 to 500 different protein antigenic determinants, and a smaller number of different carbohydrate determinants. No attempt was made, at that time, to estimate the number of antigens of other chemical constitutions. Although these figures appear very small in the light of the specificity of immunity to the multitude of infectious agents, the antigen com- plexes of the organisms represent an array of many different determinants and their over-all specificity can be much larger than that of a single antigen. The present investigation is an attempt to measure antigenic specificity. The general plan of the experiment is based on information theory; the specific methods are based on agar diffusion precipitin tests developed by Oudin (2). * Work performed under the auspices of the U.S. Atomic Energy Commission. 211 212 Bernard N. Jaroslow and Henry Quastler II. NOMENCLATURE AND MODEL An antigen, GN, consists of a specific detenninant G and a carrier moiety A'^ which is a macromolecule (MW > 10,000), i.e. protein, hpo-protein, glyco- protein, etc. G may consist of three or four amino acid residues, a mono- or disaccharide, an aliphatic chain, etc., with its specificity dependent upon order, size, polarity, or optical configuration of the residues. There may be a number of antigenic determinants, of the same or different specificities, on the surface of N, so that one molecule of antigen may combine with several molecules of antibody, e.g. 5 for ovalbumin to over 200 for hemocyanins (3). Combining capacity as well as antigenicity (ability to induce antibody formation) is usually proportional to the molecular weight of A'^. An antibody, AB, consists of a specific combining site A and a carrier moiety B which is always a globulin, usually y-globulin. The combining site is believed to be a chemical and/or spatial configuration which combines with the specific antigenic determinant through hydrogen bonds. The number of combining sites per antibody molecule is thought to be two or one. The A-G reaction may be manifested in a number of ways: precipitation of a soluble antigen, agglutination of a particulate or cellular antigen, or lysis of a cellular antigen in the presence of complement. We consider a heterophile reaction as one in which the reaction between C2 and y^i is indistinguishable from the homologous reaction of G^ and A-^ although G2 and G^ come from different sources. By cross reaction we mean the phenomenon wherein G2 reacts with A^ but the strength of reaction is less than that of the homologous one {G^ and A^. We can describe an antigenic determinant, G, as a 'word' of k 'letters'. By letter we mean antigenically active residues such as amino acids, mono- saccharides, etc. Let r be the size of the alphabet, i.e. the number of available letters; then H (letter) = a • logg r. a is a constant which ranges from zero to one. Its upper limit occurs if all 'letters' occur with equal probabihties, and no two letters can ever have equiva- lent effects. The average information content of a worfi? averaging k letters is given by: H(G) = ^- k- H (letter). /9 is a constant which ranges from zero to one. Its upper limit occurs when all letter combinations are equally probable, i.e. if there is no 'intersymbol influence'. The lower limit would obtain if there existed only one antigenically active combination of letters. To fix the ideas on the measuring of /-, H (letter), k and H(G), we give the corresponding values for printed English: r = 26 k {=ii4.5 log2 r = 4.7 ^ f^ 0.6 a ^ 0.87 H{G) ^ 10 //(letter) = 4.1 Antigenic specificity 213 III. EXPERIMENTAL TESTS 1. Occurrence of the Heterophile Reaction The incidence of heterophile reactions will depend on the number and relative frequencies of the various antigenic determinants. As nothing is known so far about relative frequencies, we assume them to be equal; this will yield a lower bound of the number of different G's. Under the assumption of equi- probability, we use the following argument (4): Let Q and C_, be different and (as far as known) unrelated antigen complexes; let m^ and nVj denote the number of antigens in the complexes which can be differentiated and demonstrated, by a given technique, by reaction with the specific antisera S^ and Sj\ let hij be the number of heterophile reactions observed ; let A^ be the total number of different antigenic determinants which this technique will differentiate. Then, the maximum likelihood estimate N of N is given by: "a Assuming ft to be one, we have: H(G) ^ logo N and This is a preliminary test of r'^'^. 2. Classification of Cross Reactions The strength of the cross reaction presumably depends on the number of letters in common, and on the nature of these letters. We assume as a working hypothesis that the former factor is the leading one. Then, if we grade the strengths of many cross reactions, we expect to find a distribution into clearly separated groups such as strong, less strong, weak, . . . etc., cross reactions. We suspect that the strong cross reactions are those in which the G-pair has (^ — 1) letters in common, the next class those with (k — 2) letters, and the weakest observable reactions those with one common letter. Then, the number of distinguishable classes of strengths of cross reactions should be (k — 1). This may develop into a test of k. 3. Ratio of Incidence of Heterophiles to Incidence of Strong Cross Reactions By our hypotheses, the probability of occurrence of heterophiles is the probability of having all letters in common; now, for k letters, assuming p to be one, the number A^ of different words is: TV = r'^'^. Then, probability of a given word = {\jrY^' and, probability of a heterophile = (1//-) 15 ■ % \ • » • (between divisions) \\. /' (membrane barrier) B. Cytoplasmic Organelle Systems TO ACT (represents collisions of microsomes-mitochondria) TO BEAT TOPOLARIZEJi i • # TO SCAN TO ANCHOR (represents linked TO LINK kj" ^ Y ^TOLI^Kkj-r PeHide-kineties) TO PACK ^\ 0 TO HOLD OPEN; TO LINK kj,., TOLINKkj(#^ # TO SWEEP (represents kinety) i TO FORMIC (represents gullet system) Fig. 3. Partial primordial graphs for organelles in Paramecium. A. Organelles engaged in communications between nucleus and cytoplasm and in construction of complexes and systems of organelles. B. Organelles engaged in systems for feeding, for locomotion and for structural integration. up to 1 // in diameter during fission (15).) The first function of the primary organelle is to contain the message introduced presumably by the chromosomal genes at time of nucleolar synthesis. Its next function is to move and to transport; the function shown in the figure is to move 'out of the nucleus; this is followed by to act (collide, fuse, develop), an accomplishment in which the message and its vehicle are also partial power supply and building stone to make cilium, trichocyst, mitochondrion, or whatever is dictated by uncertainty loss. The decision tree (Fig. 2) is itself a partial primordial graph representing the diversity of these acts. The cytoplasmic organelle system (Fig. 3b) operates through similarly gross functions. The function of a pellicle unit is to beat (row, propel), to pack (to fit as a block in the pellicle wall), and to link similar units longitudinally (k^) and latitudinally {k,)\ a column of such units (one kinety) is aligned with directional reference to each of its ciliary bases (kinetosome) and basal fibrils (kinetodesma); each kinetodesmal fibril lies to the right of its kinetosome; and each is over- lapped by and in turn overlaps two or three of the fibrils antero-posteriorly •* - ^, . . - 4 ■ vir •:, / j4 .. .^y^,.:^;^-ar- Plate I The appearance of organelles in Paranicciuin under phase conlrasl and electron microscopy. (In A and B the line represents 10 /' ; in C. D. and E it represents 1 //.) A. Food-intake system, compression-dissected from an unfi.xed cell. From left to right, the non-ciliated 'granules' of the ribbed wall complex followed by four columns of the ciliated quadrulus complex and eight columns of the ciliated peniculus complex (26). Phase. B. Macronucleus anlage in an exconjugant during extrusion of young nucleoli: net-like figure in the center is fusion product of old nucleoli (15). Phase. C and D. Electron micrographs of a similar stage. The smaller dark bodies are chromatic elements of the nuclear matrix: the larger bodies at the left are young nucleoli, and at the right in the cytoplasm are mitochondria (15). E. Electron micrograph of a single packing unit of the hexagonal complex of the pellicle system (4iS). Note the cross-sectioned kinetodesmal fibrils in the bays of cytoplasm at each side of thecilium base; at the left, a portion of a trichocyst with its golf-lee-like head: at the right a tubular mitochondrion (S. 9). None of these photographs has been previously published. Each is a part of the work cited in parentheses, and done in collaboration with Drs E. L. Powers and L. E. Roth of Argonnc National Laboratory. to face p. 224 Information Content and Biotopology of the Cell in Terms of Cell Organelles 225 along the kinety; these units act to polarize; the array of kinetics (or entire pelhcle system) functions to envelop the endoplasm of the whole animal. It also functions to anchor in place other systems such as the food intake or gullet system. The anchorage confers a new level of asymmetry, resulting in the swimming function to scan or spiral; the function of the peniculus and quadrulus is to sweep food particles down the intake tube; their terminal cilia act to form food vacuoles (FV); the cilia-free ribbed wall functions to confer rigidity and to hold open the tube, which lies within the endoplasm; the gullet system also functions to envelop (k,',-) the endoplasm. This graph may be read in the following way: effective beating of a kinety cilium requires polarization; both functions require positioning upon the pellicle by latitudinal linking to adjacent kinetics (A:/s) ; the previous functions require packing (a continuously surfaced pellicle). Longitudinal linkage {k^) is required to prevent the dispersion of the surface blocks from within a kinety. Each kinety organelle may perform every function, but locally any function may be by-passed. The function to scan requires anchorage of the gullet organelle-complexes in position on the animal; the gullet's general function is to feed. An effective food vacuole requires that food be swept into it by the cilia of the peniculus and quadrulus ; these functions require that the gullet be held open by the gullet tube-wall, which requires anchorage to the pellicle. The general function of envelopment requires the linkage of all ^/s and A:/s. (A nearly unique quality of the ciliated protozoa, which, however, should not be entirely ignored in the transformations to higher forms, is the presence of what might be called 'linkage groupings' in the cytoplasm: organelle patterns, far more complex than any known in the metazoa, appear to be as much dependent upon the previous existence of a related pattern in the cytoplasm as upon any nuclear genes (49); even Stentor, with its remarkable capacity to regenerate 'kinetics' (actually pig- mented stripes), rebuilds its mouth organelles only when a particular juncture of maximum anisotropy in the stripe pattern is available (35).) Organelle- functions are therefore given not in the terminology of the molecular level (whose necessary though not strictly pertinent relations are partially represented for the whole cell topologically in such graphs as those of the glycolytic and citric acid cycles (50)) but in the correspondingly appropriate terms of the gross operations performed. According to this concept, the cell is entirely describable in minute detail of anatomical pattern without reference to either power or fuel. It does not matter whether the oar-like cilia are tugged by galley-slaves, gasoline engines or a creatine-phosphate-ATP system. However, the universal usage by cells of such engines places some restriction at the systems-coupling level, and probably represents a nearly unique solution of the bioenergetics problem. If the model is correct, the most complex patterns are entirely derivable by just such remarkably simple interactions as those first explicitly delineated by D'Arcy Thompson (12). In summary, at the organelle level fundamental topological sets are recog- nized of two classes: those that are periodically disjoined (intranuclear from extranuclear organelles), and those that are continuously joined at non-empty intersections (cytoplasmic organelle-systems). Periodic coupling processes (such as during mitosis and nuclear membrane disappearance) occur to form non- empty intersections at all disjunctions of the first class. Below this dimensional 226 Charles F. Ehret level and within the spheres of intracellular and extracellular molecular inter- action, each set of the higher two classes is continuously capable of communica- tion with every other set by means of diffusion and convection transport phenomena. V. CONCLUSION In the introduction to this conference Bigelow suggested that in biological systems with long time-constants, for a message to be useful at the receiving point, 'all messages must be enormously complex groups of messages rather than simple ones' (51). This expression is clearly related to the limited span proposition (52), that is, that span of diversity is limited by difficulties of internal control. Therefore it is not too surprising to find that epigenetic control systems of the cell (whose internal difficulties are in the form of long time functions and thermodynamic vulnerability) solve these difficulties by the method of 'chunkmg' complex groups of messages into structurally and func- tionally unitized subsystems. That the subsystems of primary organelles appear to be phylogenetically ubiquitous might also have been predicted from the principles of biochemical evolution. But that they are structurally so alike is indeed a striking fact; although this is not to say that we should now expect the cilium of whale bronchus, the axial fiber of fern sperm, the connecting fibril of toad retina, the sweeping cilium of paramecium peniculus or the mantle cilium of a mollusc to be exactly alike. We know that such organelles are capable of antigenic distinctions even amongst the various stocks within a species (53); indeed, the mechanism of such distinctions constitutes a most crucial problem of molecular biology. That the functional and structural diagrams of an organism in temis of its organelles are topologically homeo- morphic is consistent with parallel relations at other levels; in its functions between molecular and cellular levels of organization, the cell organelle fills the last gap in a complex hierarchy of unitized subsystems that characterize the organism from the atomic to the social level. The method of integrating these hierarchies and of extracting quantity of infoiTnation from any organism that employs such mechanisms remains to be accomplished. Acknowledgment — I wish to thank my many colleagues, particularly Howard DucofF, E. Lawrence Powers, and Robert Schweisthal of Argonne National Laboratory for their interest and cooperation, and Henry Quastler of Brookhaven National Laboratory for his very helpful suggestions. REFERENCES* 1. S. M. Dancoff and H. Quastler: The information content and error rate of living things. In: Information Theory in Biology, ed. by H. Quastler, 263-273, University of Illinois Press, Urbana (1953). 2. L. Augenstine: Protein structure and information content. This volume. 3. L. Monne: Functioning of the cytoplasm. Advances in Enzynwlogy 8, 1-69 (1948). 4. D. E. Green: Organized enzyme systems. J. Cell. Camp. Physiol. 39, Suppl. 2, 75-111 (1952). 5. W. T. Astbury: Introductory review: fibrous proteins and their biological significance. Symp. Soc. Exp. Biol. 9, 1-9 (1955). Information Content and Biotopology of the Cell in Terms of Cell Organelles 227 6. E. De Robertis: Electron microscope observations on the submicroscopic organization of the retinal rods. /. Biophys. Biochem. Cytol. 2, 319-330 (1956). 7. G. E. Palade: An electron microscope study of the mitochondrial structure. J. Histochem. Cytochem. 1, 188-211 (1953). 8. E. L. Powers, C. F. Ehret, and L. E. Roth: Mitochondrial structure in Paramecium as revealed by electron microscopy. Biol. Bull. 108, 182-195 (1955). 9. E. L. Powers, C. F. Ehret, L. E. Roth, and O. T. Minick: The internal organization of mitochondria. /. Biophys. Biochem. Cytol., Suppl. 2, 341-346 (1956). 10. J. Rhodin and T. Dalhamn: Electron microscopy of the tracheal ciliated mucosa in rat. Z. Zellforsch. 44, 345^12 (1956). 11. K. E. Wohlfarth-Bottermann : Protistenstudien. VII Die feinstruktur der mitochondrien von Paramecium caudatum. Z. Naturf. 11, 578-581 (1956). 12. W. D'Arcy Thompson: Growth and Form, Cambridge University Press, Cambridge, England (1952). 13. T. N. Tahmisian, E. L. Powers, and R. L. Devine: Light and electron microscope studies of morphological changes of mitochondria during spermatogenesis in the grass- hopper. /. Biophys. Biochem. Cytol., Suppl. 2, 325-330 (1956). 14. H. W. Beams, T. N. Tahmisian, R. L. Devine, and E. Anderson: Electron microscope studies on the dictyosomes and acroblasts in the male germ cells of the cricket. /. Biophys. Biochem. Cytol., Suppl. 2, 123-128 (1956). 15. C. F. Ehret and E. L. Powers: Macronuclear and nucleolar development in Paramecium bursaria. Exp. Cell Res. 9, 241-257 (1955). 16. A. J. Hodge: A possible mechanism for the morphogenesis of lamellar systems in plant cells. J. Biophys. Biochem. Cytol. 2, 597-608 (1956). 17. C. Levinthal: The mechanism of DNA replication and genetic recombination in phage. Proc. Nat. Acad. Sci., Wash. 42, 394-404 (1956). 18. J. H. Taylor, P. S. Woods, and W. L. Hughes: The organization and duplication of chromosomes as revealed by autoradiographic studies using tritium-labelled thymidine. Proc. Nat. Acad. Sci., Wash. 43, 122-128 (1957). 19. A. F. Huettner: Continuity of the centrioles in Drosophila melanogaster. Z. Zellforsch. 19, 119-135(1933). 20. E. Chatton and J. Seguela: La continuite genetique des formations ciliaires chez les cilies hypotriches. Le cinetome et I'argyrome au cours de la division. Bull. Biol, 74. 349^142 (1940). 21. Akira Yuasa: Critical observations on the origin of the blepharoplast and centrosome in plants. Jap. J. Bot. 10, 259-284 (1939). 22. R. Lepper: The plant centrosome and the centrosome-blepharoplast homology. Bot. Rev. 22, 375-417 (1956). 23. H. W. Beams, T. N. Tahmisian, and R. L. Devine: Electron microscope studies on the cells of the malpighian tubules of the grasshopper. J. Biophys. Biochem. Cytol. 1, 197-202 (1955). 24. D. W. Fawcett and K. R. Porter: A study of the fine structure of ciliated epithelia. J. Morph. 94, 221-281 (1954). 25. M. L. Watson: The nuclear envelope. /. Biophys. Biochem. Cytol. 1, 257-270 (1955). 26. C. F. Ehret and E. L. Powers: The organization of gullet organelles in Paramecium bursaria. J. Protozoal. 4, 55-59 (1957). 27. L. E. Roth: Aspects of ciliary fine structure in Euplotes patella. J. Biophys. Biochem. Cytol, Suppl. 2, 235-240 (1956). 28. C. B. Metz, D. R. Pitelka, and J. A. Westfall: The fibrillar systems of ciliates as revealed by the electron microscope. I — Paramecium. Biol. Bull. 104, 408-425 (1953). 29. A. Lwoff: Problems of Morphogenesis in Ciliates, J. Wiley and Sons, New York (1950). 228 Charles F. Ehret 30. J. F. Danielli: Cytoplasmic inheritance. Nature, Lond. 178, 214-215 (1956). 31. F. Wassermann: The intercellular components of connective tissue: origin, structure, and interrelationship of fibers and ground substance. Ergebn. Anat. Entwgesch. 35, 240-333 (1956). 32. J. Gross: The behavior of collagen units as a model in morphogenesis. /. Biophys. Biochem. CytoL, Siippl. 2, 261-274 (1956). 33. M. R. Cohen and E. Nagel: An Introduction to Logic and Scientific Method, Harcourt- Brace, New York (1934). 34. E. L. Powers, C. F. Ehret, and L. E. Roth: An hypothesis of mitochondrial origin in Paramecium. Rep. Argonne National Laboratory, ANL-52A1, 55-60 (1954). 35. V. Tartar: Pattern and substance in Stentor. In: Cellular Mechanisms in Differentiation and Growth, ed. by D. Rudnick, 73-100 Princeton University Press (1956). 36. L. I. Rebhun: Electron microscopy of basophilic structures of some invertebrate oocytes. 1. Periodic lamellae and the nuclear envelope. /. Biophys. Biochem. Cytol. 2, 93-104 (1956). 37. H. Gay: Nucleo-cytoplasmic relations in salivary gland cells of Drosophila. Proc. Nat. Acad. Sci., Wash. 41, 370-375 (1955). 38. H. Swift: The fine structure of annulate lamellae. J. Biophys. Biochem. Cytol., Suppl. 2, 415-418 (1956). 39. G. D. Pappas: The fine structure of the nuclear envelope of amoeba proteus. /. Biophys. Biochem. Cytol, Suppl. 2, 431-434 (1956). 40. A. Garen and N. D. Zinder: Radiological evidence for partial genetic homology between bacteriophage and host bacteria. Virology 1, 347-376 (1955). 41. A. Lwoff: Control and interrelations of metabolic and viral diseases of bacteria. Harvey Lectures {Ser. L), 1954-1955, 92-111, Academic Press, New York (1956). 42. J. H. Northrop and J. S. Murphy: Appearance of new phage types and new lysogenic strains after adaptation of lysogenic B. megatherium to ammonium sulfate culture medium. /. Gen. Physiol. 39, 607-624 (1956). 43. G. N. Calkins: The life history of Cytoryctes variolae, Guarnieri. /. Med. Res. 11, 136-172 (1904). 44. W. R. Duryee : Seminar on transmission studies on renal adenocarcinoma of the frog. /. Franklin Inst. 261, 377-379 (1956). 45. T. H. Dunnebacke: Cytopathic changes associated with poliomyelitis infections in human amnion cells. Virology 2, 811-819 (1956). 46. C. Morgan, H. Calderon, H. Rose, and D. Moore: Structure and development of viruses observed in the electron microscope. IV — Viruses of the Rl-APC group. /. Biophys Biochem. Cytol. 2, 351-360 (1956). 47. N. Rashevsky: Life, information theory and topology. Bull. Math. Biophys. 17,229-235 (1955). 48. C. F. Ehret and E. L. Powers: The systems and complexes of primary organelles in Paramecium. J. Protozoal., 3 Suppl. 3, 5 (1956). 49. P. B. Weisz: Morphogenesis in protozoa. Quart. Rev. Biol. 29, 207-229 (1954). 50. N. Rashevsky: Contributions to topological biology: some considerations on the primordial graph and on some possible transformations. Bull. Math. Biophys. 18, 113-128(1956). 51. J. N. Bigelow: Unpublished. 52. H. Quastler: Unitization in biological systems. Essays on Biological Unitization, Report No. R-52, Control Systems Laboratory, University of Illinois, Urbana (1953). 53. J. R. Preer: Gel diffusion studies on the serotypes of Paramecium. Jordan Hall Dedi- cation Program, Indiana University (1956). 54. E. Borysko and F. B. Bang : Electron microscope observations on nucleolar structure. J. Appl. Phys. 23, 163 (1952). 55. H. J. Weinstein: An electron microscope study of cardiac muscle. Exp. Cell. Res. 7, 130-146 (1954). Information Content and Biotopology of the Cell in Terms of Cell Organelles 229 56. A. Moscona: Cytoplasmic granules in myogenic cells. Exp. Cell Res. 9, 377-380 (1955). 57. G. E. Palade: The endoplasmic reticulum. J. Biophys. Biochem. Cytol., Suppl.2, 85-98 (1956). 58. A.J.Hodge: The fine structure of striated muscle. J. Biophys. Biochem. Cytol., Siippl. 2, 131-142(1956). 59. J. J. Wolken: A molecular morphology of Eitglena gracilis. J. Protozool. 3, 211-221 (1956). 60. B. A. Afzelius: The fine structure of the sea urchin spermatozoa as revealed by the electron microscope. Z. Zellforsch. 42, 134-148 (1955). 61. B. P. Potts: Electron microscope observations on trichocysts. Biochiin. Biophys. Acta 16, 464^70 (1955). * The last eight references are not specifically discussed in the text, but are referred to as examples in the Decision-Tree diagram, Fig. 2. Many of the cytological references cited concur with the data rather than with the interpretations of the authors. 16 QUANTIFICATION OF PERFORMANCE IN A LOGICAL TASK WITH UNCERTAINTY A. Rapoport Mental Health Research Institute, University of Michigan Abstract — Tasks to which information theory has been apphed characteristically do not involve 'reasoning', i.e. the drawing of inferences. The present paper explores the possibility of applying information theory to measuring performance of logical tasks. We note at once that any task in which a necessary conclusion must be reached from given information has formally speaking no information content. From the information-theoretical point of view, therefore, no information is gained in the process of solving a purely mathematical or logical problem, no matter how 'complex'. There are problems, however, in which in addition to the making of inferences, information must be obtained in the process of solution. Success of solution can be measured by the rate of obtaining such information and by the degree of completeness with which it is utilized. Assuming complete utilization at each step, the efficiency of solution depends on the efficiency with which information is obtained. A classical example is the coin-weighing problem in which a deviant coin and the direction of its deviation must be determined in the fewest possible weighings. Information theory provides not only the minimum number of weighings for such a problem but also a method for constructing the best 'strategy'. In the present paper a particular logical task with uncertainty is discussed from the infor- mation-theoretical point of view. It is shown that the construction of an information-getting strategy depends very strongly on the instructions given the subjects and on the inferences which the subjects make from the instructions. Thus the practical problem of quantifying the performance of a logical task carries within it certain ambiguities which must be resolved if information theory is to be of use in psychological tests based on such tasks. Information theory is mainly concerned with a quantity called the amount of uncertainty associated with a situation in which choices or guesses are made. This uncertainty can be viewed as a measure of ignorance. For example, we are the more ignorant of the value about to be assumed by a random variable with a discrete domain, the more values it can assume and the more nearly equi-probable these values are. Defining every situation of ignorance is a set of postulates with a subjective flavor. Somebody is ignorant. At least this is the case in real situations involving subjects whose state of ignorance is to be inferred. It may be argued from certain philosophical points of view that this intrusion of subjective concepts is unsatis- factory, and attempts should be made to circumvent them or to eradicate them altogether. I don't want to take sides on this question, but only to point to some of its manifestations by way of indicating its persistence. The question has been raised in connection with the foundations of probabihty theory. There the attempts to circumvent the subjective element have given rise to the so-called 'objectivist school', which sought to define probabilities of events 'objectively' in terms of the relative frequencies of the events. Opposition to 230 Quantification of Performance in a Logical Task With Uncertainty 231 'subjectivist' notions are also, I think, at the root of many philosophical objec- tions to quantum theories. The question of where to draw the line between subjective and objective frames of reference also arises in connection with attempts to link information theory with thermodynamics so as to make it useful in theoretical molecular biology. Whatever the merits of the attempts may be, this area of investigation does bring out rather pointedly the necessity of examining the possibly 'subjec- tive' postulates underlying the description of the situations studied, indeed, the philosophical question 'What is a subjective postulate?' comes to the forefront in the region where thermodynamics, quantum theory and information theory, meet — a triple point. There is an area of possible application of information theory, however, where clearly subjective postulates are not only unavoidable but central. This is the area of psychology. Psychology is a study of behavior. Since psychology has gradually outgrown the austere positivistic restrictions of strict behaviorism, it has become respectable again to include into psychological theory considera- tions based on how the situation looks from inside the subject. To be sure, these matters must somehow be inferred from overt behavior, but once they are inferred there is no reason why these 'subjective variables', for example, sub- jective probabilities, utility functions, and so on, cannot enter as parameters in a theory. Indeed, if these parameters are determinable and stable, they serve to 'objectify' the subjective and thus contribute to the success of psychology as a science. Information theory was applied from its very inception to psychological investigations. These applications have often been criticized. The grounds for criticism have been many, but a recurrent theme has been the failure of many psychologists to realize that information theory is worthless without an under- lying set of postulates for each situation. Just as the application of probability theory to any situation necessitates the determination of a 'sample space', that is, a set of elementary events with a priori assignment of probabilities or a probability distribution function, so is the case with information theory. Yet it was shown by de Finetti (1), Savage (2), and others that a rigorous theory of probability could be constructed backwards. That is to say, beginning with certain preferences of individuals for certain outcomes as reflected in their choices of actions under uncertainty, a set of subjective probabilities of events could be inferred, provided certain 'rationality criteria' of behavior were satisfied by the individuals. The question to what degree such rationality criteria are in actuahty satisfied is another question which has led to many interesting investigations in their own right; so it is not entirely an unfortunate one. It must in any case be admitted that the 'subjective probability' of an event can in principle be defined, and thus statements such as, 'The Democrats will with probability 0.6 win in I960,' are not wholly devoid of operational sense, pro- vided the expressed 'subjective' probability is inferrable by explicit rules from observed behavior and enjoys a certain stability. Such assertions have no sense in the conceptual framework of the objectivist school, since the election of 1960 is a unique event whose 'probability' cannot be deduced from a fre- quency of occurrence. The operational definition of subjective probability introduces probability 232 A. Rapoport theory into psychology in a significant way, not as a mere appendage to statistics. I think the situation is similar in the case of information theory. Instead of lamenting the ambiguity of the 'universe of discourse' in psychological situations, which stands in the way of a straight-forward appHcation of infonnation theory, we may well seek to infer the universe of discourse as it looks from inside the subject. This is, indeed, a central task of the psychologist, and it is improper for him to shun it. It remains true, however (and it is just as it should have been), that the early psychological experiments based on information-theoretical considerations were constructed in such a way as to eliminate idiosyncratic subjective probabili- ties. In memory tasks one starts with some set of thoroughly randomized and presumably equalized stimuli. One introduces redundancies in terms of actual biases of occurrence-frequencies objectively determined. The same techniques prevail in experiments in which the capacity of the individual as a channel is measured. These are all attempts to translate into experimental psychology situations occurring in communication engineering. The human being is studied as a piece of communication apparatus. I beheve this strategy to be entirely correct as far as it has gone. I am sure, however, that its limitations are apparent most of all to the investigators who pursue it. Somewhere along the line a transition must be made which will allow the application of information theory to psychology as distinct from psychophysics. In other words, the perceptual world of the subject must eventually become a focus of interest. There is no reason why information theory should not become a useful tool in such investiga- tions. Characteristically, the tasks just described (such as rote learning, multiple choice responses, and so on) do not involve the deductive process. Indeed, from the formal information-theoretical point of view, the results of infomiation theory are not applicable to a deductive process, because there is no 'uncertainty' in such processes. The solution of a mathematical problem, no matter how complex, yields no information from the information-theoretical point of view. From either the common-sense or the psychological point of view such a con- clusion seems bizarre. The information-theorist can, of course, argue that his technical definition of information departs from common-sense and psycho- logical notions of what constitutes infonnation, and he is technically correct. Yet it might be instructive to try to bring the two concepts of 'information' into closer agreement. Formally speaking, no information is gained in the solution of a purely mathematical or logical problem, because the solution is implicitly contained in the already known conditions. But the solution is not initially known to the subject. Is there a way to measure the extent of his ignorance ? There might be, if we are willing to abandon the omniscient position from which the solution is seen as a necessary inference, hence of zero uncertainty, and enter the per- ceptive of the cognitive field of the subject, to whom only a range of possible solutions and, perhaps, associated subjective probabilities present themselves. But how does one get into this perceptive field? Obviously by observing the subject's behavior. But how does one make inferences from the observations to what that perceptive field may be ? It would be gratifying to be able to say that for every configuration of the cognitive field, there is a specific behavior pattern, Quantification of Performance in a Logical Task With Uncertainty 233 but unfortunately this is not the case, as will be seen in the following example. Suppose we present the subject with the famous coin-weighing problem. One of twelve coins is of odd weight. It is required to detennine the coin and whether it is lighter or heavier than the rest in a minimum number of weighings on a balance using only the coins for weights. Information theory not only reveals that three weighings are necessary and sufficient but also indicates the strategy. Obviously there are loga 24 = 4.59 bits of information (uncertainty) in the problem. A weighing can yield a maximum of logo 3 = 1.59 bits. Therefore, at least three weighings are necessary and may be sufficient. Further analysis shows that the first weighing can yield the full 1.59 bits and that only if four coins are weighed against four. The second weighing must involve six coins chosen in such a way that the three outcomes have probabihties 3/8, 3/8, 2/8, which yields 1.55 bits. The third weighing will therefore involve three coins, that is, 1.59 bits, in three fourths of the cases and two coins or 1 bit in one fourth of the cases, i.e. an average of 1.45 bits. The total information is 4.59, exactly equal to the initial uncertainty. Now the uncertainty in the problem as it is presented is clearly perceived. At least it is easy to recognize that there are initially twenty-four possibilities. It takes some effort to determine the remaining uncertainty after each weighing, but it is none too difficult to do so. We may therefore suppose that in most instances the 'uncertainty' of the problem is perceived by a fairly intelligent subject correctly, that is, in accordance with the 'objective' assignment of uncertainty. However, it is by no means true that the majority of subjects proceed to the solution in the optimal way. That is, they cannot deduce the 'correct' strategy, even when they perceived the 'actual' amount of uncertainty in the problem. It appears, therefore, that it is too much to expect to be able to deduce the subject's personal evaluation of uncertainty from his strategy in the solution of a problem in which both the deductive process and resolution of actual uncertainty must operate. However, this circumstance only reveals the situa- tions to be more 'psychological' than they appear in the light of the personal evaluation of uncertainty. Not only is this evaluation personal but also the choice of strategy is, and the latter is by no means always optimal relative to the uncertainty perceived. We are reminded of a similar difficulty in the psychology of decisions in which subjective estimates of probabilities and subjective utility functions are intimately intertwined. As pointed out, ours is a similar problem. Assuming that the solution of a logical task with uncertainty will be determined by two 'subjective' characteris- tics, namely, (a) the amount of uncertainty perceived by the subject at each step, and (b) his preference of strategy for a given amount of perceived un- certainty, then our problem is to determine these subjective characteristics in the course of a solution of a problem. It should be mentioned that some obvious techniques for detennining subjective uncertainty are in most cases unusable. If, for example, the solving process is interrupted to ask the subject what he does or does not know, the subject may through these questions become aware of relations he had not been aware of or he may doubt some assumptions he had been making correctly but with insufficient justification. 234 A. Rapoport I will now describe a task which has been adapted to an analysis of the problem-solving process in such a situation.* The subject is faced with a board on which nine numbered light bulbs are arranged in a circle, at the center of which is a tenth bulb. Each of the peri- pheral bulbs may be lit by an adjacent button. Moreover, relays are so arranged inside the apparatus that lighting of certain lights may result in the lighting of other lights following a constant 'synaptic delay' of three seconds. A 'prob- lem' is a programming in the apparatus so that certain causal relations are established among the lights. These causal relations are only partially represented by arrows on a chart attached to the mounting board. Examples are given in Figs. 1 and 2. Fig. 1. Problem 2 on PSI Fig. 2. Problem 3 on PSI The point of the problem is that the meanings of the arrows on the chart are ambiguous. An arrow from A to B may mean that A is necessary to light B or sufficient, or both, or that A inhibits B. The subject's task is to obtain sufficient information about these relations, by pushing any button he chooses, to be able to cause the center bulb to be lit by manipulating buttons 4, 5, and 6 only. We will refer to these as the circled buttons. There is a unique solution to each problem, consisting of a certain sequence of pushes of the circled buttons or of their combinations. For example, the solution to problem 2 (Fig. 1) is the pushing of buttons 4, 6, 5, 6 in the successive time periods. The solution to Problem 3 (Fig. 2) is 5, 0, 45, 6, 45. There are a number of 'rational' approaches to the problem. Let us begin by making a chart of the connections indicated by the arrows. Figs. 1 and 2 are formally equivalent (as linear graphs) to Figs. 3 and 4. However, many characteristics of the problems are visually displayed in Figs. 3 and 4, which immediately suggest various lines of attack. These charts display what the * The apparatus to be described, 'PSI', based on the isomorphism ofcertain networks of relays and the calculus of propositions (previously discovered by Shannon (4) and by McCulloch and Pitts (5)) was developed in Chicago by R. John, J. G. Miller, S. Molnar and H. J. A. RiMOLDi. The adaptation of the instrument to experiments of the type described is largely due to John (3), who has listed a great number of performance parameters to be observed in the problem solving process. Of these the so-called 'inferential lag' (defined below) seems to me of particular importance. John's terminology and definitions diff"er somewhat from those in this paper, but the basic ideas (as yet unpublished) have been the point of departure for the present analysis. Quantification of Performance in a Logical Task With Uncertainty 235 subject does not know. For example, in Fig. 3, the convergence of arrows leading from lights I, 2, and 6 upon C (the center light) induces the following questions: Which combinations of 1, 2, and 6 are necessary or sufficient or both to light C? Is one or more of them an inhibitor? If so, which one? Similar ambiguities are apparent at the other three junctures, that is at lights Fig. 3. Problem 2 displayed in time sequence 1, 2, and 9. A single arrow leading to a light presents no ambiguity. This is in consequence of the condition explained to the subject that each arrow 'means' something, hence a single arrow can mean only 'necessary and sufficient', otherwise its presence or absence would make no difference. The subject can now ask specific questions. He can ask, for example, a question about each converging juncture. The question can pertain to the d)— ^c Fig. 4. Problem 3 displayed in time sequence meaning of the arrows or to the combinations necessary or sufficient to light the bulb on which the arrows of the juncture converge. In the first case, he will be labeling arrows, in the second case the lights. Or he may proceed in a different way. Noting that all the possibilities of solution are displayed by the circled buttons in their proper time sequence, he may ask how each button is involved in the solution when its turn comes, by being pushed or not by being pushed. Note that each of these perceptions of the problem implies a different 'information content'. According to one scheme, one seeks a 'yes' or 'no' answer to every non-null combination at a juncture. There are sixteen such combinations in all in both problems, hence sixteen bits of uncertainty if the 'yes's' and 'no's' are assumed independent. They are not independent, but this interdependence can be arrived at a priori only by deduction, wliich may or may not be made. Thus the uncertainty of the situation depends on the 236 A. Rapoport State of mind of the subject. According to another scheme (labeHng arrows), one can assign the values to the arrows in eighteen different ways at the triple juncture! and in four different ways at each of the three double junctures. Hence there are 18 x 64 different ways. This gives a little over ten bits of uncertainty. According to the last scheme, one has to decide whether to push or not to push each of the circled buttons in the time period when they appear on the chart. Here Problem 2 seems to have four bits of uncertainty and Problem 3 seems to have six bits. Clearly, the amounts of uncertainty associated with each scheme are different, but so are the yields of each trial, because one counts the yield in different kinds of statements, which have different a priori probabili- ties of being true. One can push the analysis still further and thus reduce the information content of each problem by utiHzing the rule that each problem has a unique solution. In this analysis the 'sample space' v/ould be all possible problems having unique solutions involving the circled buttons at the proper time periods. Several of such problems would 'map' on each solution, and since the number of problems mapping on each solution are not equal, neither are the probabihties of the respective solutions. The value 4 bits for Problem 2 is a consequence of the equi-probability of all sixteen solutions (strictly speaking fifteen, barring the null solution where no button is pushed). If the solutions are not equi- probable the infonnation content is correspondingly reduced. This calcu'ation is extremely tedious and has not been carried out. It is mentioned only to stress the general idea that the information content of the PSI problems depends significantly on the 'sample space' according to wliich probabilities are assigned. This sample space is presumably chosen (perhaps unconsciously) by the subject; hence the amount of uncertainty in the problem is a 'subjective' quantity, difficult to ascertain but in principle inferrable from a thoroughgoing analysis of the problem solving process. One sees thus that even pursuing a far-reaching analysis and assuming perfect memory, it is not easy to derive the best strategy in the sense of minimiz- ing exploratory trials. When one takes into consideration the ambiguities present in the subject's mind, who may not even have the convenient visual representation of the time sequence in his mind's eye, one realizes that far more psychology than can be formally treated by information theory at this time is involved in the problem. Nevertheless, it is possible to cast the problem into information theoretical terms. One hopes, at any rate, that the concepts of information theory can be extended to cover situations where the subject's perception of the problem is an important unknown, That is, information theory may help formulate such situations in quantitative and analytic language. We have attempted to do so in the following way. We record the successive trials. Each trial must yield at least one of the sixteen 'crude facts', i.e. combinations of lights at each juncture t In view of the rule that each arrow must have a meaning, the number of ways values can be assigned to the arrows equals the number of distinct irreducible disjunctions among the subsets of the arrows. Thus for three converging arrows there are seven non-null subsets (i.e. 'disjunctions' involving only one subset), tliree disjunctions among the singles involving two singles, three among the doubles involving two doubles, three involving a single and a double, one involving all three doubles, and one involving all three singles, eighteen in all. Quantification of Performance in a Logical Task With Uncertainty 237 w hich do or do not light the node of the juncture. Some trials yield more than one fact, but some yield no new facts. The fact yield is also recorded, counting the facts which the subject had the opportunity to observe, but not counting the inferences which he could have made. Combinations of these facts make possible inferences about the meanings of the arrows at each juncture. For example, in Fig. 4, the facts '1 does not light 8' and '6 does not light 8' allow the inference that the arrows at this juncture should be labeled as shown in Fig. 5. 1 Fig. 5. These possible juncture inferences are also recorded, and thus their rate of accumulation. Finally, the juncture inferences collate into the bits of infonnation directly related to the solution of the problem — whether or not to push the circled buttons involved in the successive time periods. When all this information is available (by inference, of course), there is no formal uncertainty left in the problem. However, in most cases the problem is not yet solved. The extra trials made by the subject, who has the solution available by inference, constitute the 'inferential lag'. We thus have various possible measures of subjective uncertainty over and above the 'objective' measures. The most obvious difference is revealed in the repetitions of trials (ordinary failure to record information obtained). Next we have the explicitly redundant trials, that is, those which while being new trials yield no new facts. Next the inferential lag already mentioned. All these can be measured both in time units and in numbers of trials. The apparatus and the analysis of the problem solving process offer many opportunities for elaborate experimental designs, but they all hang on the question of how 'standard' these tasks are. In other words one needs to answer the question of whether there is a level of performance on each problem characteristic of a given subject, so that the variance in performance in a population of subjects can be adequately accounted for by a variance of some inherent abihty. Although this question has not yet been answered definitively, there are indications of a certain stability of performance. A set of experiments was performed at the Mental Health Research Institute, University of Michigan, in which the 'subject' in each case was a group of three students who solved the problems cooperatively by discussing each move and by coming to unanimous decisions on which move to make next. Eight such groups solved Problem 2 and then went on to solve Problem 3. The average number of moves for Problem 2 was about thirteen and for Problem 3 about nineteen. This is a first indication of the relative difficulty of the problems. That this difference is real is indicated 238 A. Rapoport by the fact that all eight groups increased the number of moves from Problem 2 to Problem 3. When the groups were rank-ordered on their performance on Problem 2, the rank order was preserved (with just one reversal of two con- secutive groups) on Problem 3. Another set of eight groups was given Problem 2 and then Problem 3 with a money incentive to minimize the number of moves on the latter. Under these conditions again all eight groups decreased the number of moves (averaging only eight on Problem 3). In spite of the radically changed situation, the rank order of these eight groups was again preserved from Problem 2 to Problem 3 (again with a reversal of only one pair of consecutive groups). When groups are rank-ordered according to time of solution, no discernible correlation appears from one problem to another. These results point to a possible stable relation between the complexity of the problem and the effec- tiveness of solution strategy adopted by each of our trios of subjects. The lack of correlation in rates of perfonnance points to possible extraneous effects such as the nature of the discussion process itself. At any rate the fact that the most prominent regularities are found in the performances as measured by the number of moves raises the hope that these regularities are the reflections of the uncertainty content of the problems as perceived by the subjects. It is noteworthy that, while on the level of observable crude facts and on the level of inferences about the meanings of the arrows, the two problems have the same uncertainty contents (about sixteen and ten bits each), on the level of major inference involving the circled buttons. Problem 2 has four bits of uncer- tainty while Problem 3 has six. The approximately 50 per cent increase in the average number of moves from Problem 2 to Problem 3 may well be a reflection of the increase in uncertainty on that level. Whatever the case may be, the results warrant further experimentation with a view of establishing the expected level of performance of a given subject on a given problem, once the set of uncertainties on various levels of observation and inference characteristic of the problem and certain factors of strategy efficiency characteristic of the subject are known. It is evident that the number of various problems which can be programmed into the PSl apparatus is astronomical. REFERENCES 1. B. DeFinetti: La Prevision : ses lois logiques, ses sources subjectives. Ann. Inst. Henri Poincare 7, 1-68 (1937). 2. L. J. Savage: The Foundations of Statistics, J. Wiley and Sons, New York (1954). 3. E. R. John: Contributions to t/ie study of the problem-solving process. Preprint No. 1, Mental Health Research Institute, University of Michigan, Ann Arbor (1956). 4. C. E. Shannon: A symbolic analysis of relay and switching circuits. Trans. Amer. Inst. Elect. Eng. 57, 713-723 (1938). 5. W. S. McCuLLOCH and W. Pitts: A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 5, 115-133 (1943). PART IV DESTRUCTION OF INFORMATION BY IONIZING RADIATION The disorganization of highly ordered macromolecules of biological importance by the action of ionizing radiation is a field of study owning a half-century of history, a tremendous literature, and possibly a somewhat feeble accom- plishment in terms of clear and unexceptionable conclusions. With the develop- ment of information theory, and its subsequent application to biological systems, there appears to be substantial basis for cherishing the hope that it may constitute a valuable tool in the analysis of the experimental results of radiobiology and their translation into knowledge concerning biological phenomena. The present section of the symposium is dedicated to this goal. The first two papers, by GoRDY and by Platzman and Franck, explore different aspects of the inter- pretation of physical and chemical effects of ionizing radiation on proteins and related substances; for without some measure of fundamental physical insight into the mechanisms of this action, the utilization of information theory in radiobiology would appear unlikely to emerge from an ineffectual state of pleasant vagueness. In the third paper, by Morowitz, positive steps are taken in the analysis of some relationships between information theory and ionizing- radiation action. The following two short papers, which stem from discussion by Koch and by Augenstine, are devoted to the almost perennial question of the role of sulfur-bonding in radiobiology, as was also, to a large extent, the further discussion at the symposium, part of which is summarized in the final pages of the section. It is disquieting to have to record that the views on this perplexing problem are still seriously discordant. R. L. P. 239 ELECTRON SPIN RESONANCE IN THE STUDY OF RADIATION DAMAGE* Walter Gordy Department of Physics, Duke University, Durham, North Carolina Abstract — It has been demonstrated by a Duke University microwave group that the electron spin resonance of the resuUing unpaired electron can give specific information about the radiation damage in proteins, nucleic acids, and many other biologically significant chemicals. The structures of their electron resonances show that free radicals of various types are formed from the different amino acids and simpler peptides by ionizing radiations. However, in numerous proteins only two structural patterns are obtained, either separately or in com- bination. One of these is like the common pattern obtained for cysteine, cystine, and gluta- thione and is believed to arise from an unpaired electron (electron hole) on the protein sulfur. The other pattern (obtained alone in proteins which have no sulfur) is a doublet characteristic of the interaction of the electron spin with the spin of a single proton. The latter appears to arise from an electron on a carbonyl oxygen interacting with a proton of the hydrogen bridge, or possibly on a — CH — of the peptide chain which has lost an R side group. There is no evidence that the ionizing radiation breaks the polypeptide backbone structure of the proteins. The results seem to require that an electron hole or vacancy created at a given location in the protein molecule can migrate to other locations where it has lower energy. I. INTRODUCTION Yesterday evening when coming over from the airport I discovered that I was in the car with a biologist. After making this discovery, about half way over, I asked my new acquaintance what it is that the biologists expect of the physicists, what help — if any — we physicists can be to them. He told me that we could give them better instruments. What they need as biologists, he said, are newer and better instruments to see into things. He made no mention of information or theory. I didn't ask him whether we were to bring the instruments or just send them by mail. Nevertheless, I think that a physical instrument which brings information out of biological things should be accepted as a ticket of admission to a discussion of infonnation theory in biology, especi- ally one held under the auspices of physicists ! The instrument which I offer as an admission ticket was not invented by me. Electron magnetic resonance was discovered in 1945 by a Russian scientist, Zavoisky (1). Nor can I claim to be the first to apply electron resonance to the study of radiation damage. That, I believe, was first accomplished by Hutchison (2), who in 1949 detected /'-center resonance in neutron-irradiated alkali halides. Our group at Duke University, we are proud to say, was among the first to show the applicability of electron magnetic resonance in the study of biological substances, and the first, we think, to detect such resonances in irradiated proteins. Combrisson and Uebersfeld (3), independently of our * This research was supported by the United States Air Force through the Air Force Office of Scientific Research ARDC contract No. AF 18(600)-497. 241 242 Walter Gordy work, found resonances in certain amino acids. Their results did not agree with ours, except with those for glycine. Our group has now obtained electron spin resonances of scores of biological substances which have been subjected to ionizing radiation. These include amino acids (4), peptides (5), fatty acids (6), nucleic acids (7), various proteins (4, 8), enzymes (8), homiones (9), and vitamins (9). Some of these results we think we understand, at least partially; others we do not pretend to understand. This does not discourage us, however. Some twenty to thirty years were required for obtaining reasonably definitive interpretations of x-ray diffraction patterns of a few of the simpler proteins. Nevertheless, it must have been apparent from the first that these patterns contained a wealth of information which would eventually be decoded by the persistent scientist. In electron spin resonance we now have a direct method for studying radiation damage which is comparable to the x-ray diffraction method for the study of structures. It is, in fact, a specific for such studies, for it 'sees' not the normal biological matter but the radicals, or broken pieces of molecules torn apart by ionizing radiations. Descriptions of microwave spectrometers for observation of electron magnetic resonances are available (10, 11). Such spectrometers can now be obtained commercially. Descriptions of theoretical methods and applications to chemical and biochemical problems are given in recent publications (11, 12, 13, 14, 15, 16). In the observation of electron magnetic resonance the sample to be investi- gated is placed in a microwave cavity at a point where the magnetic component of the microwave radiation is strongest. The cavity containing the sample is so placed in a d.c. magnetic field that the lines of the d.c. field lie perpendicular to the magnetic component of the microwave radiation. When the d.c. field is adjusted to the proper strength for resonance, microwave radiation will be absorbed. The value of the field for resonance is: Numerically, H (gauss) = 0.7 HSi' (Mc/sec)/^ (2) where g is the spectroscopic splitting factor for the paramagnetic species. It is found that for practically all organic free radicals, including those produced in solids by ionizing radiation, the g value is very close, within a fraction of a per cent, to the g factor for the free electron spin, 2.0023. This comes about because possible orbital moments are largely averaged out by the motion of the unpaired electron, or by the spreading out over a number of atoms (delocali- zation) of its molecular orbital. The persistent observation of a ^ factor near that of the free electron spin has led to the designation of this resonance as electron spin resonance. In the vector model, the electron spin vector would precess about the direction of the applied field H. Quantum mechanically there are only two stable orien- tations for this precessing vector, which represents an average or the 'expectation value' for the electron spin momentum. These correspond to the two observable components, +| and — |, of the electron spin vector along a fixed direction in space. Because of the interaction of the magnetic moment of the spinning Electron Spin Resonance in the Study of Radiation Damage 243 electron with H, the potential energy of the electron is slightly greater for one of the orientations than for the other. The difference in energy for the two orientations is equal to the microwave quantum energy hv which will induce the spin vector to flip over from one orientation to the other. The classical Larmor precessional frequency of the electron spin vector about the direction of Hh equal to the absorbed microwave frequency. Thus the precessing electron is in tune with, or at resonance with, the microwave radiation. In normal organic matter about us, the electrons are all — or nearly all — in the lowest orbital levels, with the maximum limit of two electrons in each molecular orbital. According to the Pauli principle, two electrons can share an orbital only if their spins are aligned in an antiparallel manner. If it is assumed that the spin vector of one electron flips over in an imposed field, that of its orbital mate must flip in the opposite direction at the same time, thus preventing any observable absorption or emission of radiation. To produce an observable electron spin resonance in normal organic matter, one must by some means lift electrons out of the completely filled orbitals of the ground level. Strong ionizing quanta, such as those of x-rays, can eject electrons from ground molecular orbitals with sufficient energy to free them entirely from the parent molecule. If a molecule loses a single electron through ionizing irradia- tion, the ionized molecule — if it holds together — will have a single unpaired electron in one of its orbitals. This electron is now free to flip over in an external field without the opposite flipping of a partner. The singly ionized molecule is thus paramagnetic and can execute electron spin resonance. Further- more, the electron which is knocked away from one molecule may become attached to a neighboring molecule and thus convert it into a negatively charged radical. Since the latter molecule is presumed to have all its bonding orbitals filled, the new arrival must go into an orbital of higher energy and remain unpaired. Thus resonance of electrons on negatively charged molecules might Hkewise be detected. If the electron is ejected with sufficient energy it may, of course, ionize several molecules before coming under the control of a particular molecule. The end result is the same, however, except that a single quantum has, in effect, been able to ionize more than one molecule. Two types of charged radicals are thus produced. If the barrier to the return passage of the electron between the molecules is high, as is the case in most organic solids, a sufficiently high concentration of charged radicals can be built up in this way to give a detectable electron spin resonance. The molecules may be small ones such as the amino acids or long-chain macromolecules such as the proteins or nucleic acids. The only requirement is that the separated electrons cannot easily become paired again, i.e. that the radicals produced by ionizing radiation have a lifetime sufficiently long for a detectable quantity to be built up. II. NATURE OF INFORMATION CONTENT IN ELECTRON SPIN RESONANCE If the spin of an odd electron of a radical were entirely free from perturbing influence of its environment, its resonance would be a single, sharp, isotropic line with a g factor of 2.0023. Not much information is contained in such a simple signal, although one could measure the lifetime of the radical from 244 Walter Gordy its rate of decay. Also, the very fact that electrons could achieve such freedom witliin an organic solid might itself be classed as desirable information. For- tunately, however, the electron resonance signals are often rich with information about the environment of the unpaired electrons. Our problem is to decode their messages. There are at least tliree important sources of information in these resonances. The first is the hyperfine structure arising from inter- actions of the electron spin moment with magnetic moments of various nuclei around or near the unpaired electron. The second is the small residual spin- orbit interaction which in some instances makes the g factor slightly anisotropic and different from the free spin value of 2.0023. The third is the information which can be obtained from the line widths and shapes. The most important of these sources is the nuclear hyperfine structure. Most instruments used for detection of electron spin resonance plot the intensity of absorption at a particular frequency as a function of d.c. magnetic field. The appearance of the plot depends upon the instrument as well as upon the actual, intrinsic shape of the resonance. I shall not discuss possible variations in the actual line-shapes, but shall here assume that the resonances have gaussian shape when the intensity of absorption at a constant frequency is plotted as a function of d.c. magnetic field. A high-fidelity receiver and recorder (or cathode ray scope) would reproduce closely the actual shape of the resonance curve, as shown in Fig. 1(a). The high-fidelity systems are not, however, the most sensitive systems. The most sensitive methods of detection employ modulation of the resonance relative to the observation frequency. A narrow-band amplifier is tuned either to the modulation frequency or to a higher harmonic of tliis frequency. If one uses a frequency modulation which is very small as compared to the width of the resonance and a phase-sensitive amplifier tuned to the modulation frequency, a curve like that in Fig. 1(b) is obtained. This curve represents the first derivative of the actual line-shape. If one uses such a receiver and tunes to the second harmonic of the modulation frequency, a curve hke that in Fig. 1(c) is obtained. This curve represents the second derivative of the actual fine-shape. Both the first and second derivative curves are commonly employed in display of electron spin resonances. In interpretation of the curves it is desirable to know what method of detection has been employed, especially when there are structural components incom- pletely resolved. In the illustrations which follow we shall sometimes use first and sometimes second derivative displays. This brief description of the appearance of the signals and the simplified theory of the structure of the resonance given below will, I hope, make it possible for you, whether you are a biologist, chemist, physicist, or hybrid, to share with us some of the fun of trying to decode the complex microwave messages which we have been receiving from biological substances. You will be able, I hope, to decide for yourself what is definitely proved by the resonances, what is strongly suggested but not proved, and what is merely hinted. 1 . Nuclear Hyperfine Structure The hydrogen nucleus, with a relatively large magnetic moment, 2.79 nm, and nuclear spin of |, is abundant in all organic matter. The only other nucleus with a non-zero spin abundantly found in biochemicals is N^^ (/ = 1 and Electron Spin Resonance in the Study of Radiation Damage 245 jij = 0.40 nm). Carbon, oxygen, and sulfur are of course also prominent constituents of biochemical matter, but their most abundant isotopes have zero nuclear spins and hence cannot interact with the electron spin. In strong resonances one might detect effects caused by C^^ (spin \ and natural abundance 1.12 per cent) or S'^^ (spin 3/2 and natural abundance 0.74 per cent). For some (o) ACTUAL LINE SHAPE (c) SECOND DERIVATIVE Fig. 1 . Appearance of resonance signals as detected in various ways : (a) High fidelity, (b) First derivative curve obtained by small modulation of the resonance with a phase-sensitive receiver tuned to the modulating frequency, (c) Second derivative curve obtained by small modulation of the resonance with phase- sensitive receiver tuned to twice the modulation frequency. substances one can obtain samples concentrated with C^^, S^ or O^'^. Hyperfine structure of their miclei thus obtained will greatly augment the information obtained from proton hyperfine structure, but it is fortunate for these studies that C^^ is not the more abundant isotope of carbon. If hyperfine structure from all the nuclei were present at one time, the resulting pattern would often be unresolvable and its decoding thus more uncertain. As it is, there is seldom 17 246 Walter Gordy any ambiguity about the identity of the nucleus which gives rise to the nuclear hyperfine structure of electron resonances in irradiated organic matter. Usually, it must be hydrogen. By substitution of deuterium for hydrogen, one should often be able to learn which hydrogens give rise to a particular splitting. When the electron spin resonances of organic radicals are observed in the microwave region at frequencies of 30,000 Mc/sec, the corresponding magnetic field required is 10,700 gauss. A magnetic field of such strength is usually sufficient to produce the Paschen-Back effect, in which the / • S coupling is broken down and both / and S precess about the direction of H. Under these conditions the resonance frequencies of the various components at constant field strength Hq can be expressed as: hv = gpoH, + 2 A,m, (3) i where A^ is the coupHng constant of the electron for a particular nucleus / with spin /^ and where the magnetic quantum numbers have the values : nu = I,, 7,-1, ••• -4 (4) Usually the resonances are observed at a fixed frequency, Vq, by variation of the d.c. magnetic field. The resonant field strengths for the various hyperfine components are then: //=7/o + ^p,m, (5) = /^o + 2 Ai/,m, (6) i The summation is again taken over all the coupling nuclei for each combination of the magnetic quantum numbers. All orientations of a given nucleus (all values of its m) are equally probable and independent of those of the other nuclei. In this expression Hq = hvjg^ is the resonant field strength for the central component of the structure at the observation frequency, Vq, or that for resonance if there were no nuclear perturbation; AH^ is the component separation (in magnetic field units) caused by a particular nucleus /. Obviously, A/f. = AJgfi. In these applications we can set g as equal to 2.00 and write : A/f, (gauss) = Ai (Mc)/2.80. (7) If all the coupling nuclei in a given free radical have the same coupling to the electron spin, one can define 7^=2 4 (8) i and mj, = T, r-1, r-2, •••, -T, (9) and can write equation (6) in the simpler form: H^Ho + (AH)Mj. (10) There will be {2T -\- 1) components corresponding to the different values of M,p. This simplification is often possible in organic free radicals in solids where the coupling nuclei are all hydrogens. It is apparent that, where all Electron Spin Resonance in the Study of Radiation Damage 247 the equally coupling nuclei have the same spin, 7= nl, and the total number N of components of the multiplet will be: N = 2nl-\\ (11) or Thus n equally coupling hydrogens (/ = \) gives n + 1 components. The intensities of the components are proportional to the number of different combinations of the /n/s which give the same sum 2 '"j or same value of M^.. i Because all the -\-h and —\ orientations of n hydrogens are equally probable and mutually independent, the intensities of a multiplet formed by equally coupling hydrogens will be gaussian. The interaction constant A^ of the electron spin with the moment of a particular nucleus may contain both an isotropic and an anisotropic component. The isotropic component, the Fermi term, is independent of the orientation of the sample in the magnetic field and arises from the non- vanishing probability density, ^'q t/'o*? of the electronic wave function at the nucleus in question. Since only the s atomic orbitals are non- vanishing at the nucleus (radius r = 0), the presence of an isotropic coupling tenn for a particular atom in a molecule generally indicates 5 character in the bonding orbitals of that atom. For an unpaired electron occupying wholly an s orbital of a particular atom, the coupling to the nucleus of that atom arises entirely from the non- vanishing density ipQ i^)q* at the nucleus and has the value (17): A. = y fif^igi Wo n* = 3 ^^3 (13) where /9 is the Bohr magneton; /5j, the nuclear magneton; gj, the g factor (/ij//) of the nucleus; /;, Planck's constant; c, the velocity of light; R, the Rydberg constant; a, the fine structure constant; Z, the effective nuclear charge; and «, the total quantum number. For atomic hydrogen in the ground state, A is known to be 1420 Mc/sec. This value with equation (7) gives A/Z^ = 507 gauss as the expected splitting for the atomic hydrogen doublet for the strong- field case {H ^ ^H^). The non-isotropic components are zero because of the spherical symmetry of the s orbital. Thus the isotropic coupling to the nucleus of a particular atom gives a good measure of the s orbital contribution of that atom to the molecular wave function of the odd electron in a free radical. An electron at a fixed distance from a nucleus / with non-zero spin will experience a magnetic field component arising from the magnetic moment of the nucleus. If the spin vectors of both the electron and the nucleus precess about the direction of an applied field H (this corresponds to the strong-field, Paschen-Back case), the non-vanishing field component at the electron, A//, caused by the nucleus, will lie along H and will have the value: (A//) = {jY^.l^^i^y^O C0S2 0 - 1) (14) 248 Walter Gordy where /, m and ^7 are the spin, magnetic quantum number, and magnetic moment (in nm units) of the nucleus, f^j is the nuclear magneton, r is the radius vector from the nucleus to the electron, and d is the angle between r and H. Although the nucleus may be regarded as located at a fixed point within the molecule or crystal, the electron definitely cannot be so regarded. Hence, to find the averaged or effective (A/r)eff acting on the electron in a molecular orbital ip, we must average the above quantity over the orbital ip. Thus {^H\s^i^-^[^J^,^w{^J^Oco^''0-\)^*dT . (15) Since the coordinates are separable, we can write this equation as Av (A//)eff = (^- j /^2/^Z^73/^ <3 C0S2 0 - 1 >^„ (16) where /'■'K ipr* d7 m and (3 cos^ 0 — 1)^^ = WeO cos^ 0 — \)fe* dr . To attack such a problem one can assume, as is usually done in other calculations of molecular orbitals, that ip is a linear combination of atomic orbitals, ip^, tPi, etc. We then readily get a part of the solution for we already know, at least to a fair approximation, \^/ and (3cos-0— 1>av for electrons various kinds of atomic orbitals. Expressions for these to various degrees of approximation together with couphng constants actually measured are available in standard texts on atomic spectra (17, 18). There is more to the problem than this, however. Although an overlap or cross temi of the forni y>,lllr^)(3 cos- 0 — 1)^)* may possibly be neglected, an electron in an atomic orbital of atom B might have a significant interaction with the nucleus of an adjacent atom A. It is thus necessary to include terms of the form : Jv.,(;l-J(3cos2 0„,-l)v',r/T, (17) where /•„;, and O^j, are the coordinates of an electron on atom B referred to the nucleus of atom A as the origin. The values of these terms are sensitive functions of the hybridization of the atomic orbitals and of the direction of the projections of the major lobes of the hybridized orbitals. As we get greater skill in the procedure, these orientation-dependent couphng terms should give additional information about orbitals of radicals. Expressed in convenient numerical units equation (16) becomes: A// (in gauss) = 5.05 ^ (4) (3 cos^ d - 1>av, (18) ^ X'' / AV where /Uj is in nm units and r is in A. Electron Spin Resonance in the Study of Radiation Damage 249 In simple cases where single crystals can be prepared, it should be possible to measure (1//'^>av for the interaction of an electron in atomic orbital of atom B interacting with the nucleus of another atom A. Such applications are made later in the discussion. If (1//'^>av~* is greater than the interatomic distance, the electron may be in a hybridized orbital of B which projects away from A. If it is less than the atomic distance, the electron may be in a hybridized orbital which projects toward A. In some instances {\lr^)A\ may be so large that the field of the electron at the nucleus may be greater than the applied field. The nucleus would not then necessarily precess about the direction of H, and the above fommla would not hold for all values of 0. It should still hold when 0 equals zero or 90°, for then the field of the electron at the location of the nucleus would have, on the average, the same direction as H. If the cloud of the electron is symmetrical about the bond axis between A and B, the angle 6 would, in effect, measure the orientation of the bond axis in the field H. For this case (3 cos^ d — 1)av equals 2 for (? = 0 (bond axis parallel to //), and (3 cos^ 0 — l)^^ equals — 1 for 0 = 90° (bond axis perpendicular to H). Thus the A// should have twice the value for 0 equal to zero as that for d equal to 90°. The dipole-dipole interaction of the electron with the nucleus averages to zero when the electron is entirely outside the nucleus and is moving in such a manner that its averaged density achieves spherical symmetry about the nucleus during its lifetime in a spin state. Nuclear hyperfine structure of any type becomes independent of the mag- netic field strength after the field becomes sufiiciently strong to achieve the Paschen-Back case, which is assumed in the above treatment. Thus nuclear hyperfine structure can be readily distinguished from the splitting wliich arises from anisotropy in the g factor, discussed below, if measurements are made at two or more frequencies, both with strong fields. Although the direct- dipole type interaction with the nucleus varies with orientation in the field, it does not vary with the magnitude of the field after the strong field case is reached. Figure 2 shows the type of hyperfine structure theoretically predicted for the strong field case for various radicals with equally coupling nuclei having spins of I {H or F, for example). Figure 3 illustrates a few cases where the coupling of one or two of the nuclei differs from that of the others. It is appa- rent that these cases are easily distinguishable. 2. Residual Spin-Orbit Coupling If the odd-electron density of a radical is largely concentrated on a non-5 orbital of a single atom of a radical or is shared mainly by only two atoms, as it is on the — N — N — group of diphenyl picryl hydrazyl (DPPH), effects of spin-orbit interaction are not entirely negligible. The orbital momentum will be oriented by the strong electrical force of the chemical bond and will not be free to precess about the applied field. Bond-oriented orbital components will give rise to an observable anisotropy in the magnetic susceptibility and thus in the observed g factor. If the odd electron wave function is symmetric about a particular bond as in DPPH, the observed g factor will reflect this symmetry: if all such bonds in a given sample were oriented along the applied H, the ^n factor would differ from the g^^ observed when the bonds are all oriented 250 Walter Gordy perpenidcular to //. For an arbitrary orientation 0 of the bond axis with H, the observed ge factor would have the value: ge V^ii^ cos'^ 0 + g^^^ sin^ 0 (19) In a sample in which the bond angles have arbitrary orientations in the field H such as would be true in a powder or polycrystalline sample, the NO. OF EQUALLY COUPLING HYDROGEN SPINS EXPECTED PATTERN OF RESONANCE Fig. 2. Types of hyperfine structure predicted for strong-field case for various radicals having ditTerent numbers of equally coupling hydrogens or other nuclei of spin |. resonance absorption would spread over all values of the field intermediate between that corresponding to the resonance value for g,, and g^^. The ^j^ would apply for any orientation of // in a plane perpendicular to the bond axis, whereas the g,, value would apply only for H along the bond axis. Thus for random orientations in the polycrystalline samples, the g^^ value has the greater weight, and the resonance has an asymmetric form with the highest Electron Spin Resonance in the Study of Radiation Damage 251 peak corresponding to the g^^ value. Such a resonance will have a shoulder or shelf on one side with the edge of the shoulder corresponding to ^i,. First derivatives of an asymmetric resonance arising from an anisotropic g factor in x-irradiated cystine are shown in Fig. 4 for three different observation frequencies. That the apparent structure in these curves is due to anisotropy NO OF RELATIVE HYDROGENS COUPLING EXPECTED PATTERN OF RESONANCE Fig. 3. Some illustrative theoretical hyperfine patterns of radicals with two sets of H nuclei (or other nuclei of spin |). All nuclei of one set have the same coupling, but those of the two sets dilTer as indicated. in the g factor has been established by measurements on a single crystal of cystine at different orientations in the field (19). Such curves show some differences depending upon amount of modulation, variations in natural line widths, degree of anisotropy in g, as well as variations in observation frequency or H value. Nevertheless, there should always be a bend point in the derivative curves corresponding to the H for resonance at ^^j^ and a lesser one for _^||. This fortunate circumstance allows measurement of ^^^[ and gj_ even in poly- crystalline samples. The identity of these bend points, if in doubt, can usually 252 Walter Gordy be established by variation of the modulation amphtude and observation frequency. The outermost bend points will in general correspond to g,^ and g^^. III. FREE RADICALS IN IRRADIATED AMINO ACIDS AND SIMPLE PEPTIDES The work of our group at Duke University has revealed that the isotropic ^orbital contributions of hydrogen atoms in aliphatic hydrocarbon radicals are very significant and that they give rise to hyperfine structure in the spin resonance of these radicals v/hich is frequently of the order of 100, and sometimes as much as 200, gauss. This couphng is an order of magnitude greater than that generally found for the aromatic ringed radicals (14, 15, 20) which can be prepared chemically and observed in solution. Furthermore, the first measure- ments indicated, and later work on single crystals (21) confirmed, that the isotropic j-orbital coupling to the hydrogen nuclei in aliphatic hydrocarbon radicals is generally much greater than the orientation-dependent, dipole-dipole component. This very fortunate circumstance makes possible detection and often identification of the aliphatic hydrocarbon radicals made by irradiation of solid matter in the polycrystalline powder and even in impure biological solids. In other words, it seems possible with microwave spectroscopy to 'fingerprint' many of the common radicals produced within soUd matter by irradiation. I need not emphasize the usefulness of such a set of fingerprints for the study of radiation damage. There are two important factors which we beheve to be mainly responsible for the reduction of the anisotropic nuclear coupling in hydrocarbon radicals. One of these is the spreading of the odd electron density over a large molecular orbital so that there is no appreciable fraction of the total density near a parti- cular nucleus. The other is the twisting, turning, tunneUng, tumbling, or other motion of the radicals, or their parts, within the solid cages in wliich they are trapped. The first is generally more important for large radicals than for small ones, and the latter is generally more important for room temperature and elevated temperatures than for lower ones. These properties of aliphatic free radicals and their remarkably long Hfetime within solids were not predicted by theory. The conclusions were forced upon us from the experimental evidence for them. Furthermore, this pronounced isotropic interaction through the 5-orbitals immediately gives much information about the electronic wave functions and structure of hydrocarbon radicals. The large coupling to the H nuclei in the CH3 radical (total spread of quartet 70 gauss) indicates that this radical is not planar. Amazingly, the characteristic pattern of the ethyl free radical, C2H5, is a symmetrical sextet (or approximately so) of about 130 gauss spread. This indicates equivalent, or nearly equivalent, coupling to the electron spin of all five protons. Fig. 5 illustrates some characteristic hyperfine patterns of hydrocarbon radicals produced by x-irradiation of some simple peptides. Compare these with the theoretical patterns for different numbers of equally coupling protons in Fig. 2. Similar patterns have been obtained by irradiation of amino acids (4) and other compounds (6, 22, 23) with x-rays and with ultraviolet light (24). Figs. 6 and 7 illustrate somewhat more complex resonances. Fig. 4. First derivative curves at different frequencies of powdered cystine after x-irradiation in a vacuum. The markers at the base are 68 gauss apart. The top curve for 2.7 kMc requires a magnetic field of 960 gauss; the middle curve at 9 kMc, one of 3200 gauss; the bottom curve at 23 kMc, one of 8200 gauss. to face p. 250 GUYCYL GLYCYL GLYCINE / CHuaRACETYL DL ALANINE Fig. 5. Some illustrative patterns of resonances of x-irradiated peptides (second derivative curves). The markers at the base are spaced 68 gauss apart. The observation frequency is 9 kMc. From G. McCormick and W. Gordy (5).) GLYCYL DL VALINE \ \ \ Fig. 6. Hyperfine pattern (second derivative curve) of the radical produced by x-irradiation of glycyl DL-valine. Marker spacing is 68 gauss. Observation frequency, 9 kMc. (From G. McCormick and W. Gordy (5).) Fig. 7. Hyperfine pattern (second derivative curve) of the radical produced by x-irradiation of acetyl DL-valine. Marker spacing, 68 gauss. Observation frequency, 9 kMc. (From G. McCormick and W. Gordy (5).) Fig. 8. Resonances (first derivative curves) obtained for \-irradiatcd silk witii strands oriented parallel and perpendicular to the magnetic field. The obser\ation frequency is 23 kMc. (From W. Gordy and H. Shiklds (8).) — - CYSTINE \ \ \— \ Fig. 9. Resonances (first derivative curves) of x-irradiated hair compared with similar resonances for cystine and cysteine. Mariner spacing at base, 68 gauss. Observation frequency, 23 kMc. (From W. Gordy and H. Shields (8).) Fig. 10. Resonance (first derivative curve) of bovine albumin which represents a combination pattern of the glycyl-glycine (or silk) doublet and cysteine (or hair) resonance. Observed at 23 kMc. (From W. Gordy and H. Shields (8).) Fig. 11. Resonances (first derivative curve) of x-irradiated feather quill (of a goose) at 23 kMc for parallel and perpendicular orientation in the magnetic field. (From W. Gordy and H. Shields (8).) Fig. 12. Resonance (first derivative curve) of x-irradiated insulin observed at 23 kMc. (From W. Gordy and H. Shields (8).) Cholesterol, Cg^H^^OH X-rayed in vacuum (indicating the top curve) X-rayed in air (indicating the bottom curve) Fig. 13. Resonance (first derivative curve) of cholesterol at 2.7 kMc, x-irradiated in a vacuum (upper figure) and in air (lower figure). (From H. N. Rexroad and W. GoRDY (24).) Electron Spin Resonance in the Study of Radiation Damage 253 There are far too many radicals already observed in irradiated amino acids and peptides to discuss them here. I should like to mention one more, however. The pattern of Fig. 7 for acetyl valine consists mainly of a set of nine syimTietrical doublets spread over 200 gauss. There is another resonance near the center of the group which I ignore for the present discussion. Seemingly, the nine doublets must arise from eight equally-coupling protons and a ninth with coupling only about half as much as each of the eight at room temperature, and only about a fourth as much at liquid air temperature. This pattern requires an almost unimaginable radical. The odd electron must spread two-fifths of its total density in \s orbitals of the eight equivalent hydrogens. This indicates a radical with a high concentration of hydrogens. It is difficult to design a radical with eight equally coupling hydrogens, especially with a ninth coupling differently. The (CH3)3C radical would have nine equally coupling hydrogens wliich would be expected to give a hyperfine spread of the order of 200 gauss. If we should assume that one of the hydrogens in (CH3)3C is replaced by a group RH with only one coupling hydrogen (such as OH) and one which does not noticeably disturb the couphng of the other two, we would have a radical which might account for the acetyl valine pattern of nine doublets. IV. RADIATION DAMAGE IN PROTEINS In contrast to the varied hyperfine patterns found for the resonances of the x-irradiated amino acids and simple peptides, we have found mainly (but not exclusively) two patterns either singly or in combination for numerous proteins. One of these patterns consists of a simple doublet arising from interaction of the odd electron spin Vv'ith a single proton spin. The other pattern is a field-dependent one like that of powdered or polycrystalline cystine, cysteine, or glutathione. Fig. 8 illustrates the first type; Fig. 9, the second; and Fig. 10 is a combination of the two patterns. In our first papers on electron resonances in irradiated proteins (4), we suggested that the doublet pattern in the proteins might arise from an odd electron localized mainly on an oxygen joined by a hydrogen bridge as indicated in Structure II. Model I represents a structural section of the unirradiated /?-keratin protein. The doublet structure, we thought, might arise from dipole-dipole interaction of the electron spin with the proton of the hydrogen bridge. Partly to test this hypothesis, H. W. Shields and the author (25) have made observations on strands of irradiated silk directed along the applied magnetic field, and also on strands directed perpendicular to it. It is known from infrared and x-ray studies (26) that hydrogen bridges in silk he approximately in a plane perpen- dicular to the direction of the silk strands. If we assume, for simplicity, that the odd electron density is symmetrically localized on the oxygen, the 0 of equation (18) would measure the angle of the O — H axis with the magnetic field. Hence, when the silk strands are along the apphed field, 6 equals 90° for all hydrogen bridges, and the doublet splitting is the same for all radicals of the silk. Under these conditions one would expect a clear resolution of the doublet. When the silk strands are perpendicular to the applied field, the 254 Walter Gordy hydrogen bridges have all orientations with the field from 0 to 180°. With this arrangement one would expect the individual components of the doublet to be broader and less well resolved than for the parallel case. These features are not completely in accord with the observed results on silk. The doublet splitting for 0 = 0 (parallel case) is found to be approximately 25 gauss, some- what larger than that previously estimated from the polycrystalline material, and also significantly larger than that expected for the hydrogen bonded model. H O C c \ / \ / CRN O H H O N R C / \ / \ . C N H O C CRN H /b O I H N RIC / \\A \ C \ N + H H H (I) H (II) Furthermore, the separation of the doublet seems to be greater for the parallel case. It should be appreciated that what is proved for silk is simply that the radical formed is one in which the odd electron interacts with one and only one proton, and that this interaction is at least partly anisotropic. Later we hope to obtain more specific evidence from deuterium substitution in glycyl glycine, which appears to have the same doublet as that for silk. Irradiated feather quill gives a composite pattern of a doublet and the cysteine-like resonance. However, the doublet is not as widely spaced as that for silk and is not resolved for a polyoriented sample. It has been found (19) that the strong component to the left of the cysteine-like resonance in feather quill (Fig. 11) is partially resolved into a doublet when the feather quill is arranged parallel to the applied magnetic field, whereas it has only about half the width of the unresolved resonance for the perpendicular orientation. Presumably the structure of the feather quill is that of the alpha helix of Pauling and Corey (27), with the helix axis along that of the quill. Interestingly, the cysteine- like component of the resonance is not orientation-dependent. We believe for reasons given later that this situation indicates that the S — S or the C — S bonds of the quill have many different orientations with respect to the quill axis. A resonance found to be prominent in x-irradiated proteins which contain sulfur is like that of cystine, shown in Fig. 4. Biological substances such as hair (Fig. 9), hoof, horn, and feather have this as the predominant if not the Electron Spin Resonance in the Study of Radiation Damage 255 only pattern, despite the fact that the cystine or cysteine content is only a few percent. The fact that the same pattern, but one very dilTerent from any so far obtained from non-sulfur compounds, is observed for many sulfur-containing proteins and for cysteine, cystine, and glutathione convinces us that the odd electron giving these resonances is essentially localized on sulfur. Whether it is on a single sulfur or is shared by two sulfurs of the S — S link, as originally suggested, remains a question to be answered by later work. That the odd electron is localized mainly on one or two atoms is borne out by the large amount of residual spin orbit coupling evidenced by the anisotropy in the observed g factor, as already explained. Because cysteine with only — SH sulfur gives the same type of resonance as cystine with — SS — sulfur, it is uncertain whether the electron wliich gives rise to the characteristic resonance of Figs. 4 and 9 is on a single S or is shared between two sulfurs to form a 'three-electron bond'. When the plus charge accompanying the odd character arrives at the S of the — SH of cysteine, it would probably 'shock' off either the naked proton to leave the neutral free radical RS-, or the H atom to leave RS+, where R represents the part of the cysteine exclusive of the SH group. In the latter case, the H atom would escape through the lattice or react with something. (We have been unable to detect the free hydrogen radical at room temperature in any irradiated substances.) We do not know at tliis time which if either of these two events occurs. Inter- estingly, RS+ is not a free radical, and no resonance would be detected for this case until further events had transpired. At room temperature, however, the molecules may flop about sufficiently to allow the RS+ to react with the — SH of a neighbor and to release another H and form the same charged disulfide radical which has been postulated for irradiated cystine. The common patterns of cystine and cysteine might be thus explained. I should say, however, that the two patterns although alike are not identical: the resonance pattern of cysteine has a slightly greater over-all width than that of cystine, a variation which we believe arises from the difference in dielectric medium. If the radicals were diff'erent — if one were RS- and the other were R • (SS)'" • R — a much greater diff'erence would be expected. If the resonance in irradiated cysteine arises from RS- mentioned above, the resonance of cystine must arise from the same radical, which would result first from ionization of the cystine molecule and later from rupture of the S — S bond to leave RS- and RS+. There is no evident mechanism by which the positive charge could disrupt the S — S bond other than the initial 'shock' of the sudden arrival of the charge. Such 'shock' effects can be anticipated from the Franck-Condon principle (28). They would hardly be expected to break the S — S link, because its potential curve would be lowered and its bond length shortened by the removal of an anti-bonding electron. The positive charge would have no preference for either sulfur; and, if the S — S bond holds, the odd electron would be shared equally by both sulfurs to form an additional half-bond. The Franck-Rabinowitch caging principle (28) would also tend to prevent the breaking of the S — S link by the 'shock' efi'ect. The two S atoms are in a sense caged and hindered from flying apart by the large inert R groups to which they are attached. Any 'shock' energy would probably be dissipated as vibrational energy throughout the whole dimeric molecule. Such a charged 256 Walter Gordy link would of course tend to attract other agents such as O2 or H2O which might later sever the bond or an electron which would restore the normal S— S hnk. Although we are not yet certain whether the cystine or cysteine-hke resonance arises from radicals of the type R — (S • • • S)+ — R or RS-, we are inclined to favor the latter. It would seem that the neutral radical would enjoy the longer life and hence be the more probable one to be detected. Furthennore, in the RSH compounds the formation of the RS- radical would require the simpler process. With the present information we are inclined to beheve that R — (S • • • S)+ — R is the primary radical formed by ionization of the disulfide compounds but that the healing of the molecule through capture of an electron or later rupture of the charged link, probably by attraction of other groups, or molecules, may occur so rapidly that this charged radical is not the one detected, but rather the neutral radical R — S. However, our interpretations are still tentative. Because we consider the question an important one we are continuing to investigate it experimentally. Studies using S^^ can clear up this uncertainty. What already seems established is that the odd electron giving rise to the pattern is essentially localized on the sulfur. The large anisotropy in the g factor for the cystine-type resonance suggests the potential usefulness of this resonance for obtaining structural information about the proteins. Studies by Shields and the author on single crystals of cystine (19) showed a resonance simpler and much narrower than that for polycrystalline cystine, and one which shifted position sensitively with orienta- tion in the magnetic field. After this observation the same crystal was crushed up and found to give the resonance pattern characteristic of polycrystalline cystine, shown in Fig. 4. Observations (19) on strands of hair and on feather quill at various orientations in the d.c. magnetic field shovved only the poly- crystalline type of cystine resonance for all orientations. It is significant, we think, that the cystine-like resonance in these proteins is not orientation- dependent, for that fact gives convincing proof that the bonds to sulfur, either the C — S or S — S links, in the keratins are randomly oriented (in contrast to hydrogen bridges). We have also made measurements (19) on stretched and unstretched hair and found no significant variation in its cystine-hke resonance pattern. In all cases it is like that of the polycrystalline cystine. The resonance of x-irradiated insuhn may exliibit a third type of protein resonance (cf. Fig. 12). It has the characteristic sulfur or cystine-like pattern but with a relatively sharp resonance superimposed (at the left of the pattern). Although it has the same g factor — that of the free electron spin — this com- ponent to the left seems too sharp to be classified as an unresolved doublet like that of feather quill or silk. Possibly this sharp component of the insulin resonance may arise from an electron trapped in one of the unsaturated ringed structures known to be on the side chains of this protein. The ringed structure may act as a sink or trap for the odd electron produced by ionizing radiation. We have found a similarly sharp resonance (29) for x-irradiated polystyrene, where the odd electron observed is believed to be trapped in the aromatic rings attached to the backbone structure. Considering the varied patterns which we found for the resonances of the different amino acids and simple peptides, it was at first surprising to us that Electron Spin Resonance in the Study of Radiation Damage 257 the proteins gave such shnple patterns with the same few features, described above, repeating so often either singly or together. We were forced to conclude that the electron hole or vacancy created by an ionizing quantum or particle at any given locality in the protein can move through the polypeptide chain until it reaches one of a few traps or sinks where it becomes lodged. One such low-energy trap we believe is sulfur. Both — SH and — S — S — groups are effective traps. Possibly the unsaturated rings of certain side chains are an important trap. Furthemiore, we must postulate that there are effective traps for the electrons knocked away in the ionization process since these do not always seem to be able to return readily to fill the hole. Because they have not given recognizable resonances, we do not speculate on the negative traps. For most of them, the resonances may be too broad for detection. V. PROTECTIVE MECHANISMS I am sure that there are many who have suspected that some proteins when ionized can hold together and conduct the electron hole to certain side-chain groups such as the sulfur link. I think that I have heard Professor E. C. Pollard, of Yale, and members of his group express such views. However, from my brief and sketchy acquaintance with the hterature in tliis field I surmise that this question has been a highly debatable one. In the microwave resonances we have a new and perhaps more direct type of evidence in favor of the migra- tion of the electron holes to certain side-chain groups. Now that there is new evidence for effective resistance to the breaking of the polypeptide backbone of the proteins by ionization, it is interesting to speculate on the reasons why this is true. If one of the electrons of a locahzed, covalent bond between two atoms were suddenly removed, the two atoms might — according to Franck-Condon principle — become dissociated wliile trying to adjust to the new and shallower potential curve with the longer equilibrium distance commensurate with the 'one-electron bond'. It might be supposed that the Franck-Rabinowitch caging would help to prevent any two atoms of a protein chain faced with such an emergency from becoming dis- sociated. However, the evidence which we have obtained for the migration of the electron vacancy to a sink in the side chains indicates that a particular bond of the polypeptide chain does not have to face the Franck-Condon catastrophe because the bonds are not in a strict sense localized. If we imagine that charge density equivalent to a single electron is removed completely from the localized region of two adjacent atoms along the main chain, we must, at the same time imagine that tliis charge density is restored quickly, before the atoms have time to move significantly apart, by the flow of electronic charge from a side chain group such as the S— S hnk. It might be better to think of the ionization as taking place only at these sites where the electron vacancy is detected. A molecular chain or polymer which can conduct a hole out to a non-essential side-chain sink or to a point where a simple recapture of an electron restores the status quo has, in effect, a built-in, remarkably effective method of self- protection from radiation damage. Such polymers have a high survival value in a world where ionizing radiations are ever present. 258 Walter Gordy Although our measurements were made in dry — reasonably dry — samples, it seems likely that the same transfer of an electron hole to low-energy sites, such as the side-chain sulfur, would take place in the proteins of living systems. The better mobility of charges in the more fluid systems should only speed up the recapture of an electron and hence the recovery of the system. Of course the attack on the charged radicals such as — (S — S)+ — by molecules like H.2O would also be speeded up in the living systems, but in the living systems the electron recovery might well be the more rapid. Even if a break in the S — S bond should occur, this might be less damaging and more easily healed than a break in the polypeptide trunk line. We seem to be proposing here a self-protective mechanism which would prevent almost any radiation damage whatever to proteins. This is not true for several reasons, one of which is that not all proteins have — S — S — hnks in their side chains. There are other traps for the 'hole' where bonds are probably broken as postulated for silk, or for the sulfhydryl group, where the hydrogen atom or proton is believed to be freed. A free hydrogen atom could cause trouble in the living system, even though it could be temporarily spared from the S — H group of the protein. Moreover, not all damage to proteins in the hving systems is due to the direct ionization of the protein which we have been discussing here. Much of the damage (30) is thought to be done by radicals such as H, OH, and OOH produced by radiation in the inter-pene- trating fluid, which later attack and damage the protein. These are the so-called indirect effects. About the time of our initial experiment on the proteins, a very significant experiment of an entirely different kind was in progress by Eldjarn, Pihl, and Shapiro (31) which indicated that the indirect effects are probably not as significant as had been previously thought, and that a high degree of protection could be achieved by previously converting the — SH groups in proteins to — S — S — links. Their experiments are of a chemical nature and employ tagged sulfur (S^^) in cysteamine (NH2C2H4SH). I shall not attempt to give the details of their experiments but merely to connect their results with ours. The interdependence of the two apparently different types of results has been pointed out in an interesting paper by Ehrenberg and Zimmer (32). Our results indicate that any ionization of a protein which contains S — H groups would always tend to dissociate the — SH group through the migration of the 'hole' or positive charge to the S. Because of the large cross-section of the proteins there would be a large release of H atoms by this mechanism unless there were many competing — S — ^S — links or other traps in the protein to protect the — SH. The experiment of Eldjarn et al. would seem to 'protect' the — SH group by first destroying it! By carrying the hydrogen away peacefully in a harmless molecule they prevent its being released by the irradia- tion as a damaging free radical. Later, after the upheaval is past, it can be restored peacefully if needed. Our results, as well as those of Eldjarn et al., suggest that some agents may exert their protective eff'ects by becoming temporarily attached through a chemical bond to the protein or other thing which they protect. Cystine, glutathione, or other agent which gives up electrons easily is needed for protection against the damaging effects of positive holes. Cysteine, glutathione Electron Spin Resonance in the Study of Radiation Damage 259 in the reduced form, or other — SH agents may exert their protective effects by forming an — S — S — link with an — SH of the protected molecule, as Eldjarn ct al. proved for cysteamine. Electron sinks which collect the electrons knocked out of the holes, and thus prevent them from causing damaging reactions, would also be protective agents. The most desirable electron storage tank would be a molecule which would accept the electron without itself becoming dissociated, would hold it loosely, and would give it up easily when it was needed elsewhere. I should like to add that the electron sources (traps for electron holes) attached to side chains are not necessarily restricted to protective action against direct hits : they may also protect from some of the indirect effects. Certain free radicals produced in the medium around the protein might exert their damage simply by steahng away an electron from some point in the protein. This would of course be replaced by an electron borrowed from the protective group, just as if the electron had been removed by irradiation. The effects of ionized Oo or HoO — if there are such things — would be, I suppose, to ionize the protein when they came near it. An OH radical might react with the protein molecule, or it might simply ionize the protein and form 0H~. I do not know which would happen in a particular case. I simply wish to illustrate a possible unrecognized protective mechanism against indirect action of the radiation. Some specialists on radiation effects evidently believe that the damage to the protein of the cells is due mainly to the indirect effect of radicals produced in the medium around the protein molecules and that the protective action of such agents as cystine or glutathione is entirely the elimination of these radicals before they get to the protein. I do not mean to imply that such effects and the mechanism proposed to protect against them are not very im- portant. What seems clear is that protection is also needed against direct hits as well, and if possible against those radicals or charges produced in the medium which survive long enough to reach the protein. Because of the ability of the protein to transfer a charge, it now seems possible to provide this type of pro- tection too. In fact it has already been achieved in some measure by Eldjarn e/ a/. The protective mechanism which I have proposed is strikingly related to enzyme activity. Pollard's group at Yale, and perhaps others, have been making experiments which show, I beheve, that a single hit in a large enzyme molecule by an ionizing particle is enough to destroy the enzyme activity of that molecule. This is not strange if the sensitive sites for the enzyme activity are synonymous with the sinks or sources for electrons about which we have been talking, and the ability of the enzyme molecule to conduct a hole or excitation is required for enzyme action. I hke to think of the protective agents which are described here as enzymes which prevent reactions. I know, of course, that the normal function of an enzyme is to cause reactions. Some enzymes, so I understand, exert their catalytic action by accepting spare electrons for a time and giving them up again later. In the vivid language of Professor Henry Eyring, they take over the unnecessary children (electrons) during the divorce proceedings and give them back after the remarriages have taken place. One kind of 'protective enzyme' supplies children to prevent divorces (broken bonds) and then later recovers children indistinguishable from those given up (electrons all). Another 260 Walter Gordy kind provides temporary abode for the disrupted children to prevent their disturbing the neighbors. Our Hving systems probably have already built in both types of protective agents in sufficient quantity to provide reasonably good protection from ionizing radiation encountered in normal living of the past. For the future we may need to add some. I have not space to discuss radiation damage to other substances — such as fatty acids, nucleic acids, and hormones — for which our group has obtained many spin resonance data similar to that described here. I have not space to discuss the important effects of oxygen on radiation damage to molecules, about which we have obtained information from spin resonance, of the type shown in Fig. 13. I hope that I have described enough of the results to convince you — and the biologist who rode with me in the car — that microwave electron- spin resonance is an important new way of 'seeing into' biological things. ACKNOWLEDGEMENTS The experimental results I have discussed here were obtained largely by my students — W. B. Ard, H. W. Shields, C. G. McCormick, H. N. Rexroad, A. VAN RoGGEN and C. F. Luck. Others now active in the project are J. H. Hadley, F. W. Patten, and Dr. I. Miyagawa. For their valuable assistance I am grateful. Our research is supported by the Office of Scientific Research, U.S. Air Force, and the Office of Ordnance Research, U.S. Army. Without implying that he is to any degree responsible for any misinterpreta- tions I may have made, I wish to acknowledge with thanks some stimulating and enlightening discussions with Dr. James Franck during the preparation of this paper. REFERENCES 1. E. Zavoisky: Paramagnetic relaxation of liquid solutions for perpendicular fields. /. Phys., Moscow 9, 211-216 (1945); E. Zavoisky: Spin magnetic resonance in the decimetre-wave region. /. Phys., Moscow 10, 197-198 (1946). 2. C.A.Hutchison: Paramagnetic resonance absorption in crystals colored by irradiation. Phys. Rev. 75, 1769-1770 (1949). 3. J. CoMBRissoN, and J. Uebersfeld: Detection de la Resonance Paramagnetique dans Certaines Substances Organiques Irradiees. C.R. Acad. Sci., Paris 238, 1397-1983 (1954); W. Gordy: Quadrupole couplings, dipole moments and the chemical bond. Disc. Faraday Soc. 19, 14-29 (1955). 4. W. Gordy, W. B. Ard, and H. Shields: Microwave spectroscopy of biological substances. I Paramagnetic resonance in x-irradiated amino acids and proteins. Proc. Nat. Acad. Sci., Wash. 41, 983-996 (1955). 5. G. McCormick, and W. Gordy: Electron-spin resonance in x-irradiated peptides. Bull. Amer. Phys. Soc. 1, 200 (1956). 6. W. Gordy, W. B. Ard, and H. Shields: Microwave spectroscopy of biological substances. II Paramagnetic resonance in x-irradiated carboxylic and hydroxy acids. Proc. Nat. Acad. Sci., Wash. 41, 996-1004 (1955). 7. H. Shields, and W. Gordy: Electron-spin resonance of x-irradiated nucleic acids. Bull. Amer. Phys. Soc. 1, 267 (1956). 8. W. Gordy, and H. Shields: Electron-spin resonance in x-irradiated proteins. Bull. Amer. Phys. Soc. 1, 199 (1956). 9. H. N. Rexroad, and W. Gordy : Electron-spin resonance in x-irradiated hormones and vitamins. Bull. Amer. Phys. Soc. 1, 200 (1956). Electron Spin Resonance in the Study of Radiation Damage 261 10. W. GoRDY, W. V. Smith, and R. Trambarulo: Microwave Spectroscopy, J. Wiley and Sons, New York (1953). 11. D. J. E. Ingram: Spectroscopy at Radio and Microwave Frequencies, Butterworth, London (1955). 12. B. Bleaney and K. W. H. Stevens: Paramagnetic Resonance 1. Rep. Progr. Ph)s 16, 108-159(1953). 13. K. D. Bowers and J. Owen: Paramagnetic Resonance II. Rep. Progr. Phys. 18, 304-373 (1955). 14. W. Gordy: Microwave and Radiofrequency Spectroscopy. In: Chemical Applications of Spectroscopy, ed. by W. West, chap. 3, Interscience Publishers, New York (1956). 15. J. E. Wertz: Nuclear and electronic spin resonance. Chem. Rev. 55, 829-955 (1955). 16. W. Gordy: Microwave spectroscopy. In: Handbuch der P/iysi/c, ed. by S. Fliigge, Vol. 28, 1-78, Springer- Verlag, Heidelberg (1957). 17. N. F. Ramsey: Nuclear Moments, J. Wiley and Sons, New York (1953). 18. H. E. White: Introduction to Atomic Spectra, McGraw-Hill, New York (1934); E. U. Condon and G. H. Shortley: Theory of Atomic Spectra, Macmillan, New York (1935). 19. H. Shields and W. Gordy: unpublished data. 20. H. S. Jarrett and G. J. Sloan: Paramagnetic resonance absorption in triphenylmethyl and dimesitylmethyl. J. Chem. Phys. 22, 1783-1784 (1954); G. E. Pake, S. I. Weissman, and J. Townsend: Paramagnetic resonance of free radicals. Disc. Faraday Soc. 19, 147-158 (1955). 21. A. Van Roggen, L. Van Roggen, and W. Gordy: Electron-spin resonance in x- irradiated single crystals of amino acids. Bull. Amer. Phys. Soc. 1, 266-267 (1956). 22. C. G. McCoRMiCK and W. Gordy: Microwave investigations of radiation effects in solids : Methyl and ethyl compounds of Sn, Zn and Hg. /. Amer. Chem. Soc. 78, 3243- 3246 (1956). 23. C. F. Luck and W. Gordy: Effects of x-irradiation upon some substances in solid state: Simple alcohols, amines, amides and mercaptans. /. Amer. Chem. Soc. 78, 3240-3243 (1956). 24. H. N. Rexroad and W. Gordy: Electron spin resonance of radicals produced by ultraviolet irradiation. Bull. Amer. Phys. Soc. 2, 227 (1957). 25. H. W. Shields and W. Gordy: details to be published. 26. B. W. Low: The structure and configuration of amino acids, peptides and proteins. In: The Proteins, ed. by H. Neurath and K. Bailey, Vol. lA, chap. 4, Academic Press, New York (1953). 27. L. Pauling and R. B. Corey: Two hydrogen-bonded spiral configurations of the poly- peptide chain. /. Amer. Chem. Soc. 11, 5349 (1950). 28. J. Franck and R. Platzman: Physical principles underlying photochemical, radiation- chemical, and radiobiological reactions. In: Radiation Biology, ed. by A. Hollaender, Vol. I, Chap. 3 McGraw-Hill, New York (1954). 29. H. W. Shields, W. B. Ard, and W. Gordy: Electron-spin resonance in x-irradiated plastics and synthetic fibers. Bull. Amer. Phys. Soc. 2, 227 (1957). 30. W.M.Dale: Basic radiation biochemistry, Chap. 4; and E. S. G. Barron: The effect of x-rays on systems of biological importance, Chap. 5, In: Radiation Biology, ed. by A. HoUaender, Vol. I, McGraw-Hill, New York (1954). 31. L. Eldjarn, a. Pihl, and B. Shapiro: Cysteamine-Cystamine : On the mechanism for the protective action against ionizing radiation. In: Proceedings of the International Conference on the Peaceful Uses of Atomic Energy, Vol. II, 335-342, United Nations, New York (1956). 32. L. Ehrenberg and K. G. Zimmer: Die Bedeutung der Mikrowellen-spektroskopie fiir die Strahlen-biologie. Hereditas 42, 515-521 (1956). 18 A PHYSICAL MECHANISM FOR THE INACTIVATION OF PROTEINS BY IONIZING RADIATION Robert Platzman and James Franck Department of Physics, Purdue University, Lafayette, Indiana and Research Institutes {Pels Fund), The University of Chicago, Chicago, Illinois Abstract — The extraordinary sensitivity of living systems and certain of their components to ionizing radiation must stem, at least in part, from a great sensitivity of individual molecular or macromolecular species. Analysis of the interaction of such species and swiftly moving charged atomic particles shows that the initial events of energy transfer cannot be responsible for this sensitivity, but that events immediately subsequent to ionization acts definitely can be. This is because the time scale for the production of new electric charges is so short as to evoke a violent reaction of the medium, a reaction which is related intimately to the dielectric be- havior at very great frequencies. Such behavior is not as yet fully explored experimentally, for many of the frequencies concerned lie between the readily accessible infrared and the microwave regions, but the known dielectric properties of highly polar systems like protein and nucleic-acid macromolecules do disclose the existence of regions of strong dielectric absorption, which are to be identified with dipole oscillations and rotations of polar atoms and molecular groupings. The wave of polarization to which a sudden production of electric charge in the interior of the macromolecule gives rise must cause profound degradation of the molecular organization. This may be viewed as resulting from rupture of many weak polar bonds (such as hydrogen bonds) which maintain the intricate organization and which are involved in the above-mentioned dielectric absorption, the ruptures being essentially simul- taneous. The dynamic effect on the molecular structure therefore is without parallel in any other variety of action presently accessible to experimental study — for example, thermal, chemical, or electrochemical action, all of which are in the present context essentially adiabatic in character. The mechanism clearly explains at once the striking difference in sensitivity of the media to ionizing radiation and to ultraviolet light. An approximate quantitative analysis suggests that inactivation of common proteins by a single ionization act is unlikely, but rather that several may be required. Since the effects of the ionizations in a particle track or electron spur are additive (these events being virtually simultaneous), the familiar influence of spatial correlations of ionization is qualitatively explained. The greater radiation sensitivity at elevated temperatures is another obvious consequence. Other predictions of the theory are a dependence of radiation sensitivity upon molecular anisotropy, and a wide variation in the injury to identical molecules exposed to ionizing radiation. I. INTRODUCTION Living systems embody two distinct varieties of intricate organization. One is the complex static structure of the macromolecules which are essential com- ponents of cells; the other is dynamic and is manifested in the delicate organiza- tion whereby the functions of the cell and of the organism are achieved. Living systems are extraordinarily sensitive to ionizing radiation. This is perhaps the most striking single result of experiments in radiobiology and has been emphasized repeatedly (1,2, 3, 4). It is customary, and plausible, to identify 262 A Physical Mechanism for the Inactivation of Proteins by Ionizing Radiation 263 this great sensitivity with a disruption of complex organization, but it is not known which of the two varieties is so highly susceptible. Indeed, both are likely ultimately to be involved. In the case of the functional organization, many proposals have been made concerning the initial point of attack. Thus, the destruction or transformation of sulfur-containing groups, of critical enzymes, and of various other essential constituents present in small or in trace amounts, or the production of powerful poisons, have been implicated. The answer is unlikely to be unique, and from its pursuit, which must involve biological questions of the highest complexity, most of the contributions of radiobiology to the science of biology probably will devolve. The degradation of crucial macromolecules by ionizing radiation is, on the other hand, amenable to in vitro experimentation and to purely physico- chemical theoretical analysis. It is the purpose of this paper to examine the possible explanations for such disruption from a simple physical point of view, and to present and investigate one mechanism which is based realistically upon physical and chemical principles and is in full accord, at least qualitatively, with the results of experiment. A paramount experimental fact is the exceedingly great sensitivity of these macromolecules to ionizing radiation compared to their sensitivity to ultraviolet light. This fact is without parallel in the radiation chemistry of simple organic or inorganic systems, and no clue is provided by the conventional theory of the interaction of swiftly moving charged atomic particles with simple molecules. It is therefore imperative to reanalyze this interaction, with specific regard to the character of the absorbing medium. The primary processes through which the radiation affects the medium cannot differ qualitatively from those in a simple molecular system. In the case of proteins, for example, the very weak bonds which bind the polypeptides together, and even the peptide bonds, must be essentially without influence on the optical dispersive properties of the medium,* and hence the varieties of energy transfer from charged particles, their statistical distribution, and even their spatial dis- tribution will differ only slightly from the corresponding quantities for a simple mixture of amino acids having the same over-all composition. (This statement is not correct with respect to the energy dissipated by moderation of the subex- citation electrons, but that portion of the energy transfer cannot alone be responsible for the great sensitivity.) Many of the events immediately subsequent to primary absorption of the incident radiation, on the other hand, must differ strikingly from those in a simple system. The reason (5) is that the course of such events is determined by the dielectric properties of the medium, and these, in contrast to the behavior at high (optical) frequencies, are profoundly different for a condensed system composed of highly polar molecules. Ionization in nonpolar substances is usually followed more or less quickly by recombination, so that the chemical consequences of absorption of ionizing radiation are very similar to chemical changes induced by ultraviolet light of appropriate frequency. In a system * This is even approximately true for the near-ultraviolet absorption spectrum, which is more sensitive to such bonds than the excitations of greater energy that dominate the phenomena of energy transfer from ionizing radiations. 264 Robert Platzman and James Franck having a great dielectric constant, however, most or all of the initial recombina- tion is inhibited (6), and the chemical effects of ionization stem from interaction with the medium of spatially separated electric charges. This interaction is intimately related to the processes of dielectric absorption characteristic of the medium. It may be noted parenthetically that nonpolar macromolecules, such as many plastics, display utterly different dielectric behavior — in particular showing far less dielectric absorption — and, therefore, that discussions of mechanisms in radiobiology based upon analogies with such polymers, however attractive and despite their current vogue, are perilous, to say the least. II. CONSEQUENCES OF IONIZATION IN A POLAR MEDIUM Before treating the particular case of biologically important macromolecules it will be useful to consider briefly the general consequences of an ionization act in a condensed polar medium. By polar medium is meant one with a high value of the static dielectric constant e^. The condition e^ ^ 1 implies (with only one exception, which is not germane*) the existence of dipolar molecules, and these must always possess regions of strong dielectric absorption at greater frequencies than those at which the dipoles relax. This dielectric absorption embraces all of the familiar infrared absorption, and more: it includes the region from about 30 /^ to 1 mm, which at the present time is not accessible with commercial instruments and is therefore virtually unexplored (although a few investigations of simple molecules have been made, particularly in recent years, and more are now under way). The dispersive properties of proteins, for example, are almost completely unknown in this spectral region. But it is certain that such absorption must be common and intense among complex polar substances, for only those resonances arising from strong bonds and small masses lie in the 1 to 20-/< region ; weaker bonds and greater masses entail absorption at greater wavelengths. The production of an electric charge in a medium will ultimately induce a strong polarization similar to that produced, say, in a condenser filled with the same substance. Tliis polarization will not grow uniformly, however, but rather in several stages, each increase in polarization occurring at a time corresponding in order of magnitude to the reciprocal of its characteristic frequency (7) ; the entire spectrum of dielectric absorption is, of course, involved. The total energy transferred to the medium as the result of an ionization act, excluding the kinetic energy of the ejected electron, can be divided into three parts: that involved in the polarization about the positive ion, a similar quantity released about the electron, or the negative ion which it produces, and finally (and usually much later) the thermal recombination energy of positive and negative ion. The last of these is small (in the case of liquid water, for example, it amounts only to a few per cent of the total) and may be ignored. It merits emphasis that an ionization act per se does not usually cause a mole- cular dissociation process. Analysis of results from mass-spectrographic investigations shows that stable parent ions are the rule, and that the dissociation of ions, which is usually the main subject of such studies, results from additional * This refers to substances, such as certain semiconductors, in which intense electronic absorption at comparatively low frequencies gives rise to a high value of e^. A Physical Mechanism for the Inactivation of Proteins by Ionizing Radiation 265 energy coinmunicated in conjunction with the ionization act in the form oi' electronic excitation energy of the parent ion. In the vapor phase this additional energy may have time to be concentrated in a particular degree of freedom, but in condensed phases it will ordinarily be dissipated by internal conversion and thermal conduction, leaving the parent ion in its ground electronic state, which is usually stable. Hence the simple identification of an ionization act with splitting of a molecule, which is so common in the literature of radiation chemistry and radiobiology, must be viewed with skepticism in so far as con- densed phases are concerned. Such rupture may indeed occur, however, during the growth of polarization about a freshly formed electric charge. It is then very much a consequence of the interaction of the ion with its environment, and in the case of valence-bond breakage imposes special energy requirements with regard to both availability and mode of communication. These require- ments may be satisfied, for example, in instances where dissociation would lead to much greater localization of the charge, and therefore to greater polariza- tion energy. An important property of the energy transferred to the medium by virtue of the growing polarization about a freshly produced electric charge is that most of it is transferred in initial stages, during which the electric field strength is very great. This follows from the Born formula ^E ^ (^2/2/?)[(l/e(r,)) - (l/e(?2))] (1) giving the difference in self-energy of the electric field about a charge of magnitude e outside the sphere of radius R, between instants when the eff'ective dielectric constant has the values ^{t-^ and €{t<^. Since the electronic polarization (oj/-^ 10^^ sec~^) is effective virtually instantaneously, the initial value of e is approximately equal to the square of the ordinary refractive index n, or about 1 .5. By the time that e has increased to (say) 5, most of the total energy (e^l2R){lln^ — 1/eJ will have been dissipated. (Paradoxically, if e^,^ 1 the behavior under discussion is nearly independent of the magnitude of e^.) This argument has the important consequence that a major portion of energy trans- ferred to a polar medium by virtue of an ionization act will be communicated, in a very short time, to degrees of freedom associated with weak polar ('secondary') bonds, and that the region receiving this energy will be considerably more extensive than that affected by a slow change in electric field intensity. The total energy so communicated will be of the order of magnitude of 100 kcal/mole for each (electronic) charge produced, but will depend upon the 'size' of the positive ion, or, in general, upon the structure at a molecular level. If the bonds aflTected are very weak, a substantial fraction of them will be broken, so that the corresponding amount of energy cannot properly be said to have been converted to heat: a portion of it truly has been used to 'melt' a certain structure; obviously, subsequent 'resolidification' often will occur and then will release part or all of the energy as heat. The fate of the ejected electron is similar. It will progress a distance more than sufficient for the developing polarization to inhibit initial recombination (6), and eventually will attach itself to a negative-ion forming group such as OH, if one is available, or to some other entity capable of binding it. The net 266 Robert Platzman and James Franck energy released in such an attachment process is small, apart from the contribu- tion from polarization, and may be positive or negative, since the energy evolved in negative-ion formation (electron affinity) must, if it is substantial, compensate for the energy required to rupture a chemical bond. (This is a consequence of the fact that electron affinities of molecules always are small; the only large electron aflinities are those of certain atoms and radicals, which by their nature must be present in bound states.) Thus the effect on the medium will be very much like that for the positive ion. III. PHYSICAL CONSEQUENCES OF IONIZATION IN PROTEINS The ejection of an electron in an ionization act by even a fairly slow secon- dary electron is an exceedingly quick process and may be considered to have duration of at most 10~^^ sec. The response of a highly polar medium to such an event has been analyzed qualitatively above. After subtraction of the electronic polarization, equation (1) shows that a total amount of energy approximately equal to e^flrfiR will be dissipated during the several subsequent stages of polarization ; here R denotes a distance of the order of magnitude of the mean separation of polar molecular groups, and for proteins must be only slightly greater than atomic dimensions. A value of R of about 2 A thus corresponds to an energy dissipation of 60 kcal/mole. If all of this energy were expended in dissociating secondary bonds, which may be considered to have dissociation energies of approximately 5 kcal/mole, on the average, a rupture of some twelve secondary bonds would be expected. A more detailed analysis for the particular case of proteins, which leads to the same conclusion, will now be presented. Although it is again based upon the Born formula, which applies strictly only to a continuous dielectric, the error caused by neglect of molecular inhomogeneity will not alter the result in order of magnitude. The development of polarization ensuant to electric charge localization in the medium may be divided into four stages. These stages, although distinct in character, are by no means without mutual effects, but such interactions can be disregarded in the present analysis, which is only semi-quantitative. 1. Electronic Polarization — This effect, in contrast to the others, is strongly coupled to the physical processes which lead to localization of the charge, and indeed to the initial ionization act itself (8). Its inffiience on the secondary-bond structure probably is negligible. 2. 'Infrared', or True Vibrational Polarization — The polarization resulting from degrees of freedom corresponding to the characteristic infrared oscillations is developed during a period corresponding to the longest wavelength of such oscillations, or about 3 X 10"^^ sec. With a plausible value of 1.7 for the dielectric constant after this stage of polarization, equation (1) yields an energy dissipation of 1 1 kcal/mole. Thus at most two secondary bonds can be broken, and more probably none are. - 3. Secondary- Bond Polarization— li has already been emphasized that pro- teins and similar substances must possess regions of intense dielectric absorp- tion at frequencies between the accessible infrared and the radiofrequency or even microwave regions. Part, or the whole of this absorption, which has yet to be investigated experimentally, stems from the highly polar A Physical Mechanism for the Inactivation of Proteins by Ionizing Radiation 267 secondary bonds (hydrogen bonds of various sorts, salt linkages, etc.) upon which in large measure maintenance of the structure of the macromolecule depends. The polarization corresponding to this absorption is established during the period of from roughly 10^^ to 10^^ sec following charge localiza- tion. Information concerning the magnitude of the dielectric constant sub- sequent to such polarization apparently is lacking, but fortunately, according to equation (1), the exact value is comparatively unimportant provided that it is much greater than unity, which is certainly the case. (For water it is approxi- mately five.) Thus an energy of about 35 kcal/mole is released during this stage of polarization. It would be erroneous, however, to suppose that this amount comprises the total effect. The Born energy of polarization is the electrostatic energy difference between unpolarized and polarized states (free energy), and this net diminution in energy includes a positive energy of the various degrees of freedom (active coordinates of secondary bonds) equal in magnitude to this net energy. Thus about 35 kcal/mole are dissipated to the medium and 35 kcal/mole reside in the 'bonds' as potential energy resulting from deformation and cleavage. A maximum of about fourteen secondary linkages may therefore be broken. 4. Orientation Polarization — This is the type of process usually considered in studies (9) of dielectric relaxation of proteins (and, regrettably, often imagined to be the only variety of dielectric absorption). It occurs at far greater wave- lengths than the preceding types (e.g. at relaxation times of order of magnitude 10"'^ sec) and is without influence on the secondary bonds. (Thus, these electric waves do not denature proteins, whereas intense irradiation in the 20 to 50-/^ region would very likely do so.) To summarize, energy sufficient to dissociate about sixteen secondary linkages will be released within an extremely short time interval after localization of an electric charge of magnitude e in a protein molecule. Not all of this energy need be used in bond rupture: a portion will be communicated to heat— for example, to quantum oscillations, both primary and secondary, and also to waves of long wavelength. But since the major interaction is with the secondary- bond degrees of freedom, it is likely that the actual number of broken bonds is a substantial fraction of the maximum number. A conservative estimate would be ten. It is obvious that this consequence of ionization* will have a profound influence on the structure. It is now universally believed that, as first proposed by MiRSKY and Pauling (10), the organization of the macromolecules is achieved and sustained by a very great number of secondary bonds, and that the primary-bond structure is identical in native and denatured states. Modern elaborations have refined details while retaining the basic ideas of Mirsky and Pauling. (Thus Lumry and Eyring (11) distinguish between several different arrangements of secondary bonds — for example, in the states of reversible and irreversible denaturation — and propose the useful term conformation changes for these variations.) In some respects denaturation may be viewed as a quasi phase-transition, on a submacroscopic level. Although isolated secondary bonds are continually being opened in random fashion by thermal energy, each bond * The essential idea was described briefly in a previous publication (5). 268 Robert Platzman and James Franck is normally re-formed in the same configuration and the structure maintained by the constraints imposed by neighboring bonds ; only if a number of distur- bances overlap will there be a chance that closure of the bonds occurs in im- proper fashion, and that the disorder becomes irreparable. The model makes possible a satisfactory interpretation of thermodynamic and even kinetic data for thermal denaturation of many proteins (12). It immediately suggests that explanation of the great radiation sensitivity of proteins must be sought in a means of communicating energy from a swiftly moving charged particle to the secondary bonds. Direct energy transfer is negligible, for the coupling is too small (6). The process here advanced provides the required mechanism. Although the analysis given above is admittedly crude (a circumstance for which the lack of relevant and important information on protein structure is chiefly responsible) it is certainly not speculative: it is based upon well established physical principles which are, perhaps, unfamiliar in their present implication. The simultaneous cleavage* of approximately ten secondary bonds following charge localization constitutes a violent perturbation of the protein structure, but probably does not suffice to denature most proteins, at least at ordinary temperatures. This conclusion is suggested by an examination of representative data (Table 1) from analysis of thermal-inactivation kinetics, taken from the Table I. Critical Number of Hydrogen-Bond Ruptures From Thermal Inactivation Rates Molecular ^HX ^S% A'l N^ WlNi weight ( W) Insulin 12,000 35.6 23.8 7 2 1700 Trypsin 24,000 40.2 44.7 8 4 3000 Pepsin 37,000 55.6 113 11 9 3400 Peroxidase (milk) 40,000 185 466 37 39 1100 Ovalbumin 43,000 132 316 26 26 1700 Hemoglobin 68,000 75.6 153 15 13 4500 Yeast invertase 120,000 110 263 22 22 5500 (based chiefly upon reference (12)). A// j is the enthalpy of activation, in kcal/mole, ASJ is the entropy of activation, in cal/mole deg, and the values of A'' are calculated by: A/^i = A// J/5, Ni = ASt/12. work of Stearn (12). Stearn proposed a calculation of N, the number of secondary bonds which have been ruptured in the activated complex (i.e. the critical number for disordering of the conformation), by assuming an average energy requirement of 5 kcal/mole (A^^), or, alternatively, an average entropy incre'ase of 12 cal/mole deg {N^. The values of N^ and N^ so calculated from velocities of thermal denaturation are in impressive accord with one another, * The simultaneity of secondary-bond cleavage, which plays a decisive role in the mecha- nism here proposed, has not been accorded much attention in radiobiology heretofore. It necessarily underlies much of the thinking about mechanisms in thermal denaturation, at least implicitly, and has been invoked, for example, in connection with a model of chemical de- naturation by Kauzmann (13). A Physical Mechanism for the Inactivation of Proteins by Ionizing Radiation 269 except in the case of the smaller proteins, for which the energy and entropy requirements probably do not correspond well with the averages assumed. However, an entropy increase of 12 cal/mole deg is much greater than would be expected for simple cleavage of a secondary bond. This suggests the entirely reasonable conclusion that unfolding occurs when there are broken, not any N secondary bonds, but a particular selection of A^ of them. Clearly, the selection must be a very special one, embracing bonds at certain decisive locations. (Because of cooperative effects this would be true even if all of the bonds were equivalent in their stabilizing action, which is unlikely to be the case.) In the case of secondary-bond rupture following ionization, the bonds aff"ected are more or less localized, and therefore less effective, on the average, than the numbers A'^ listed in the table. Hence the required number of ruptured bonds for ionizing radiation, TV,, must substantially exceed A'^. Since A^ is in the neighborhood of ten for even the smallest enzyme molecules, it is evident that the effect of a single electronic charge is almost, but not quite violent enough to inactivate a typical, small protein macromolecule. Even the combined eff"ects of the positive and negative charges, if they are localized in the same molecule, which must usually be the case, may be expected to be just 'subcritical' (except, possibly, in the case of the smallest molecules). This conclusion leads immediately to the following important consequences. 1 . Variation in the Response of Various Proteins — Because of differences in structural features among proteins of comparable size, the effectiveness of one or two charges may have wide variation. Furthermore, Ni would be expected to increase with the molecular volume, but not necessarily in a simple way. (Note that A^ in Table I shows a definite correlation with molecular weight ( IV), but that W/N is by no means constant.) In all hkelihood NJN would also exhibit interesting differences. 2. Effect of Temperature on Radiation Sensitivity — ^In cases in which a single electric charge (or a pair of them) is subcritical, its effect may be critical at elevated temperatures, because of the augmented probability that the ambient thermal disorder can supplement the radiation eflTect and bring it past the threshold for denaturation. This explains, qualitatively, the pronounced thennal sensitivity which has been observed for some inactivation cross sections (14, 15) and which apparently has not received a satisfactory interpretation here- tofore (14). 3. Effect of Anisotropy on Radiation Sensitivity — ^Anisotropy of the structure (at the microscopic level) may contribute greatly to the radiation sensitivity. An extreme example would be DNA, which is stabilized by numerous secondary bonds having a degree of freedom for oscillation with an almost common direction. Abrupt production of electric charge would rupture many of these bonds simultaneously, causing a portion of the structure to collapse. It is entirely possible that the great radiation sensitivity of DNA, which is found in a variety of experiments (16), may have its origin, at least to some extent, in this effect. The predicted role of molecular anisotropy might be tested experimentally, since proteins and allied molecules exhibit interesting differences in this characteristic. 4. Collective Effect of Individual Activations on Radiation Sensitivity — Although a single electric charge may be insufficient to eff'ect denaturation, a very small number of them would suffice. This points to the importance of 270 Robert Platzman and James Franck the analysis of spatial correlations of charge production* (also called 'spur' or 'cluster' distribution), a subject that has not yet been brought to a quantitative basis, j (It should be emphasized that the bond ruptures caused by positive and negative charges arising from a single spur are all essentially simultaneous.) If accurate information concerning these correlations were available, the path would be open for study of A^^-, and it could be anticipated that both N^ and the ratio NJN would prove helpful in the study of protein structure. A closely related subject is the dependence of denaturation efficiency on the so-called density of ionization; the mechanism predicts that as this parameter increases, the effectiveness first rises (as more of the charges are formed in proximity to one another) but ultimately declines, on an energy basis (as the number of bonds ruptured in a molecule exceeds the minimum number required for unfolding). This is indeed observed in many types of experiment, although the rising or the declining portion of the dependence may be enhanced or suppressed in individual cases, depending upon the specific effect. Such behavior should be distinguished from the corresponding one of simple radia- tion-chemical systems (18, 19), which stem from secondary chemical reactions occurring subsequent to the primary processes; the distinction is not trivial. (In the case of complex biological systems, such as whole cells, the dependence, although it often appears to be similar and may be closely related, must clearly have a far more complex origin (20).) It may be noted that for large protein molecules the disorganization about even a densely ionized track may be insufficiently extensive to produce denaturation. Hence there would be antici- pated some theiTnal sensitivity, although in general less than in the case of sparsely ionizing radiations. Combination of the disorder produced by several localized electric charges is by no means the only possible kind of collective effect, for contributions may be made by excitation events and even by energy transfer to valence-bond and secondary-bond oscillations from subexcitation electrons, both of which must by themselves be minor influences. Changes in certain molecular properties may indeed demand such collective action. For example, permanent dissociation of a valence bond following an excitation act is very unlikely in proteins, but if a dissociative excitation and an ionization should occur close together, the secondary-bond breakage caused by the ionization would prevent heahng of the rupture. Subsequent thermal action would then denature and fragment the molecule. It is a suggestive possibility * Following Lea (1), many investigators have inferred from their experimental data that inactivation is accomplished by a single 'average' primary ionization. This is in rough agree- ment with the general conclusion reached above, but it is not a quantitative statement. Most analyses of experimental data currently being offered appear to be insufficiently detailed and accurate to refine it. t The proposal that Auger cascades may have an important role in the chemical and biological effectiveness of ionizing radiation (17) is highly relevant to the conclusion that a single electronic charge will in general be subcritical, for each cascade must unquestionably result in destruction of the secondary-bond structure on an extensive scale. (One factor is shown by equation (1): the polarization energy is proportional to the square of the electric charge.) In this connection it may be mentioned that the detailed calculations in the paper cited apply only to heavy-particle irradiation; for fast electrons and gamma-rays the yield of Auger cascades is very much greater, being of the order of magnitude of a few per cent of all ionization events, in proteins. A Physical Mechanism for the Inactivation of Proteins by Ionizing Radiation 271 that interchain disulfide bonds may be sensitive regions for such an effect, especially in view of the prominent contribution made by cystine absorption to inactivation of proteins by ultraviolet light (21, 22), but there is no convincing evidence that disulfide-bond cleavage is a major factor in protein denaturation by ionizing radiation. Even the connection with ultraviolet inactivation is ambiguous, because of the difTerent character of excitation produced by charged particles (cf. infra). 5. Ejject of the Environment on Radiation Sensitivity — Since the external environment of the protein can and does participate in the structural stabiliza- tion of the molecule, it may alter the effectiveness of the various possible dis- turbances; the temperature effect already discussed is an instance of this. For example, the medium can contribute externally and internally attached water molecules, various interacting ions, and even chemical influences, and the altered array of secondary bonds may clearly respond differently to the disturbances caused by irradiation. After irradiation and resultant unfolding the imposed forces may impede further unfolding and may, indeed, with the help of thermal agitation, promote healing of the disorganization. On the other hand they may under certain circumstances enhance the radiation sensitivity. This accounts in a general way for pH and other solvent effects. There is in principle no simple way to correlate such solvent influences with their effects in ordinary thermal or biochemical inactivation, since the response to sudden charge localization is completely different in character from that involved in such phenomena. 6. Spectrum of Radiation Injury — The previous considerations show clearly that in a system of identical protein molecules exposed to any variety of ionizing radiation, a broad range of effects on the molecules must occur. This variability has its origin in (a), variations in the disturbance following localization of a single charge, owing to both the intrinsic variability of the effect of the charge at a given position, and to its localization at different possible sites (e.g. in the interior or on the periphery of the molecule); and (b), in variations in the cooperative effects discussed above, which can differ in number, degree, and proximity (extent of overlapping of regions of charge-induced disorder is obviously a cardinal factor). (Thus A'^^ certainly is not unique.) The consequence is a wide range of change in properties, different molecules exhibiting qualitative as well as quantitative differences. This spectrum of radiation injury is manifest when appropriate measures are taken to detect it, and the suspicion arises that the common conclusion from irradiation experiments that proteins are either inactivated or unaffected cannot possibly be general, and may often be an oversimplification or even an artifact imposed either by the conditions of an experiment or its interpretation. That thermal denaturation is not a unique transformation is, of course, elementary knowledge; on the basis of the present analysis it appears likely that radiation denaturation may cover an even broader range. Ample proof of the spectrum of radiation injury is provided by the work of Fricke (23, 24). Thus partial unfolding of the main chains, in addition to denaturation, is indicated by changes in optical rotation, serological response, and A//:J:, and there is evidence for a small amount of fragmentation, with formation of a variety of products of lower molecular weight. The so-called 'after-effect', a diminished thennal stability of irradiated proteins, is simply 272 Robert Platzman and James Franck interpreted as stemming from the portion of the spectrum that is subcritical. (Butler has shown (3, 16) that DNA is also more sensitive to thermal destruc- tion after irradiation.) Fricke has even specifically resolved the thermally labile component into a number of fractions with differing thennal response (23). In the case of ovalbumin irradiated with gamma-rays he found the denatured product to be less degraded than the thermally denatured product; this is as expected, since large-scale unfolding can only occur thermally. Another mani- festation of the spectrum of radiation injury is the differing reaction to post- irradiative environment that is occasionally observed. This phenomenon, of which the after-effect is a special case, is related to the effect of radiative environ- ment, discussed above, but it clearly involves a later phase of the injury — in particular, partial damage will have been stabiUzed by closure of many hydrogen bonds, although often in an incorrect way. (This can be inferred from the very low values of lieats of denaturation, which show that in thermal denaturation most of the bonds do fonn again after unfolding.) Such disordered molecules may be further altered by certain external influences and may be restored, at least in part, by others. It has been remarked (15) that a dependence of the inactivation of irradiated hemoglobin (and other proteins) on the pH of the solvent in which they are dissolved after irradiation is anomalous, but according to the views set forth here such a dependence is not unexpected. In the above discussion the term 'localized electric charge' was used in place of 'ionization act' to denote the center of the polarization wave. The motion of an electron vacancy produced by ionization in a protein has been the subject of much conjecture, but a cogent analysis has yet to be made. Although it is certainly true that in (for example) a simple, isolated organic molecule, the precise designation of an original site of ejection of a valence electron has little meaning, this cannot be taken as proof that an electron vacancy has unlimited ability to migrate along the skeleton of a protein or similar macromolecule. One reason is the low degree of symmetry of the molecule, and its greatly differing regions of potential. Another, which often is overlooked, is the influence of the external polarization on this migration. The electronic part of the polarization sets in almost immediately at ionization, and the various low- frequency varieties follow as discussed above. All of them severely limit the mobility of both positive and negative charges. It therefore appears unlikely that an electron vacancy can cross a secondary linkage, or possibly, indeed, even a peptide bond. In the case that several electron vacancies are produced within the same molecule, whatever motion may be possible must enhance the potency of the effect for disordering, for the coulomb repulsion will tend to separate the final sites of localization, thus preventing diminution in effectiveness by too great confinement. It should be emphasized that the mechanism developed in this paper applies strictly only to the effect of radiation on an isolated macromolecule, a somewhat hypothetical situation approximated in experimental work on 'dry' proteins (1, 4). Immediate effects of the environment have also been touched upon. For proteins in solution, or in living cells, indirect effects, of a simple or complex chemical nature, must also contribute to the observed behavior, and no general- izations regarding the relative potency of the two, except that neither is likely to be negligible, seem warranted. A Physical Mechanism for the Inactivation of Proteins by Ionizing Radiation 273 It may also be noted that the chemical effects may be reversed, but that the disorganization caused by localization in a macromolecule of several freshly created electric charges cannot be; hence protection from or cure of radiation damage at the molecular level cannot possibly be complete, even in principle. IV. THE ROLE OF EXCITATION Absorption of ionizing radiation leading to the formation of a certain number of ion pairs must also produce a comparable number of electronically excited molecules. This is true for the effects of primary charged particles as well as for secondary ones, and is an elementary consequence of electromagnetic theory. Indeed, the ratio of total number of excited to total number of ionized molecules is, except for slow electrons, simply related to familiar optical pro- perties of the absorbing matter, and the available evidence shows that this ratio is unlikely to depart from unity by more than a factor of about 2, even allowing for the disturbing effects of slow electrons. The ratio is known accurately, at present, only for the noble gases, for which it is 0.4. For all molecular systems it must be greater. Just as in the case of ionization, which was discussed above, excitation — whether produced by absorption of ionizing radiation or of ultraviolet light — does not itself 'break bonds'. The initial acts of energy transfer are all* ones in which energy is communicated to the electronic systems of molecules; subsequent rearrangement of atomic positions may then result in dissociation. For polyatomic molecules the probability that bond rupture will follow excitation is by no means unity, and may be quite small. In molecules like amino acids, polypeptides, and proteins, excitation commonly is followed by dissociation or by internal conversion, but only very rarely by luminescence (5). In general, experimental work (which has usually been restricted, for practical reasons, to wavelengths greater than about 2200 A) indicates small quantum yields for inactivation, of order of magnitude 10 "^ to 10"^. Analysis of the absorption processes has not progressed to the stage of identifying them either with dissociation or with internal conversion, but the following explanations for the low efficiency seem attractive. In the case of dissociation, that is, cleavage of a primary (valence) bond, the secondary-bond structure may prove capable of sustaining the conformation, at least until activation energy becomes available for healing the rupture. (Thus the cage effect is enhanced.) This proposal is supported by the fact that dissociation by moderate or long-wavelength ultraviolet radiation does not provide much energy in excess of the bond dissociation energy; thus at 2200 A, not more than several hydrogen bonds could be broken. The structure should therefore remain otherwise intact, with closure of the bond a much more likely ultimate result than denaturation. Internal conversion, on the other hand, releases a substantial quantity of energy to oscillational modes, but the coupling is chiefly with valence oscillations (C — H, C — C, etc.); by the time the energy reaches the secondary bonds it will have been dissipated too extensively to have much effect, t * The only direct bond breakage arises from momentum transfer to atoms from swiftly mo\ing particles, in so-called nuclear collisions (17). This is usually a minor effect. t However, individual internal conversion processes may be responsible for isomerization, and thus for such biological phenomena as gene mutation. 274 Robert Platzman and James Franck Another factor which diminishes the effectiveness of excitation by uhraviolet light is the spatial isolation of the individual absorption events. Thus the second- ary bonds ruptured as a consequence of a single internal conversion process may heal before serious unfolding occurs. In the case of excitation by charged particles, however, the excited molecules are often produced in close proximity (and simultaneity) to other activations, and hence must undoubtedly contribute to the disorganizing action. Such collective effects have already been discussed. One characteristic of ionizing radiation which always should be kept in mind is the difference in nature of the excited molecules produced from those that have been studied photochemically : they correspond, for the most part, to radiation in the vacuum ultraviolet region, where most of the optical transition probability invariably lies. Little is known of polyatomic molecules with regard to optical phenomena and to processes following excitation in this spectral region. However, such radiation may have far greater potency than the readily accessible ultraviolet, for either in dissociation or in internal conversion processes, it always releases sufficient kinetic energy to break many more secondary bonds. There is indeed some experimental indication of this in the rise of quantum yields at the shortest wavelengths studied (25, 22). Thus the role of excitation in radiobiology probably is greater than usually is (cf., e.g., (14)) supposed. V. CONCLUSION The mechanism considered here provides a realistic physical basis for understanding the remarkable fact that a polar macromolecule of molecular weight as great as 10^ can be inactivated by only a few ionization acts, and it is capable of explaining qualitatively a variety of experimental results. It replaces the notion that ionizing radiation acts merely by breaking chemical bonds directly, which, apart from its superficiality, does not actually explain denatura- tion at all. No attempt has been made in the present paper to analyze in detail the myriad data on numerous kinds of radiation effect for varying quality and quantity of radiation, varying environment, etc. Indeed, further development must await, in most points, the further elucidation of protein structure, especially in its dependence upon the secondary-bond configuration. In particular, the number, disposition, and mutual dynamical behavior of the secondary bonds, as well as the character of their large-scale stabilizing action, must be more fully understood. At some future stage of development radiation studies may provide a valuable tool in advancing this knowledge, for the action of ionization, as described here, is completely different in character from other types of attack which are investigated, such as heat, salts and other chemically inert solutes, and chemical agents, all of which act essentially adiabatically at the atomic level. In essence it has been demonstrated that the marvellous stabilizing action manifested in natural polar macromolecules is intrinsically ineffectual against the nonadiabatic disturbance of an ionization act. REFERENCES 1. D. E. Lea: Actions of Radiations on Living Cells, Cambridge University Press, Cambridge, England (1946). 2. L. H. Gray: The initiation and development of cellular damage by ionizing radiations. Brit. J. Radiol. 26, 609-618 (1953). A Physical Mechanism for the Inactivation of Proteins by Ionizing Radiation 275 3. J. A. V. Butler: The action of ionizing radiations on biological materials. Facts and theories. Radiat. Res. 4, 20-32 (1956). 4. E. Pollard: Primary ionization as a test of molecular organization. In: Advances in Biological and Medical Physics, ed. by J. H. Lawrence and C. A. Tobias, 3, 153-189 (1953). 5. J. Franck and R. Platzman: Physical principles underlying photochemical, radiation- chemical, and radiobiological reactions. Radiation Biology, ed. by A. Hollaender, Vol. I, Chap. 3, McGraw-Hill, New York (1954). 6. R. L. Platzman: Energy transfer from secondary electrons to matter. In: Basic Mechanisms in Radiobiology , 11-7,1, Nat. Res. Council (U.S.) Publ. No. 305 (1953). 7. R. Platzman and J. Franck : Theory of the absorption spectra of halide ions in solution. In: L. Farkas Memorial Volume, 21-36, Research Council (Israel), Jerusalem (1952). 8. R. L. Platzman: Influences of details of electronic binding on penetration phenomena. In: Symposium on Radiobiology, ed. by J. J. Nickson, Chap. 9 (cf. §VIII), J.Wiley and Sons, New York (1952). 9. J. T. Edsall: The size, shape and hydration of protein molecules. In: The Proteins, ed. by H. Neurath and K. Bailey, Vol. I-B, 698-712, Academic Press, New York (1953). 10. A. E. Mirsky and L. Pauling: On the structure of native, denatured, and coagulated proteins. Proc. Nat. Acad. Sci., Wash. 22, 439^W7 (1936). 11. R. LuMRY and H. Eyring: Conformation changes of proteins. /. Phys. Chem. 58, 110-120(1954). 12. A. E. Stearn: Kinetics of biological reactions with special reference to enzymic pro- cesses. Advances in Enzymology, ed. by F. F. Nord, 9, 25-74 (1949). 13. W. Kauzmann: Denaturation of proteins and enzymes. In: The Mechanism of Enzyme Action,ed.hy'W. D. McEIroyand B. Glass, 70-110, Johns Hopkins Press, Baltimore (1954). 14. E. C. Pollard, W. R. Guild, F. Hutchinson, and R. B. Setlow: The direct action of ionizing radiation on enzymes and antigens. In: Progress in Biophysics, ed. by J, A. V. Butler and J. P. Randall, 5, 72-108, Pergamon Press, London and New York (1955). 15. R. Setlow: Radiation studies of proteins and enzymes. Ann. N.Y. Acad. Sci. 59, 471^83 (1955). 16. J. A. V. Butler: Effects of x-rays and radiomimetic agents on nucleic acids and nucleo- proteins. In: Ciba Foundation Symposium on Ionizing Radiations and Cell Metabolism, ed. by G. E. W. Wolstenholme and C. M. O'Connor, 59-69, Churchill, London (1956). 17. R. L. Platzman: On the primary processes in radiation chemistry and biology. In: Symposium on Radiobiology, ed. by J. J. Nickson, Chap. 7, J. Wiley and Sons, New York (1952). 18. E.J. Hart, W.J. Ramler, and S. R. Rocklin: Chemical yields of ionizing particles in aque- ous solutions : Eff'ect of energy of protons and deuterons. Radiat. Res. 4, 378-393 (1956). 19. R. H. Schuler and A. O. Allen: Radiation chemistry studies with cyclotron beams of variable energy — yields in aerated ferrous sulfate solution. J. Amer. Chem. Soc. 79, 1565-1572(1957). 20. R. E. Zirkle: The radiobiological importance of linear energy transfer. In: Radiation Biology, ed. by A. Hollaender, Vol. I, Chap. 6, McGraw-Hill, New York (1954). 21. R. Setlow: A relation between cystine content and ultraviolet sensitivity of proteins. Biochim. Biophys. Acta 16, 444 445 (1955). 22. R. Setlow and B. Doyle: The action of monochromatic ultraviolet light on proteins. Biochim. Biophys. Acta 24, 27-41 (1957). 23. H. Fricke: Kinetics of thermal denaturation of x-rayed egg albumin. /. Phys. Chem. 56, 789-795 (1952). 24. H. Fricke, C. Leone, and W. Landmann: The role of molecular degradation in the loss of serological activity in gamma-irradiated ovalbumin. Radiat. Res. 7, 316 (1957). 25. A. D. McLaren: Photochemistry of enzymes, proteins, and viruses. Advances in Enzymology, ed. by F. F. Nord, 9, 75-170 (1949). INFORMATION AND INACTIVATION OF BIOLOGICAL MATERIAL* Harold J. Morowitz Department of Biophysics, Yale University, New Haven, Connecticut Abstract — An analysis of target theory has been carried out in terms of the language of in- formation theory. Certain results suggest that radiation and thermal inactivation experiments can be used to set limits on the values of information content of biological structures. A group of such limits has been discussed, as well as a suggestion for using 'radioactive suicide' experi- ments to evaluate information content. Information theory provides a discipline for quantifying order and specificity in biological structures. Ionizing radiation and heat provide more or less random methods of disordering biological structures. Therefore, we may anticipate that infonnation theory and studies of the biological effects of heat and ionizing radiation may in some way complement each other. In particular, if we can make some quantitative statements about the amount of disordering necessary for loss of biological function, we are then able to say something about how much order is involved in specifying the system. The concept of target volume has an analogue in the representation of a structure in terms of a series of symbols. If inactivation curves are exponential and the target volume is less than the volume of the structure, we may conclude that part of the structure (the critical target) has an information density higher than the rest of the structure. That is, a subset of syinbols in the array require much closer specification than the rest of the array. If no energy is transferred and there are b symbols in the subset, the target volume will be bVjM, where V is the total volume of the structure and M is the total number of symbols of equal volume needed to specify the structure. If there is energy transfer with high efficiency over g atoms, the volume will be of the order of bg^VjM, assuming no overlap of partial volumes. In this paper we shall be concerned with those biological materials that can be dried, taken to low temperatures and then returned to a functional state without appreciable loss of activity. This class of materials includes enzymes, viruses, spores, and transforming principle. In these entities we may conclude that the information is contained in the structure. Several methods have been used to evaluate the information content of these resting systems. We shall outline briefly two methods that have been used to evaluate the infonnation content of biological materials. In both methods it is assumed that the atomic composition and volume are known. The volume may then be divided into a number of elementary atomic volumes. To specify the * Supported by a Research Grant from the John A. Hartford Foundation. 276 Information and Inactivation of Biological Material 277 system completely, we need to state, in some pre-arranged sequence, which type of atom is present in each elementary volume and the bonds between that atom and its six nearest neighbors. Our specification then consists of a message giving the appropriate symbol to each elementary volume. To cal- culate the average information per symbol, we consider the probability /7,y of having the /th type of atom in theyth bonding state. The average information is then given by H=-lPii\o%2Pii (1) a If the/j/s are assumed from the average composition of dry bacteria (hydrogen, 52.2 mole per cent; carbon, 29.9 mole per cent; nitrogen, 7.6 mole per cent; oxygen, 5.8 mole per cent; phosphorus, 2.9 mole per cent; sulfur, 1.6 mole per cent), and if we assume that all the types of bonding configurations have equal a priori probabilities, we can then calculate that H is of the order of 4.0 bits per atom. Since the different bond configurations have rather different probabilities, our figure is high and 3 bits would probably be a more realistic estimate. An alternative but equivalent method of finding the information content is to assume that all states of the system have equal a priori probability. If there are A'^ possible states and L of these are biologically functional, the probability that the system is in a functional state is LjN and the information is given by H = -log2 LIN = log2 TV - log2 L (2) If the system must be completely specified, L equals one and H takes on its maximum value, log2 A'^. We may then calculate A^ from the number of per- mutations of the atoms in the elementary atomic volumes and the permutations of the bond states (1). This leads to the same average information content per atom as the previous treatment. However, from a point of view of biology, we would like to know the actual value of H rather than //,uax- ^^ we consider a large collection of spores or viruses or enzymes in contact with a thermal reservoir at temperature T and allow the system to come to thermal equilibrium over a long time, we may regard the collection as a Gibbsian ensemble, and the ratio of the final activity to the initial activity is a measure of the a priori probability of finding the system in a functional state, in general the activity decreases with time in an exponential fashion. In all the experiments that have been carried out, the sample is too small and thermal equilibrium is never reached. This enables us to determine a lower limit of the information content, but the limit is too low to be of any practical use. For example, dry Bacillus suhtilis spores show an exponential inactivation over twelve decades. There is no indication that the system is nearing equilibrium so we may conclude that the a priori probability of finding the system in a living state is less than 10^^. H is then greater than — log2 10~^^ or greater than 42. Since the upper limit (based on L = 1) for this system is of the order of 10^^ bits, the thermal data do not help very much in bracketing the figure. Experimentally it is not feasible to carry thermal inactivation studies below an activity of 10^^^ because of difficulties in the Id 278 Harold J. Morowitz sample size and the assay in the presence of all the inactivated material. It may be noted in passing that the consideration of the system in terms of a Gibbsian ensemble may provide some insight into the origin of life or the a priori probabihty of a biologically functional structure arising de novo. In considering the information aspects of ionizing radiation, we shall confine ourselves to anhydrous systems and consider only the direct action of radiation. We must then consider the effect of a primary ionization in altering the structure of biological molecules. Present evidence indicates that primary ionizations occur in a random fashion along the track of the fast charged particle. However, the subsequent events are much less clear. It is difficult to make quantitative statements about the probability of the energy being transferred from the site of the original ionization to an energy sink in the material. For purposes of developing the theory, we shall first make the simplest possible assumption that the result of a primary ionization is a bond break, or rearrangement of bonds at the site of the ionization. Many structures are inactivated by a single ionization within the structure. If the previous hypothesis applies, such structures have an information content close to //niax5 sincc L must be unity if any rearrangement destroys the functional integrity of the structure. It should be remembered that //max is the formal upper limit if the calculation is based on atomic specification. It would be possible to start from other points of view, such as monomer specification, functional unit specification, or genetic specification, and arrive at different values of an H function for use in subsequent analysis. However, there are many indications that the simple assumption made above is not valid. For tobacco mosaic virus (2), the target volume is about half the total volume of the particle, yet the infectious unit is presumably the RNA which is only six percent of the total volume. Many enzymes show a target volume equal to the physical volume of the molecule (3), yet recent evidence suggests that several amino acids can be removed from the enzyme without loss of activity (4). It is difficult to see why bond rearrangements in these amino acids should lead to loss of function. Some enzymes show a target volume larger than the physical size of the molecule. These factors indicate energy transfer from the site of the ionization to an energy sink within the molecule. Recent studies by Gordy (5) and Setlow (6) tend to suggest that sulfur-sulfur bonds are the ionization sinks in protein. If we assume that this is the case and that the energy of a primary ionization is transferred with a high efficiency to these bonds, then we can arrive at a minimum value of the infonuation content of molecules which contain these bonds and are inactivated by a single ionization. Since about one in every 400 of the atoms is involved in an S-S bond, we may conclude that MH/400 per atom is a crude estimate of the minimum information necessary to specify the structure of M atoms. In this case radiation experiments would enable us to set a lower limit to the information content. Consider next a structure which requires several ionizations to cause an inactivation. If there are M atoms in the structure, each ionization may trans- fornn the system from its original state to any one of the MB neighboring states, where B is the average number of ways in which each atom can be bonded to its neighbors. If on the average x hits are required to inactivate the structure, Information and Inactivation of Biological Material 279 then L must be at least {MBf H is then given by H = //max — X logo MB + l0g2 x\ (3) It was argued above from equation (I) that there are about 3 bits per atom, so that //,nax is a number of the order of 3A/; thus if x is small in comparison to M, a structure requiring multiple hits will still have an information content close to //max* i^ cascs where no energy transfer is assumed. A 30x10-'° - ~s20xl0'° o - z z o 1- '^ 10x10-'° to to o 5 X 0 - e • , 1 1 1 1 1 200 600 1000 2000 LET IN EV/ 100 A 3000 4000 Fig. 1 There are cases in which x may be an appreciable fraction of M. If one plots cross-section as a function of linear energy transfer (LET), the shape of the curve gives an indication of target thickness and number of ionizations required. If it can be shown that x ionizations are required in a distance A for an inactivation, we can divide the structure into sub-structures of volume A^ in which case H is given by H = ZA^m - 1022 ^^^' (4) '- x\ where m is the number of atoms per unit volume. If x is an appreciable fraction of A^m the substructure may have an information content smaller than the maximum value. The information content of the entire structure, which is the sum of that for the substructures, will be correspondingly smaller than in the case of complete specification. Let us now consider the specific case of the irradiation of spores of Bacillus subtilis with fast charged particles. At all values of LET studied the inactivation curves are exponential functions of the dose. Fig. 1 shows the curve obtained 280 Harold J. Morowitz by Mr. J. Edward Donnellan of the Yale Biophysics Department, and indicates inactivation cross-section for colony formation as a function of LET. From the electron irradiations of Proctor et al. (7), the target volume for these spores is of the order of 10"^'^ cm^. However, this seems to be the volume of the sub- structure with the highest information density. For we see that as we increase the LET the radiation rapidly becomes more efficient in causing inactivation. What we are doing in increasing LET is to increase the probabihty of several ionizations in a given substructure of the spore. Since the cross-section then rises so dramatically, we must conclude that targets of lower information density than the one originally inactivated at low ion densities are now coming into play. Since the curves are exponential, the multiple ionizations in any sub- structure must be coming from the same fast charged particle. Under these circumstances, x must of necessity be small compared with M, and the secondary targets must still retain an information content near //max, if we ignore energy transfer. We may conclude, in general, that any large structure which is capable of being inactivated by the passage of a single fast charged particle through that structure probably has an information content which is an appreciable fraction of /^max- In general, if information can be transferred with high efficiency over g atoms, H is probably greater than H^^aJ^lgf. There seems to be a possibility of reducing energy transfer and thus getting a better estimate of information content. When enzymes are irradiated with fast charged particles and the experiments are carried out at different tempera- tures, the target cross-section is found to be an increasing function of temperature (3). The possibility exists that energy transfer is being reduced at the low temperatures, and data taken in this range might provide a better index of the actual information content. However, considerations of this type demand a thoroughgoing analysis of the physics of the situation. Another method of random disordering exists which might provide an even more powerful tool for the elucidation of information content. It has been recently shown (8) that viruses labeled with P^^ lose activity on standing, and the rate of loss is associated with the amount of P^^ incorporated. Now the decay of a radioactive atom incorporated in a biological structure, and the consequent transmutation of the atom, represents a random disordering. If the labeling is random, the rate of decay should provide a measure of the fraction of atoms of the labeled type which require precise specification in order for the structure to be functional. Such an information evaluation should be possible for phosphorus, sulfur, hydrogen, carbon, sodium and calcium. Thus we may inquire about a complicated structure like a spore: how many of the phosphorus atoms in the structure are required to specify a functional unit? Experiments and calculations of this type should serve to limit the value of the information content of biological structures. REFERENCES 1. H. J. Morowitz: Some order-disorder considerations in living systems. Bull. Math. Biophys. 17, 81-86 (1955). 2. E. C. Pollard and A. E. Dimond: The inactivation of tobacco mosaic virus by ionizing radiation. Phytopathology A6, 214-218 (1956). Information and Inactivation of Biological Material 281 3. E. C. Pollard, W. R. Guild, F. Hutchinson, and R. B. Setlow: The direct action of ionizing radiation on enzymes and antigens. In: Progress in Biophysics, 5, 72-108, Pergamon Press, London and New York (1956). 4. C. B. Anfinsen: The limited digestion of ribonuclease with pepsin. J. Biol. Cliem. 11\, 405^12 (1956). 5. W. GoRDY, W. B. Ard, and H. Shields: Microwave spectroscopy of biological substances. Proc. Nat. Acad. Sci., Wash. 41, 983-1003 (1955). 6. R. B. Setlow: A relation between cystine content and ultraviolet sensitivity of proteins. Biochim. Biophys. Acta 16, 444-445 (1955). 7. B. E. Proctor, S. A. Goldblith, E. M. Oberle, and W. C. Miller, Jr.: Radiosensitivity of 5ac/////5.sM6r///.y under different environmental conditions. Radiat. Res. 3, 295-303 (1955). 8. G. Stent and C. A. Fuerst: Inactivation of bacteriophage by decay of incorporated radioactive phosphorus. /. Gen. Physiol. 38, 444-457 (1955). DISCUSSION Quastler: Dr. Morowitz's analysis of the informational aspects of radiation effects, and his concept of information density, are very important developments. As a matter of fact, I believe them to be so important that even small differences in interpretation are worth mentioning, and this is the reason for making the following comments. To rephrase the situation: consider a structure (message) consisting of a distinct sub- structures (words) of b elements (letters) each. Let H' be the information content per letter and T' the information measure of constraints between letters. Then: H" = information content per word = b{H' — T') H" = information content of message = a(H" — T") (where T" represents the informational aspects of constraints between words) and H^jab = information density in bits per letter. If the 'letters' are atoms in living matter, then I suspect the constraints T' and T" to be quite considerable and to reduce H"'lab to rather less than three bits per atom. We introduce noise of such a character that a single noise event results in the functional destruction of a single letter, and examine the functional value of the message after it has suffered a known number of noise events. If a single event destroys the functional value, then all letters are functionally essential and the functional information density is H"'lab. If the number of noise events needed averages more than one, then the informational density must be less than maximum, and this can occur in two entirely different situations. (I) A number a^ of the letters in each word (or a number b^ of the words) is either irrelevant or can be reconstructed, provided every one of the a — a^ essential letters (or b — h^ essential words) is intact. Then the functional information density is H" a -ao H" b - bo ab a ab b and a single event can cause loss of function but does so only with probability (1 — aja) or (1 — b^jb), respectively. This is the situation where the target size is less than the total size of the structure. (II) Up to «r letters in every word (or up to br words) can be destroyed without loss of function — and these letters or words do not belong to a predetermined sub-set but can be any letters (or words). This is the case with error-correcting codes; in this case no single event can cause loss of function, but Or + 1 events will every time. The functional information density is again less than maximum, being H" a- ar H" b - br ab a ab b 282 Discussion but it is reduced by the presence of redundant information — not by irrelevant information as in the first case. These two situations must be sharply distinguished. With atoms and molecules, the error-correcting mechanism can be a cage effect or some- thing of the kind. This could be the situation where loss of function is caused by clusters of events, i.e. passage of a particle of high linear energy transfer. It may be suspected that the protective effect of redundancy of chemical structure extends only over regions of rather limited sizes which would imply that the reduction of information density could be rather substantial. THE ABSENCE OF RADIATION-INDUCED DISULFIDE INTERCHANGES* Arthur L. Koch f Division of Biological and Medical Research, Argowie National Laboratory, Lenwnt, Illinois, and Department of Biochemistry, University of Florida, Gainesville, Florida Abstract — Mixtures of cystine and its 6/i-dinitrophenyl derivative were: (a) irradiated as dry films, (b) treated in aqueous solution with Fenton's reagent, and (c) irradiated in aqueous solution. In none of these cases could any of the interchange product, mono-dinitrophenyl cystine, be detected. It is therefore inferred that disulfide interchange is not a primary cause of protein denaturation or enzyme inactivation by ionizing radiations. As a consequence of treatment of proteins and protein solutions with large doses of ionizing radiations, denaturation, as assessed by decreased solubility at the isoelectric point, frequently occurs (1). The involvement of sulfur linkages in these solubihty changes as well as in the concomitant loss of biological activity has been frequently suggested. The importance of the oxidation of existing thiol groups to disulfides is well documented (2), but another possi- bility is that polymerization results from disulfide interchange, in a manner similar to that postulated by Huggins, Tapley, and Jensen (3) to account for gel formation in the presence of urea. In their case the initiator of the chain reaction was assumed to be a sulfhydryl group exposed by the unfolding of the protein. This unfolding results from the breaking of intramolecular hydro- gen bonds by the urea. One could, however, conceive of chain reactions initiated by the free radicals produced by ionizing radiations. For example, hydroxyl radicals produced by the action of the radiation on water could react to produce a sulfenic acid and a sulfide radical, which could then react further: OH + /?iSS/?.3 > 7?iS0H + S/?2 Si?2 -r /?3SS/?4 >RS^R^ + S/?3 . Such a mechanism appears attractive in view of the properties of sulfur compounds as presented by Calvin (4, 5). Further reactions of /^^SOH could also lead to polymerization as a consequence of dismutation to the sulfide and sulfinic acid. In order to investigate these possibilities, a model system was studied. The system chosen was that utilized by Ryle and Sanger (6). These authors * This work performed under the auspices of the U.S. Atomic Energy Commission. t Present address: Department of Biochemistry, University of Florida, Gainesville, Florida. 283 284 Arthur L. Koch were particularly interested in the possibility of interchange under strong aqueous acid conditions that are present during protein hydrolysis. In our laboratory this system has been used to study the effects of strong anhydrous acid media (7). In the former case interchange was found and this could be repressed with thiol compounds; in the latter case no interchange occurs. The system consists of a mixture of /?/5-2,4-dinitrophenyl-L-cystine {bis- DNP cystine) and L-cystine. If the interchange occurs, the reaction product, mono-2,4-dinitrophenyl-L-cystine (mono-DNP cystine) may be readily measured by removing the bis compound by acid ether extraction, or by chromatography in a solvent system consisting of aqueous 5 per cent Na2HP04 overlaid with isoamyl alcohol. The spots can be visualized by observation with near ultra- violet hght. By a combination of these two techniques additional sensitivity can be obtained. Dry Irradiation Using this system it was quickly established that even with doses as large as 3 X 10'^ r of Co^° gamma rays no detectable interchange product was produced in the radiation of dry films of mixed cystine and its /?w-dinitrophenyl derivative. As very small amounts could be detected by the combination of the extraction and chromatographic techniques, it is felt that disulfide interchange cannot be of importance in the denaturation of dry protein samples, as certainly much less than one interchange per 1000 disulfide bonds could have been detected. E^ect oj OH Radicals In aqueous solution the experimental situation is quite different. We first investigated the effects of hydroxyl radicals by themselves. Experiments (Table I) with Fenton's reagent (a mixture of Fe++ and H2O2 prepared as described Table I. The Absence of Interchange Produced by OH Radicals The complete system contains 1 x 10^ M cystine; 1.25 x 10-* M 6w-DNP cystine; 1.2 x 10"^ M phosphate buffer, pH 7.9; 5 X 10-5 M pe++ (^s FeSOi); 5 x 10^ M H2O2. For analysis, 0.5 ml aliquots were added to 2.0 ml of 1 N HCl. The solution was then exhaustively extracted with ether and read at 350 m/t in the Beckman spectrophotometer. Complete Minus Fe++ Minus H2O2* Minus Fe++ and H2O2* If interchange is complete (calculated) Increase in optical density 5.5 hours 48 hours 0.048 0.062 0.004 -0.006 0.052 0.052 0.084 0.100 0.489 * The interchange observed at 48 hours in absence of H2O2 is caused by thiol compounds produced by the hydrolysis of the disulfide (6). The Absence of Radiation-Induced Disulfide Interchanges 285 by CoLLiNSON et al. (8)) did not give any increase in the content of non-ether soluble chromophoric material as compared with a control containing HaO, alone. Even at the end of forty-eight hours no increase was apparent, although ribonuclease would have been destroyed completely at the end of thirty minutes (8). The increase in optical density in the HgOg control is small but significant (20 per cent of theoretical at the end of forty-eight hours), and probably represents oxidation to ether insoluble materials such as dialkyl sulfoxides and cysteic acids. Irradiation in Aqueous Solution When aqueous solutions of mixtures of cystine and its Z)/5-dinitrophenyl derivative were irradiated with 1 X 10' r, the results obtained were equivocal because of the influence of side reactions causing change in the chromophoric moiety. Possibly this effect is akin to the well known photo-destruction of dinitro- phenyl derivatives generally. That such a process is occurring follows from the observations that the samples irradiated with 6 X 10*^ r of Co^" gamma rays were less intensely colored than the non-irradiated controls; the optical density at 350 m/^ was reduced to one third that of the controls. The bulk of the 350 mn absorbing material after irradiation was insoluble in ether. This product was clearly not the interchange product, because it possessed the wrong spectrum (high absorption at 250 mfx, with only a shoulder at 380 m//, whereas the wavelength for the maximum absorption of the mono-dinitrophenyl cystine under these conditions is 360 m//). In addition, the same material was produced (in higher yield) in the control containing no cystine. An attempt was made to lower the dose to a point where the above side reaction would not obscure the possible interchange reaction (see Table II). Table II. The Effect o/Co^" Gamma-ray Irradiation ofhis-DNP Cystine The complete system contains 1 x 10"^ M cystine; 1.25 x 10^* M bis-DNP cystine; 1.2 x 10"- M versene buflFer, pH 7.9; where indi- cated, 5 x 10"* M A^-ethyl maleimide (nemi) and 5 x 10"* M/7-chloro mercuri-benzoate (pcmb). The samples were irradiated with 4 x 10* r of Co*" gamma-rays in a 60-minute period. Optical density Before ether After ether Component with extraction extraction Rp = 0.6 Unirradiated 0.510 0.040 0 Complete 0.220 0.171 0 Plus NEMI 0.255 0.198 0 Plus PCMB 0.281 0.209 0 Minus cystine 0.155 0.106 0 After a dose of 4 X 10^ r no colored material could be detected with the Rp of 0.6, which is the R^ of mono-dinitrophenyl cystine, although significant 286 Arthur L. Koch amounts of the reactants were still present. In this experiment the dose was delivered in a one-hour period. During this time the effect of the spontaneous interchange catalyzed by thiols produced by hydrolysis of the disulfides is small (6). In the presence of versene, which prevents metal-catalyzed oxidation by molecular oxygen, the interchange is greater because of greater persistence of thiol. For this reason, controls were included containing thiol binding reagents which completely block thiol catalyzed interchanges. These preliminary experiments indicate that interchange of disulfide bonds is not a prominent feature of radiation-induced denaturation. Further work will be required to assess the role of disulfide linkages in secondary aspects of denaturation. In addition, further work should be carried out using a disulfide interchange indicator that is not itself influenced by irradiation in aqueous solution and thus affords a more sensitive assay for interchange in aqueous media. Acknowledgement — The author acknowledges with gratitude the continued interest and valuable advice of his colleagues Howard S. Ducoff, George A. Sacher, and Antreen Pfau. REFERENCES 1. H. Fricke: The denaturation of proteins by high frequency radiation. Cold Spr. Harb. Symp. Quant. Biol 6, 164-170 (1938). 2. E. S. G. Barron and P. Finkelstein: Studies on the mechanism of action of ionizing radiations — X. Effect of x-rays on some physicochemical properties of proteins. Arch. Biochem. Biophys. 41, 212-232 (1952). 3. C. HuGGiNS, D. F. Tapley, and E. V. Jensen: Sulphydryl-disulphide relationships in the induction of gels in proteins by urea. Nature, Loud. 167, 592-593 (1951). 4. M. Calvin: Mercaptans and disulfides, some physics, chemistry, and speculation. In: Glutathione, ed. by S. Colowick et al, 3-26. Academic Press, New York (1954). 5. J. A. Barltrop, p. M. Hayes, and M. Calvin: The chemistry of 1,2-dithiolane (tri- methylene disulfide) as a model for the primary quantum conversion act in photosynthesis. /. Am2r. Chem. Soc. 76, 4348-4367 (1954). 6. A. P. Ryle and F. Sanger: Bisulphide interchange reactions. Biochem. J. 60, 535-540 (1955). 7. A. L. Koch, W. A. Lamont, and J. J. Katz: The effects of anhydrous strong acid on ribonuclease and lysozyme. Arch. Biochem. Biophys. 63, 106-117 (1956). 8. E. Collinson, F. W. Dainton, and B. Holmes: Inactivation of ribonuclease in dilute aqueous solution. Inactivation by hydroxyl radicals. Nature, Lond. 165, 267-269 (1950). A PROPOSED MECHANISM OF PROTEIN IN ACTIVATION* L. G. AUGENSTINE Brookhaven National Laboratory, Upton, Long Island, New York Abstract— An hypothesis dealing with the role of disulfide bonds in protein inactivation by physical agents has been discussed with reference to material presented at this conference. It is proposed that the critical effect becomes localized at a 'weak-link', causing first the rupture of a disulfide bond, followed by the breaking of neighboring intramolecular bonds and finally the rupture of a second disulfide bond. Much of the evidence upon which these postulates are based is reviewed. The manner in which this model defines a target volume is indicated and alternative methods of disulfide splitting are discussed. The author has previously proposed (I) an hypothesis deahng with the general problems of protein inactivation and the importance of disulfide bonds in maintaining protein structure. This hypothesis was originally presented to account for heat denaturation data. It has since been extended in an attempt to account for inactivation by ultraviolet hght and by ionizing radiation ('direct effect') (2, 3, 4). The model is a special case of more general ones proposed by Mirsky and Pauling (5), Lumry and Eyring (6) and Platzman and Franck (7), and would depend for its accomphshment upon physical processes similar to those described by the latter authors. It is to be emphasized that this scheme is not advanced as the only mechanism whereby protein inactivation can occur, but rather as the most likely. It is proposed that the critical effect of the physical agents mentioned is not to cause indiscriminate molecular disorganization. Rather, their primary effect becomes preferentially locaHzed at certain points in the molecule! . Further, certain of these points (collectively called the 'weak-hnk') are involved in processes which are characteristic of all proteins and which lead to inactivation. These processes can be characterized as occurring in three distinct steps. 1. The breaking of an S — S bond; 2. The breaking of a variable number of neighboring intramolecular bonds (e.g. H-bonds); and 3. The rupture of a second S — S bond. Step 1 requires about 20 kcal/mole, or 0.9 eV per molecule (8), and a negligible entropy factor, while an appreciable entropy increase is associated with step 2. Step 3 allows the spontaneous formation of a structure incom- patible with further activity. Although irreversibility could result from the * Research carried out at Brookhaven National Laboratory under the auspices of the U.S. Atomic Energy Commission. t For instance, Platzman (4, p. 19) has pointed out that 'the most stable position for a migrating electron vacancy to become localized is at a site that can be crudely identified with the atom of lowest ionization potential'. 287 288 L. G. AUGENSTINE formation of a single new bond (4)*, it is probably produced most often by the spontaneous breaking of a large number of intramolecular bonds which is accompanied by a very large increase in entropy. Steps 1 and 2 would constitute the activated state of physical denaturation, i.e., reversible inacti- vation, whereas the rupture of the second S — S bond, step 3, allows irreversible inactivation to proceed. The large entropy change often found to be associated with irreversible denaturation indicates that a partial unfolding of the molecule usually occurs, and therefore the 'weak-link' is probably involved in latching the molecule together. Once the second S — S bond has ruptured, the degree of denaturation will depend both upon the extent to which unfolding proceeds and upon subsequent reactions of the newly exposed groups of the altered molecule. The number and the location of the intramolecular bonds involved in step 2 is thought to be essentially invariable for a given protein, and to depend upon its particular structure in the region of the weak link. It is postulated that a variable number of bonds is involved in step 2 since the enthalpy of denaturation activation varies widely between different proteins (8, 1). These generalizations are consistent with a variety of experimental findings, many of which have been discussed elsewhere (1,2,3), and therefore will be only mentioned briefly. (a) Data reported by a number of investigators give a value for the free energy of denaturation activation of AF* = 24.8 i 1-5 kcal/mole (or 1.1 eV/molecule), which is slightly in excess of that required to rupture an S — S bond (8). (b) Mild denaturation is reversible, whereas violent denaturation is not. (c) Following activation an additional 20 kcal/mole (for trypsin) is necessary to initiate a large entropy change (about four or five times that for A^* of activation). This is thought to be a large configurational change. (d) An average of two to three cysteine equivalents per insuhn molecule corresponds to a fifty per cent reduction in its biological activity (the present hypothesis would predict three cysteine equivalents per molecule, i.e. two for each reversibly inactivated and four for each irreversibly inactivated mole- cule). (e) The appearance of the full protein sulhydryl titer is invariably associated with complete loss in activity. (f) Disulfide bonds are likely involved in the latching together of large segments of the insulin molecule, since reoxidation of the reduced insulin molecule causes aggregation. (g) The ultraviolet action spectra for the inactivation of trypsin and ribo- nuclease, both of which have high cystine contents, are peaked at a wavelength corresponding to maximum cystine absorption, and the quantum efficiency is strongly correlated to their cystine content (9), (10); however, Setlow and Doyle (10) found that the action spectra for gramicidin and aldolase, which had little or no cystine, roughly paralleled the molecular absorption spectra. They concluded that although there must be more than one inactivation mechanism, a quantum absorbed by cystine could be as much as twenty-five times as effective in producing inactivation as one absorbed by an arom.atic ♦ For instance, inactivation due to freezing and drying, which is apparently not accom- panied by a gross opening of the molecule (5), may depend upon such a process. A Proposed Mechanism of Protein Inactivation 289 amino acid, and as much as fifteen times as efTective as a quantum which spht a peptide bond. (h) The electron spin resonance measurements reported by Gordy (11) indicate that irradiation of proteins usually converts some of the disulfides into free radicals. (i) The native configuration of the ribonuclease molecule can be greatly disrupted (as indicated by viscosity measurements) by destroying its H-bonds with urea without destroying its function; however, as soon as the disulfides are oxidized activity is lost immediately (12). (j) The recent findings of Leone (13) are in excellent agreement with the two main aspects of the hypothesis. First, he found that the antigenic properties of y-irradiated serum albumin were the same whether an average of 9 or 90 eV per molecule had been absorbed indicating that irradiation caused this protein to unfold in a characteristic fashion. Second, the single S— S bond of serum albumin was likely involved since ultracentrifugation patterns of the irradiated material contained only monomers, dimers and small amounts of di- and tripeptides, with no evidence of larger aggregates. Setlow and Doyle (10) also noted smaller fragments produced by ultraviolet irradiation of trypsin, but found that the degradation components produced by a wavelength very favorable to a cystine effect were more prominent and homogeneous than those produced by a less favorable wavelength. They interpreted this as evidence for more than one inactivation mechanism. (k) Studies of the inactivation of protein monolayers (2,3,14,15) yielded results which were consistent with the model; however, the data could only disprove but not prove the hypothesis. For example, molecules in compressed monolayers show reduced inactivation from both surface forces (14) and irradiation (15). This was expected, since the scheme proposed here would predict that an external force, such as the monolayer film pressure, should lower the probability of the second S — S bond being ruptured; and even if step 3 occurred, an external pressure should be able to maintain molecular structure sufficiently intact so that restitution would be enhanced. However, it was estimated that the proposed mechanism might account for no more than two-thirds of the inactivation observed. Some proteins, such as ovalbumin (16) and serum albumin, contain fewer than two S — S bonds. In such proteins other bonds which (i) have comparatively small rupture energies and (ii) are involved in latching large segments of the molecule together, would assume the functions of the cystine in this scheme. The present model provides a specification of the 'target volume' for irradia- tion inactivation. Associated with each atom is a probability that energy will be absorbed and migrate to the weak link in amounts sufficient to rupture that structure completely. The sum of these probabilities over the whole molecule gives the 'effective target volume'. Thus, the actual physical target need not have sharp boundaries (see also discussions by Lea (17), Burton (18) and Setlow and Doyle (10) of target elements having probabilities other than 0 or 1). The probabilities, and thus the 'size' of the target volume, will depend upon a number of factors. For instance, the possibility— discussed by Platz- MAN and Franck (7) — of complementary effects between thermal and absorbed 290 L. G. AUGENSTINE energies suggests that the target volume should decrease as the temperature at which protein is irradiated is decreased. Consistent with this is the fact that the x-irradiation cross-section of phage Tl has been found to be a linear function of the irradiating temperature (19); also the inactivation of trypsin ultraviolet-irradiated at 300°K is about three times as great as at 90°K (10). The target volume should also decrease as the quantum of energy absorbed is decreased: the inactivation cross-section of bovine serum albumin bombarded with very low energy electrons was found to increase with increasing ^ energy and a measurable cross-section was obtained with particle energies as low as 10 ev (20). The recent resuUs and interpretations of Yalov/ (21) are particularly per- tinent to the hypothesis discussed here. Her irradiations of insuhn, serum albumin and cystine indicated that disulfide bonds are reduced both under conditions where direct and indirect effects should predominate. However, she proposed that the splitting occurred between the C and S (leaving an S — S — C configuration) rather than between the sulfurs. Although the data cited pre- viously appear to indicate a splitting of the S — S bond, most of those same data would be compatible with a reduction of the C — S bond instead. To select between the two possibilities may be difficult, since the energy required to spht a given bond in a compound such as cystine may be quite different than that required in a protein; as Lumry and Eyring (6) point out, various of the intramolecular bond angles of proteins may be distorted in order to effect structural compromises which minimize free energy. However, Yalow (21) has pointed out that the production of a C — S — S- radical is probably more consistent with Gordy's findings than the other alternative. The failure of Koch (22) to detect radiation-induced disulfide inter- changes either in solution or the dry state does not disprove the hypothesis proposed here. The dosages they used (up to 3 X 10' r) are much larger than those required by other workers (3 X 10^ r) to liberate sulfur groups (21, 23) from similar compounds. These results indicate that although the splitting of disulfide bonds may well be critically involved in protein inactivation, seven per cent or less* of the liberated — SH groups undergo interchange. REFERENCES 1. L. Augenstine: Structural interpretations of denaturation data. In: Information Theory in Biology, ed. by H. Quastler, 119-124, University of Illinois Press, Urbana (1953). 2. L. Augenstine and R.Ray: Trypsin monolayers at the air-water interface. III. Structural postulates on inactivation. /. Phys. Chem. 61, 1385-1388 (1957). 3. L. Augenstine: Trypsin monolayers at the water-air interface. Ph.D. thesis, University of Illinois (1956). 4. E. C. Pollard: The direct effect of radiation on proteins, viruses, and other large mole- cules. In: Biochemical Aspects of Basic Mechanisms in Radiobiology, ed. by Harvey M. Patt, Nat. Res. Council Washington Publ. no. 367, 1-29 (1954). * The \le dose, D*, for a spherical molecule of mol. wt. 240 and density 1.35 is 2 x 10* r. A dose of 3 X 10' r could potentially rupture 1.5 per cent of the S— S bonds (from dNjN = dNjD* = (3 X 107)/(2 X 109)). If one interchange per 1000 disulfides could have been detected (22) but none was found, then seven per cent (1/15) or less of the 'hit' molecules were eventually involved in disulfide interchanges. A Proposed Mechanism of Protein Inactivation 291 5. A. MiRSKY and L. Pauling: On the structure of native, denatured and coagulated proteins. Proc. Nat. Acad. Sci., Wash. 22, 439^W7 (1936). 6. R. LuMRY and H. Eyring: Conformation changes of proteins. J. Phys. Chcm. 58, 110-120(1954). 7. R. Platzman and J. Franck: A physical mechanism for the inactivation of proteins by ionizing radiation. This volume. 8. A. Stearn: Kinetics of biological reactions with special reference to enzymic processes. Advances in Enzymology ed. by F. F. Nord, 9, 25-74 (1949). 9. R. Setlow: A relation between cystine content and uv sensitivity of proteins. Biochim. Biophys. Acta 16, 444-445 (1955). 10. R. Setlow and B. Doyle: The action of monochromatic ultraviolet light on proteins. Biochim. Biophys. Acta 24, 27-41 (1957). 11. W. Gordy: Electron spin resonance in the study of radiation damage. This volume. 12. C. Anfinsen and R. Redfield: Protein structure in relation to function and biosynthesis. Advances in Protein Chemistry 11, 1-100 (1956). 13. C.Leone: personal communication. 14. R. Ray and L. Augenstine: Trypsin monolayers at the air-water interface. I. Film characteristics and the recovery of enzymatic activity. /. Phys. Chem. 60, 1 193-1 199 (1956). 15. L. Augenstine and R. Ray: Trypsin monolayers at the air-water interface. II. Effect of x-ray and uv radiation upon enzymatic activity. J. Phys. Chem. 61, 1380-1384 (1957). 16. G.Tristram: Amino acid composition of purified proteins. Advances in Protein Chemistry 5, 84-148 (1949). 17. D. Lea: Actions of Radiations on Living Cells, Cambridge University Press, Cambridge, England (1946). 18. M. Burton: Elementary Chemical processes in radiobiological reactions. In: Symposium on Radiobiology, ed. by J. J. Nickson, 117-138, J. Wiley and Sons, New York (1952). 19. C. Bachofer, C. F. Ehret, S. Mayer, and E. L. Powers: The influence of temperature upon the inactivation of a bacterial virus by x-rays. Proc. Nat. Acad. Sci., Wash. 39, 744-750 (1953). 20. F. Hutchinson: Energy requirements for the inactivation of bovine serum albumin by radiation. Radiat. Res. 1, 43-52 (1954). 21. R. Yalow: Cleavage of the disulfide bond as a result of the indirect or direct effect of ionizing radiation. National Biophysics Conference, Abstract No. 118 (1957). 22. A.L.Koch: The absence of radiation-induced disulfide interchanges. This volume. 23. W. Dale and J. Davies: The liberation of hydrogen sulfide by x-radiation from cysteine and glutathione. Biochem. J. 48, 129-132 (1951). DISCUSSION Platt: This is a tangential comment which may be relevant to the paper by Professor Gordy and that by Professor Platzman and Professor Franck. It is that Dr. Meyerson* has recently done some extremely beautiful work on species of various mass fragments that break up during their flight time in the mass spectrometer, and by isotopic substitution experiments, checking the species that come out, he has proved that when isopropylbenzene is bombarded with 50-volt electrons, the side group is changed to a cyclopropyl group in which all three carbons are equivalent. He has also proved that if a molecule like toluene is bom- barded, the decay in flight indicates the existence of a species of the form of a tropylium ion, a Cy ion in which again all the carbons are equivalent. Similar results are obtained down to zero excess electron energy. * P. N. Rylander, S. Meyerson, and H. M. Grubb, "I. The Cationated and Cyclo- propane Ring" /. Amer. Chem. Soc. 78, 5799-5802 (1956): Organic ions in the gas phase. II. The tropylium ion. J. Amer. Chem. Soc. 79, 842-846 (1957). See also later papers of this series by same authors in J. Chem. Phys. and /. Phys. Chem. 292 Discussion The result, I think, is that one should probably not speak about traditional valence bonds in traditional directions in a molecular species which has been ionized by ions or radiation. The notion of the classical valence bond is peculiar to a closed shell system, that is, an electroni- cally closed shell with a filled lower energy band and with a large energy gap between the highest filled state and the lowest empty state. I think that it might not be possible to attribute the electrons whose spins Professor Gordy detects to a particular site in the original primitive molecule, because these bonds may have been completely rearranged. Also it may not be possible to attribute Professor Platzman's type of damage to a particular section of the molecule, if this sort of thing, cyclopropyl or tropylium ion, is very common. There may be a large section of the molecule which is doing a merry-go-round of interchangeable carbon atoms before the system settles down. Platzman : I should like to record my skepticism as to the ubiquitous and decisive role of disulfide-bond cleavage in radiobiology. This is not to question the participation of such breakage in denaturation by ionizing radiation or any other agent : since disulfide bonds make an important contribution to the structural stability of many proteins they must certainly be involved in structural breakdown. Dr. Augenstine has, indeed, attempted here to describe the relationship between the contribution of secondary bonds and that of disulfide bonds. However, the argument — frequently heard in recent years — that the great sensitivity of a protein molecule to ionizing radiation can be understood in terms of migration of an electron vacancy produced almost anywhere in the molecule to a disulfide bond, with resultant cleavage of that bond and unfolding of the molecule, is open to most serious objections. In the first place, such long- range migration is unsupported by independent evidence and, as indicated in my paper, is likewise unsupported by physical principles. The fact that electron holes are observed to move freely in certain nonmetallic crystals is of doubtful relevance because of the different dielectric properties of such crystals, and because of their periodic structure. Moreover, even though Professor Gordy's proof of the existence of free valences at sulfur atoms is impressive (although the precise number of such radicals in relation to the radiation dose is still uncertain), a simple causal connection between formation of the radicals and inactivation of the protein has yet to be established, and, indeed, may not exist at all. It is quite possible that they are a secondary factor in denaturation, however conspicuous they may be in the paramagnetic-resonance spectrum. Furthermore, the logical link between disulfide-bond cleavage and electron- vacancy migration is also unproven. A simpler and plausible mechanism for a strong sulfur-atom signal is given by the action of subexcitation electrons : that these can attack the disulfide bonds effectively even though the latter are present in low concentration is strongly suggested by the small dissociation energy of such bonds and also by the marked red displace- ment of the absorption spectrum of cystine in relation to the spectra of most other amino acids. Gordy : I certainly think that Professor Platzman's suggestion is worthy of consideration. I try to keep the sulphur-sulphur bonds in my brain open when discussing these complex systems. PART V AGING AND RADIATION DAMAGE A FEATURE of our timcs is that people are now living long enough so that the prob- lems and diseases of the aged have become an important medical speciality, and at the sam.e time we are, of necessity, embarking on the development of civil and military technology which generates radioactivity, an agent which, uncon- trolled, will contribute to shortening our lives. There is evidence that these two attributes of this age are more than incidentally related. It is well for us to remember that the biological effects of radiation are not new, for the same radiation by which Becquerel discovered radioactivity very soon thereafter burned his person. An understanding of these effects has come slowly. The relationship between aging and radiation damage has been dormant in the literature for a long time and has come to prominence only recently. The first papers on the effect of radiation on life span were published by W. P. Davey (1, 2) in 1917 and 1919. The care exercised in dosimetry and in showing that the observed effect is due to the x-rays and not to some experi- mental artifact was most remarkable for the time at which this work was done. Davey found that the life span of the beetle Tribolhim confusiun was shortened by large amounts of x-rays and lengthened slightly by small amounts. The first result seems to be well established today. The second result is still frequently reported. GowEN and Stadler (3) in 1952 found an increased life span for male Drosophila melanogaster given 2500 r, although the life span of the female was decreased. The effect appeared in Lorentz's data (4) on the LAF^ mouse and inbred guinea pigs receiving 0.1 1 r per day. He did not consider this statistically significant, although Sacher (5) later stated that the effect is significant and that it had been confirmed by himself and D. Grahn. Gowen (6) found a shortening of the life span in male mice from ten distinct inbred strains — even for small single doses of x-rays. However, for female mice he found an increase in life span for doses up to 320 r. But the number of litters produced was reduced even for small doses. The explanation given was that the semi-sterility induced by x-rays reduced the hazards of pregnancy. For low doses, it was argued, this more than compensated for the somatic x-ray damage. It is not generally accepted that there is a stimulation due to x-rays. Probably cases where this seems to occur can be explained as an artifact, perhaps following Gowen's explanation. At any rate further research on this point is well justified. In 1937 Russ and Scott (7) published a report on the biological effects of continuous gamma irradiation. They found the significant features known today, namely, that there is a cumulative permanent damage reflected by a death rate higher than that of the controls, sterility or semi-sterility, high infant and prenatal mortality of progeny from both male and female irradiated parents. They confirmed these results in 1939 and specifically called attention to accelerated aging in the irradiated rats (8). The invention of the nuclear reactor in 1942 added immensely to the industrial and laboratory hazards of radiation and to the concern for evaluating these hazards. Henshaw (9) in 1944 again called attention to the similarity between the pathology of aging and the pathology of radiation damage. Sacher (10) 293 20 294 Hubert P. Yockey in 1950 and Brues and Sacher (11) in 1952 considered radiation injury and lethality and normal aging from the same point of view and gave an analysis in terms of survival curves. In 1953 H. A. Blair (12) emphasized this relationship and extended the notion to internal emitters. He pointed out that the shortening of life, even with bone seekers such as Po, Pu and Ra, is not attributable solely to bone pathology since other tissues are also damaged in a way similar to total body irradiation. Blair's remark is based on observations by Boyd et al. (13) that tissue changes were of the type produced in rats by 550 r whole-body irradiation. In 1954 Furth, Upton, Christenberry, Benedict, and Moshman (14) called attention to this relationship in the case of LAF^ mice exposed to atomic bomb radiation. In the same year Upton, Furth, and Christenberry (15) made the same observa- tion with regard to late effects from thermal-neutron irradiation of RF mice. The similarity between aging and radiation damage is paralleled by chemical carcinogens. Cloudman, Hamilton, Clayton, and Brues (16) reported that mice painted with a carcinogenic agent (methylcholanthrene) exhibited a life shortening not to be explained by a single pathology. They indicated an analogy between life shortening from hydrocarbons and total-body irradiation. Russell (17) has recently found that the increased prenatal and infant mortality of offspring from irradiated parents continues throughout life and is reflected by a reduced average life. He studied only the offspring of male mice irradiated by neutrons from an atomic weapon. Presumably, the effect is general and applies to offspring of both male and female animals subjected to any ionizing radiation. The relation of this work to that of Russ and Scott is clear and the need for detailed study is paramount. Some may feel that establishing a relation between two unexplained effects gets one nowhere. However, as Platt (18) points out, such a relationship may help one effect to explain the other. The concept of premature aging as a measure of damage from various deleterious agents seems to be well enough established for practical use in understanding the nature of biological damage. Information theory may well have a contribution to make to the elucidation of these problems of our times which are so important from so many points of view. H. P. Y. REFERENCES 1. W. P. Davey: The effect of x-rays on the length of life of Tribolium confusiim. Gen. Elect. Rev. 20, 174-182 (1917). (An article nearly identical to this is to be found as follows: W. P. Davey: The effect of x-rays on the length of life of Tribolium confusum. J. Exp. Zool. 22, 573-592 (1917).) 2. W. P. Davey: Prolongation of life of Tribolium confusum apparently due to small doses of x-rays. Gen. Elect. Rev. 22, 479^83 (1919). 3. J. W. GowEN and J. Stadler: Irradiation effects on viability of Drosophila melanogaster. Genetics 31, 586-587 (1952). 4. E. Lorentz: Effects of long-continued total-body gamma irradiation on mice, guinea pigs, and rabbits. In: Biological Effects of External X and Gamma Radiation, Il(i-1A1, ed. by R. E. Zirkle, National Nuclear Energy Series IV-22B, McGraw-Hill, New York (1954). 5. G. A. Sacher: On the statistical nature of mortality, with especial references to chronic radiation mortality. Radiology 67, 250-258 (1956). 6. J. W. Gowen: private communication (1957). 7. S. Russ and G. M. Scott: Some biological effects of continuous gamma irradiation, with a note on protection. Brit. J. Radiol. 10, 619-629 (1937). Aging and Radiation Damage 295 8. S. Russ and G. M. Scott: Biological eflFccts of gamma irradiation. Bril. J. Radiol. 12, 440-441 (1939). 9. P. S. Hensiiaw: Experimental roentgen injury. iV Effects of repeated small doses of x-rays on blood picture, tissue morphology and life span in mice. /. Nat. Cancer Inst. 4,513-522(1944). 10. G. A. Sacher: The survival of mice under duration-of-life exposure to X-rays at various dose rates. Argonne Nat. Lab. Report No. ch-3900 (1950). 11. A. M. Brues and G. A. Sacher: Analysis of mammalian radiation injury and lethality. In : Symposium on Radiobiology, 441-465, ed. by J. J. Nickson, J. Wiley and Sons, New York (1952). 12. H. A. Blair: The shortening of life span by injected radium, polonium and plutonium. University of Rochester Atomic Energy Project, Document No. UR-274 (1953). 13. G. A. Boyd, H. E. Silberstein, R. M. Fink, A. Frenkel, W. L. Minto, R. G. Metcalf, G. Casarett, and G. M. Suter: Pilot studies on the intravenous lethal dosage of polonium, plutonium, and radium in rats. In: Biological Studies with Polonium, Radium and Plutonium, ed. by R. M. Fink, Chap. 7, National Nuclear Energy Series vi.-3. McGraw-Hill, New York (1950). 14. J. Furth, a. C. Upton, K. W. Christenberry, W. H. Benedict, and J. Moshman: Some late effects in mice of ionizing radiation from an experimental nuclear detonation. Radiology 63, 562-570 (1954). 15. A. C. Upton, J. Furth, and K. W. Christenberry: Late effects of thermal neutron irradiation in mice. /. Cancer Res. 14, 682-690 (1954). 16. A. M. Cloudman, K. A. Hamilton, R. S. Clayton, and A. M. Brues: Effects of combined local treatment with radioactive and chemical carcinogens. J. Nat. Cancer Inst. 15, 1077-1083 (1955). 17. W. L. Russell: Shortening of life in the offspring of male mice exposed to neutron radiation from an atomic bomb. Proc. Nat. Acad. Sci., Wash. 43, 324-329 (1957). 18. J. R. Platt: Functional geometry and the determination of pattern in mosaic receptors. This volume. A STUDY OF AGING, THERMAL KILLING, AND RADIATION DAMAGE BY INFORMATION THEORY Hubert P. Yockey Health Physics Division, Oak Ridge National Laboratory, Oak Ridge, Tennessee Abstract — The information theoretic formahsm developed in the author's paper in Part I has been applied to the calculation of survival curves. The results have been compared for a variety of organisms ranging from viruses to mammals. The deleterious agent varied in the magnitude of its quantum energy from thermal and chemical energies to several Mev. It should be emphasized that this article says very Uttle about models. In general the accepted model of the organism is taken and its behavior is calculated from those features of the model which have the aspects of a communication system. In spite of the complexity and variety of the organisms and the range of energy in the lethal agent and without ad hoc assumptions pertaining to models, we were able to account for the main features and some details of the survivorship curves. Many additional experiments will be suggested; some of them have been pointed out and some predictions have been made. The idea of the storage and transfer of information and its destruction by deleterious agents seems to link the material discussed. Much of this material may otherwise seem unrelated or only vaguely so. I. INTRODUCTION The study of the survivorship curve has contributed to the quantification of some essential but otherwise quahtative notions in biology. The effect of various insults such as ionizing radiation, ultraviolet light, temperature, disease, chemicals, and so forth, is very often measured by survivorship of a suitable test organism. On the other hand, the experimenter may be interested in the survival response of a particular organism as a function of maturation, nutrition, strain difference, or the like, and may use some convenient agent as a test stimulus. Survivorship does not contain all we feel intuitively is involved in the concepts of 'vigor' or 'fitness' but it does contain much of what can be defined and measured in an unequivocal and operational way that is associated with those ideas. These facts, together with the application in evaluating quantitatively hazards to man, make this subject one of great practical and theoretical interest. Information theory is peculiarly well qualified to provide a mathematical treatment of these matters. The survivorship curve is a property of the ensemble of organisms rather than of the individual. It reflects the generalized decay of the organization of a system. The central thesis of this paper is that aging, thennal killing, and radiation damage reflect essentially the same action, namely, the destruction of the information content of the cell. The ideas discussed in the author's previous article in this volume will be applied to the calculation of 297 298 Hubert P. Yockey hapJoid survivorship, diploid survivorship, and to the role of equivocation in the germ line. II. SURVIVORSHIP FOR THE HAPLOID CASE Otto Rahn (1,2) was the first to suggest in two pioneer papers published in 1929 and 1930 that the genetic structure is the sensitive element in the cell for radiation damage, thermal killing, and the action of some chemicals. He later reviewed the data on disinfectant action and confirmed this opinion (3). Lea (4) was a strong supporter of this idea and used it in his development of the target theory. This is generally an accepted view today (5) although the role of gene mutations and chromosome aberrations is a matter of debate (6, 7). In a previous article in this volume we showed that this notion follows directly from the application of information theory to the current conception of the storage and transfer of genetical information in the cell and the synthesis of proteins. It is therefore of interest to continue the argument and attempt to calculate survivorship curves. We also showed that error will exist in the genetical information of all real organisms. The organism will live and multiply according to Dancoff's principle (8), in spite of these errors. We argued previously that there must be a distribution of message entropy values among the elements of the ensemble of organisms. Suppose that the number of errors in the genetical infoiTnation is increased as a result of, say, radiation. Those elements of the ensemble near the lethal limit will succumb even though they were quite viable before irradiation. This is a notion peculiar to information theory. The communications analogy is Shannon's channel capacity theorem (9). This theorem shows that if a channel has a capacity C, it is possible, by proper coding, to send information at rate C or less through the channel with as small a frequency of errors as desired. Thus, though the noise level in the channel will affect the channel capacity C, it will not prevent nearly perfect transmission of information. This can be assured by proper coding. As long as this limit C is not exceeded, it is impossible for the recipient to know the noise level in the channel or information source. With these points in mind, we now return to the suggestions of Rahn and Lea, keeping further in mind the idea of Watson and Crick that a mutation is a change in the order of nucleotide bases in DNA (or some other information- bearing molecule). We have proposed that the action of radiation or other deleterious agent at the molecular level is such that the nucleotide pair mimes some other nucleotide pair insofar as protein synthesis and replication are concerned (10). The action of radiation may therefore be thought of as causing lethality through gene mutation by decreasing the message entropy of some members of the ensemble below the lethal limit. This is essentially the suggestion of Rahn and of Lea phrased in the language of information theory, and it follows from the argument given in the previous article in this volume. On this basis we may proceed to calculate the force of mortality on the ensemble. The distribution of message entropy in the ensemble will be represented by a probability distribution p(H, A), where A is a measure of the magnitude of the deleterious agent and the initial distribution is p{H, 0). This distribution will vary with the genetic character of the ensemble of organisms. It can probably be derived from first principles, at least for simple cases, when more A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 299 is known about storage and transfer of information in organisms. For the present it will be necessary to make some simpje assumptions, however. It was proposed previously (10) that death occurs when the value of //decays below some limit H^. Let / be the number of organisms in the population representing the ensemble. The probability per unit I of leaving the population {\jl){dHdX) is called the force of mortality. The force of mortality will be the probability density unit per H at //,, after exposure X multiplied by the rate of decrease of H per unit X at //,^. \_dl I dl p{Ha, A) dH ~di (1) ffrf The value of p(//, A) varies continuously with A; no organisms leave the microensemble which at A = 0 lies between H and U + dH. dX = p{H„ X) cm dX dX (2) H. The relation of p{H^, X) to p{H, 0) is as follows: piH„ X) = p(H, 0) i.) (3) ^o + ^. + iIp(0 logo ;>:(/) »j where p'i(j) is the value corresponding to H^. Equation (1) may be written p(Ha, X) ffo~-ffa + iIp(i)\og,Pi(j) 1,3 J{X) dX (4) In many cases the action of the deleterious agent will be of the first order so that J{X) = /„, a constant. Let us assume that p{H, 0) is of such shape that p{H^i, X) is a constant. Equation (4) may be integrated : log, ///o = p{H„ X) Ho-H, I lp(i) ^0^2 Pi(j) J,X (5) Equation (5) represents haploid survivorship as a function of X for many types of destructive influences under many experimental conditions — but not for all influences, or conditions, or haploid organisms. T. Alper (11) found the rate of inactivation by gamma rays of dysentery phage SI 3 to increase with increasing dose at 130 rad/min. At 5.3 rad/min the survival curves departed markedly from the exponential forni although that form was found when catalase was present. Watson (12) also reported the same phenomenon with phage T2. Alper (13) later showed that the gas treatment of phage could result in departure from the exponential form. A number of cases of a non-exponential inactivation curve for viruses are discussed by Luria in a recent review (6). Gates (14) discussed the deviations from an exponential curve for ultra- violet irradiations of bacteria. Recent work by Uretz (15) has shown that 300 Hubert P. Yockey inactivatioii of haploid yeast is exponential for x-rays and sigmoid for ultra- violet. Anderson (16) has irradiated two biochemical mutants of strain B of E. coli with x-rays, namely, the streptomycin dependent strain and the purine- less strain. An exponential survival curve is obtained in oxygen while a sigmoid curve is obtained in nitrogen for each strain. Hollaender, Baker, and Anderson (17) have discussed the effect of oxygen and other chemicals on the x-ray sensitivity for mutation production and survival. Hollaender and Stapleton (18) have shown that many types of survival curves may be obtained ranging from exponential to a very pronounced sigmoid shape depending on the experimental conditions. Stapleton, Sbarra, and Hollaender (7, 19) have studied the nutritional aspects of survival of bacteria from ionizing radiation. They showed that the B/r strain of £". co// grown on a complete medium such as nutrient broth exhibited radiation-induced requirements for nutritional factors. They presented some evidence showing that such bacteria are not stable auxotrophic mutants. The dependence of survivorship on nutritional factors is explained by ZiRKLE and Tobias (20) from hit-theory concepts. They state that : 'Accordingly, the number n of essential sites in the haploid chromosome set might vary with the composition of the medium; in general, one would expect that the richer the medium the fewer would be the observed number of essential sites. On the other hand, if the 'inactivation' of a 'site' is not a mutation, but a gross change in chromosome state or configuration, the number of sites would be independent of the composition of the medium.' The interpretation of these results given by the authors quoted loses some force since essentially the same results are found for viruses by Friedewald and Anderson (21), iDy Luria and Exner (22), and by Dale (23). The explana- tion offered by Luria and Exner is based on a two-fold action of the radiation, a direct and an indirect effect. The case where two deleterious influences operate simultaneously is interest- ing. Wood (24, 25) has studied the x-ray survival of haploid yeast as a function of temperature. The curves show a distinct tendency to be concave downward for temperatures between 45°C and 55°C. He finds a 'softening' or 'memory' of exposure to temperature and x-rays for the action of the other. Uretz (15) finds very little 'softening' in his study of the action of x-rays and ultraviolet on haploid yeast. We are not aware of a study of the ultraviolet survival as a function of temperature, although such data would be of importance to complete knowledge of these effects. Gray (26) has pointed out recently that a view is gaining general acceptance that a site may be inactivated by a single fast electron, but not by the absorption of a single photon. The site mentioned by Gray is interpreted as a nucleotide pair in the present paper. The action of the deleterious effect may be, partly at least, to throw the nucleotide pair into an excited tautomeric form. In such a form it may be more easily damaged by a successive interaction. The extent to which this occurs may very well depend on the chemical environment. At any rate, for the present purposes, there is reason to believe that J{X) may be represented by a polynomial in A. The higher order terms represent higher order reactions. In that event, it is possible to begin to understand why an exponential A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 301 survival is obtained under some conditions, but not under others. The shape of the survival ciuve depends on both the environment and on the genetic character of the organism. Thus, it may be possible to obtain some separation between purely biological and purely physical or chemical phenomena — in so far as such a separation has meaning — by means of the present theory. The function p{H, 0) is related to the distribution of genetical information in the ensemble and so is characteristic of the biology of these problems. J{X) represents the interaction of radiation and matter, and will be determined by the physics and chemistry of the situation. It is in this regard that the current controversy on the role of direct and indirect action bears on the present theory. Not all haploid organisms exhibit the exponential survivorship curve. Nybom (27) reported sigmoid x-ray survival curves for the green algae, Chlamy- (hmonas euganietos, C. moewusii, and C. reinhardi. Genetic experiments involving tetrad analysis indicate that the haploid character of these organisms is reasonably certain. Jacobson (28) has studied C. reinhardi in some detail and finds a sigmoid survival curve. The mathematical nature of this curve is such that it does not correspond with target theory calculations. This and other cases of sigmoid survival curves in haploid organisms will be discussed in the next section. III. SURVIVORSHIP FOR DIPLOID Lea (4) rejected the gene mutation suggestion for 'killing of organisms other than bacteria or viruses' in favor of the view that chromosome aberrations are the main cause for lethal effects in polyploid tissue. He did this partly on the ground that a 'recessive lethal mutation in a diploid cell will not be lethal unless it is in the X chromosome of the male, owing to the presence of a normal allelomorph in the same cell'. ZiRKLE and Tobias (20) retained the recessive lethal mutation hypothesis in their study of x-ray survival curves in yeast. It was shown by Tobias and Stepka (29) that irradiated diploid yeast exhibits an inheritable increase in radiosensitivity presumably because of an increased load of recessive lethal mutations. Mortimer and Tobias (30) obtained direct experimental evidence for x-ray induced recessive lethal mutations by demonstrating a reduction in the fraction of genninating spores produced by x-ray exposed diploid yeast cells. Mortimer (31) obtained further evidence for the existence of recessive lethal mutations in studies of the conjugation of yeast cells of opposite mating type. See results shown in Fig. 1. Mortimer argued, as Lea had, that the viability of zygotes should be unaltered because of the presence of the normal allelomorph and that therefore recessive lethal mutations could not be respon- sible for all the radiation damage. Chromosome aberrations clearly represent an increase in the equivocation in the genetic information. It is difficult to see how this is to be calculated at the present writing. It is unclear also what their role is in insults milder than damage due to ionizing radiation, such as thermal killing and aging. Sacher (32) has called attention to the need for cytological investigation of the part chromosome aberrations play in the development of late effects from radiation damage. Russell (33) has argued that chromosomal aberrations probably have little to do with radiation hazards to man. At any rate, one may argue that 302 Hubert P. Yockey recessive lethal gene mutations play an important role in the lethality of diploid cells. Since it is possible to present a calculation of the equivocation due to this process let us calculate the survivorship curve according to this notion and see how it compares with experiment. 100 50000 X-RAY DOSE.ROENTGENS 75000 100000 Fig. 1 . Percent survival for yeast with one irradiated parent. Haploid x haploid cross (oo), haploid x diploid cross (oO), etc. The first symbol represents a cell of the a-mating type, the second one of ^-mating type. A filled letter o designates the irradiated parent. Haploid dominant lethal curve: Q, 9o; %,o% ; 6, •©; 9,0*. Diploid dominant lethal curve: n^ •O' ■>0*; D, •©; Ho*. Diploid survival curve: A' • — aa diploid, ^, 9 — aa diploid. Haploid survival curve: A> • — a haploid; ^,9 — a haploid. (From ref. (61)). The decay of the correct read-off probability is given by equation (10) of my paper in Part I : ^^P.(j) = -J(^)Pi(j) + im (6) In the diploid case J(X) cannot contain a constant tenn because of the protection afforded by the unaffected allelomorph. Therefore dp/ij)ldX must depend at least linearly on ?^. The polynomial for J{X) is in this case, where J^ is a constant: m = /i^ (7) Substitute this function in equation (4) dJ 1 JX = P^^"^ ^^ J,?. (8) A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 303 The assumption of Section II concerning the nature of pC//^, A) is retained and the expression for the surxiva! curve is: log. Ilk = P(^.^ ^) //o-^. + i>;/Hoiog2/'ay) ^,} (9) That some survivorship curves have wholly or substantially the form of equation (9) for normal aging where A is the time can be shown for a wide variety of organisms. Some examples are shown in Fig. 2 and also one in II 27 43 51 59 67 75 83 91 WEEKS • (TIME UNIT ) AS INDICATED Fig. 2. Survivorship curves for two insects and two mammals plotted against the square of the age (see equation (9)). Straight lines for pure vestigial Drosophila (35) obtained from maximum likelihood estimate. Other straight lines are for comparison only. (See text.) Fig. 3. The data for pure vestigial Drosophila are from Pearl and Parker (35) and represent a life table. The animals are kept under ideal conditions and the number which die in certain time intervals is recorded. All data on Fig. 2 and Fig. 3 are obtained this way. The curve has been fitted by these authors to a function of the following form where a, h, c, d, e are positive constants. log / = €"^6 - cA + d?? - e)?) (10) Pearl and Parker were aware that this description involves too many con- stants and that irrelevant statistical fluctuations are preserved by equation (10). 304 Hubert P. Yockey Leslie and Ranson (36) fitted their data on the vole to a function of the form of equation (9). They give a conventional yj- analysis to justify the hypothesis. The y^ analysis cannot be applied to data obtained as these life tables are obtained since the points are not statistically independent. The random variable is the time of death of each animal, not the number alive at 100 FLIES (days) 25 35 42 50 55 60 65 70 74 78 62 MICE (months) Fig. 3. Normal aging survivorship of certain strains of mice (52, 53) and Droso- phila melanogaster (35). Note that the abscissa is plotted as age squared to show that the dilute brown strain follows equation (9) but that others do not. a given interval of time. We are unable to justify our hypothesis in any objective mathematical way. The standard errors given are calculated as follows where /, is the number in the / interval and /j_,_i the number in the / + 1 interval (10 The points in Fig. 2 lie very near to a straight line and so it is plausible, at least, that equation (9) represents the normal aging survivorship for some organisms. If the destruction of genetical information is the feature common to the action of the deleterious agents discussed in this article then the survivorship curve should be relatively insensitive to the character of the agent except insofar as reflected by the fonn of J(A). That such is indeed the case is shown A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 305 by the results obtained for x-ray and thermal kilhng of diploid yeast and other organisms. T. H. Wood (24, 25, 39) has reported data on the x-ray survival and thermal killing of yeast, since repeated and verified by Uretz (15). The data for diploid yeast are given in Fig. 4 and A is the time at the indicated temperature or the MINUTES AT TEMP 10 15 20 0.002 0.001 60 90 120 180 MINUTES AT 425 r PER MINUTE 240 Fig. 4. Thermal and radiation killing of yeast from Wood (25, 39). Note that the abscissa is plotted as the square of time at temperature indicated or at 425 r per min. Straight lines drawn for comparison only. X-ray dose as the case may be. The data have been fitted, using Kimball's method (40), to several curves and the results are shown in Table I. The function log ///q = n log [2e -M — e -2k„X ■'' ] was derived from hit-theory by LuRiA and Dulbecco (41) and used by Zirkle and Tobias (20) and by Wood (39). We have fitted Wood's data allowing k^^ to be determined internally by the diploid data and obtain the values shown. When the haploid value nkf^ = 2.49 X 10 * r^^ is imposed on the diploid data n = 21.3; x^ = 5.1; P > 0.5. Wood obtains graphically the slightly different values shown in Table I. It is clear from Fig. 4 and Table I that the term in ?? represents sub- stantially, if not completely, the behavior of the survival curve as a function of dose. The fact that the fit may be made satisfactory by including a small term in either /.^ or A* supports this conclusion. 306 Hubert P. Yockey Wood's survivorship curves for thermal killing in yeast (25) are also given in Fig. 4. The term in X^ again substantially represents the behavior of the survivorship curve. The curve retains its form when the temperature is changed. The -f- test shows in each case a very poor fit to A" but this presumably reflects Table I. Goodness of Fit for Wood's X-rav Survival of Diploid Yeast Function Constants r- p* log ///o = -a^' a = 0.022 60. <0.001 log ///o = ~aX' + bX^ a = 0.029; b = 7.2 X lO--" 5.5 ^0.5 log ///„ = -flP + bX^ 0 = 0.025; * = 3.2 X 10-^ 8.3 ^0.2 log///o = /;]og[2t'-M - e-2M] n = 26.3; //Atj = 2.20 X 10-* 1-1 (this paper) n = 30; nka = 2.41 >: 10^* r-i (from Wood (39)) 3.9 >0.5 * 7 degrees of freedom the existence of small higher order terms as in the x-ray case. Attention is also called to the aging of the grain beetle Calandra oryzae at 32.3°C and at 29. TC shown in Fig. 2. The survivorship curve again retains the same shape, changing only the coefficient of A^. The sensitization to thermal killing of Paramecium caudatum following x-irradiation was first reported by Giese and Heath (42). They found a slow recovery eff'ect, requiring several days. This parallels the earlier discovery of Giese and Grossman (43) of sensitization to thermal killing by ultraviolet and by visible light in the presence of photodynamic dyes (44). Baldwin (45) has pointed out the similarities between thermal killing and killing by x-rays for the hymenopterous insect Dahlbominus fuscipennis. The immediate consequence of both insults is a coma from which the insect may recover to die later of delayed eflFects. Aging decreases the tolerance for both temperature and x-rays. The dose-survivorship curve is not given accurately but it has roughly the same shape for each agent. The diploid females are more resistant than the haploid males. These observations parallel those of Wood on yeast. It was mentioned briefly in the section on haploid organisms that Nybom (27) has reported sigmoid x-ray survival curves for three species of green algae, Chlamydomonas eugametas, C. Moewusii, and C. rcinhardi. Jacobson (28) has studied C reinhardi in some detail and shows that the x-ray survivorship curve fits accurately an equation of the foitn of equation (9). He points out that this can 'be explained by a redundancy of genetic information.' Clark and Herr (46) irradiated the haploid male and diploid female of Habrobracon A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 307 uglandis at three stages of growth in air and nitrogen. Rough survivorship curves are given using eclosion as the criterion of survival. The haploid males apparently do not exhibit an exponential survival. This point should be studied further but it may be that the male haploid insects also exhibit a redundancy in the genetic information similar to Chlamydonwnas in spite of their haploid character. The argument used to derive equation (9) applies in these cases as well as in the diploid case. So far in the discussion it has been argued that deviations from the ideal fonn of the survivorship curves were due to the deleterious agent and p{H, 0) was regarded as having the same form. The fact that many organisms, parti- cularly hybrids, do not exhibit the survivorship curves corresponding to equation (9) is shown in Fig. 3. This behavior is closely associated with the genetic constitution of the organisms. There are a number of facts which support this conclusion. Consider the survivorship curves of vestigial Drosophila melanogaster in Fig. 2 which differs from the wild type, whose survivorship curve is shown in Fig. 3 by a single gene. The same general effect has been reported by Clarke and Smith (47) for Drosophila subobscura. The hybrids between two inbred lines designated 'B' and 'K' exhibit a life span essentially double that of the parent inbred strains. The data are not sufficiently extensive to determine the mathe- matical form of the survivorship curve, but the inbred strains seem to have roughly the same type as the vestigial Drosophila melanogaster of Pearl and Parker shown in Fig. 2 while the hybrid has the same form as the wild type shown in Fig. 3. This effect is also shown by mice. The survivorship curves for nonnal aging are given in Fig. 3 for two hybrid strains and for the hybrids of each with the C57 strain. It is therefore a very plausible conclusion that the survivorship curve is very sensitive to the genetical character of the ensemble and that the change of shape can be ascribed to the form of p(H, 0). This function p(H, 0) plays a role in information theory not unlike the equation of state in thermodynamics. We are at liberty to admit many types of probability distribution in //but it must be the same one in a given ensemble of organisms for all experiments. That this is the case is illustrated for mice by the resemblance between the survivorship curves for gamma- and x-irradia- tion and those for normal aging. The purposes of most of the work in this field, particularly in the case of acute killing, are served by obtaining an LDgQ. The results are ordinarily reported by probit analysis and the life table is not given. We will not attempt to review the very extensive literature, which was not developed for the present purpose. Rather we will quote one experiment which involved a very large number of mice and which has been extensively studied and reported (48, 49, 50). In Fig. 5 the acute killing from atomic bomb radiation as a function of dose is shown from Cronkite et al. (51), on LAFj mice. This is to be compared with the normal aging survivorship curve obtained from the controls. All curves on this figure are normalized by being passed through the 3 per cent survivorship point, after the custom of Pearl. The data of Murray and Hoffman (52) and Murray (53) giving normal aging life tables for hybrid and in-bred mice are also shown. The agreement, of course, is not exact but the curves for gamma-ray acute lethality 308 Hubert P. Yockey agree with the normal aging curves as well as these curves agree with each other. There are several interesting details which should be pointed out. The gamma-ray data show a remarkable coUinearity with aging data below the 10 to 20 per cent survivorship value but rise above the aging curve to a much sharper 'knee'. This effect is probably due to recovery, a phenomenon which is associated with radiation damage but not with aging. Cronkite et al. noted « (woeks) 24 100 • r DOSE (r ) 100 300 500 700 900 13 15 17 1100 21 23(months) o o (months) I 17 Fig. 5. Normalized survivorship of certain strains of mice for normal aging and acute radiation killing for LAF^ (51, 52, 53). LAFj normal aging curve is due to A. C. Upton and A. W. Kimball, personal communication (see also (48)). Radiation dose and age are plotted linearly. a small mortality below about 600 r and took pains to establish its reality as a radiation effect. This feature is also present in the aging curves — a bit more pronounced, to be sure, since there is no recovery. This feature of these curves, incidentally, fits very well with the present theory, or any theory which relates aging and radiation damage, but is otherwise quite a puzzle. These mice were presumed to be as nearly identical as possible and so should have been killed by nearly the same dose. (A discussion of a possible molecular basis for recovery has been given (10). This effect is left out of this article for simplicity.) Cole, Nowell, and Ellis (54) have reported the late survivorship of LAFj mice protected from 800 r of 250 kVP x-rays by spleen homogenate. The survivorship curve has the same shape but is shifted so that the mean age at death is eight months earher than the unirradiated control. This shows that even if a certain organ system, in this case the hematopoietic system, is so aided in repair that its role as a cause of death is greatly modified, neverthe- less, the survivorship curve has the same shape. A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 309 The discussion given above indicates that even though it is necessary to assume p{H, 0) of different shapes for different ensembles of organisms, once chosen, the same shape is required to represent survivorship data with httle regard for the nature of the deleterious agent causing mortality. The case for loss of information content by action of chemical mutagens or carcinogens is less clear than that for radiation. Radiation is no respecter of local chemical detail — it sees only an electron gas held together by positive charges. Side reactions complicate the experimental problems in obtaining good survivorship curves for the chemical radiomimetics. In fact, we know of no such curves at all, although some data of this sort are discussed by Rahn (3). Nevertheless, there is good reason to believe there is considerable miming (55, 56, 57) of radiation effects, in general, and of aging in particular. This is illustrated by a paper by Cloudman, Hamilton, Clayton and Brues (58). They studied malignant tumors and survival in CF-1 female mice painted with methylcholanthrene, irradiated with P^^ /? particles, and with these two insults in combination. A striking decrease in life span was found in those mice painted with methylcholanthrene. This was not due to any single pathologic state, not even to the pulmonary tumors generated in these animals. Survival was also shortened in those mice which did not have pulmonary tumors. The authors had the impression that life was shortened in a general way similar to the life shortening effects of total-body irradiation. The carcinogenic effects of the two agents used in the experiments were approximately additive. This observation as well as the life shortening and the carcinogenesis correlates very well with the view that these effects are manifestations of the destruction of genetic information in the somatic cells. Since exposure to such chemicals is probably on the increase there is a practical as well as a theoretical reason for pursuing this matter further. IV. THE ROLE OF EQUIVOCATION IN THE GERM LINE It was said in my previous article in this volume that the ideas developed there should be applicable both to the germ line and to the somatic line. In this section we shall consider the effect of equivocation on the ability of the germ line to transmit specificity. We do not have available as much experimental material as that which pertains to the somatic line but there are several experi- ments which are very good and are very gennane to the phase of information theory in biology discussed in this section. It should be remembered that there are a number of error correction methods peculiar to the germ line. Among these are fertilization or conjugation and the selection value of the independent existence of cells in the germ line. The germ line may therefore be expected to exhibit a recovery from damage to a degree not found in the soma. An experiment in which the germ line is propagated parthenogenically and so resembles very much the somatic line has been reported by Lansing (59, 60). He studied the effect of parental age on the survivorship of two species of the rotifer, namely, Euchlanis triquetra, which lives normally about a week, and Philodina citrina, which survives normally nearly a month. The method of the experiment was to observe the survivorship curve for 21 310 Hubert P. Yockey a series of generations each produced from eggs laid on a given day in the life of the parent. Lansing called such a series an 'orthoclone'. An orthoclone obtained from a senile stage in the life of the rotifer was designated as an old orthoclone or a 'geriaclone' whereas an orthoclone from adolescent organisms was called a young orthoclone or a 'pediaclone'. For each species it was found that the geriaclone could be followed to extinction in a few generations. In the case of Philodina citrina even the six-day orthoclone died out in the seventeenth generation. It was observed that the longevity of the five-day orthoclone tends to increase. The maximum life span of that orthoclone was not found but appeared to be indefinite. It was found for each species of rotifer that the life shortening could be reversed by starting a pediaclone as an off shoot from a geriaclone. The limit to the ability to lengthen life seemed to be the fact that egg production does not appear until about the fifth day for Philodina citrina and about the fourth day for Euchlanis triquetra. The number of animals used to establish a life table was sixty, a number too small to avoid considerable fluctuations. However, the curves shown in Lansing's papers give the impression that the shape of the survival curves is maintained. This feature seems to be in common with data discussed above in Section III, and in particular with the work of Furth et al. (48), on the late effects of ionizing radiation on mice. The decline and extinction of viability in the germ line is accounted for in the present theory by the accumulation of equivocation in the gene code as it is transmitted in the germ line. The recovery is regarded as being due to selection and propagation of that portion of the ensemble with a relatively low amount of equivocation. The explanation offered by Lansing is quite different from that given here. He attributes his results to a transmissible factor which appears at cessation of growth. In particular, his assertion that the factor is non-genic appears to contradict the point of view adopted here. Actually there is no contradiction with the latter assertion since Lansing v/as undoubtedly thinking of genetic factors in terms of the ideas concerning the gene current at the time of writing, and indeed today. However, as Lansing notes, 'it is striking that the experi- mental observations on the primitive rotifer as well as conclusions derived therefrom are entirely compatible with conclusions drawn from mammalian experiments.' The feature of these and other organisms which is the same is the chemical composition of the genetic material and for this and other reasons it seems to me that an explanation for so ubiquitous a phenomenon as aging must be related to the genome. The germ line provides an opportunity to study the error correction function of conjugation. The most extensive data relating damage in the germ line from one parent or both to survival seem to be due to Mortimer (31, 61). He obtained survivorship curves for yeast zygotes formed by the conjugation of cells of opposite mating types. The following crosses were obtained: haploid X haploid (oo), haploid X diploid (oO), diploid X haploid (Oo), and diploid X diploid (OO). In the symbolism used a capital O represents a diploid cell, a lower case o represents a haploid cell; a filled letter (#0) irradiation. The first symbol indicates the a-mating, the second the a-mating type. A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 311 The survival curve for each haploid type in Fig. 1 is of the usual exponential form, equation (5). According to the discussion in Section II above, this is to be understood as the full expression of recessive lethal mutations. The survival curve for diploid exhibits the sigmoid shape whether the irradiation is done before or after conjugation (31). Note that the abscissa in Fig. 6 is «H 100 000 200 000 X-RAY DOSE .ROENTGENS 300 000 Fig. 6. Survivorship for yeast with o'e irradiated parent. Data from ref. (61). Haploid X haploid cross ( oo ), haploid > diploid cross (oO), etc. The first symbol represents a cell of the a-mating type, the second one of ^-mating type. A filled letter o designates the irradiated parent. Haploid dominant lethal curve: o^ ^q; •, o0; 6 #0; 6- *-**• Diploid dominant lethal curve: □ , #0; ■. O©; n, 0o; h, oO. Diploid survival curve: A,* — aa diploid, A, • — aa diploid. Haploid survival curve : ▲, • — a haploid ; ▲, • — a haploid. Note that abscissa is the square of the dose. the square of the dose. The straight lines have been drawn for comparison purposes only, but, as in Figs. 2, 3 and 4, it is clear that log ///q is well represented by /r. According to the discussion of Section III, this is to be understood as indicating that an error is not expressed in the diploid cell except when errors are paired in the two sets of chromosomes. This follows from the dependence of log ///o on A^. The shielding of errors by a normal allele is seen to be very effective in the (Oo) or the (o©) case so that survival is very much greater than for the 312 Hubert P. Yockey irradiated diploid (••). This shielding seems to be complete for errors due to first order damage; that is, damage such that J{X) = /„ in equation (6). If this shielding were not complete a term in the first power of X would be apparent in Fig. 6. Thus far the application of information theory to the currently accepted model of the diploid cell succeeds very well. We find features we expect and do not find ones we do not expect. Owen and Mortimer's data on the domi- nant lethal survival enable one to study second order effects ordinarily sub- merged in those of the first order discussed in the paragraphs above. Figure 6 shows that there is a very small dominant lethal expression of damage but only when the damage is of the second order in the irradiated parent. That is, a single error or group of errors is shielded but pairs of errors or pairs of groups of errors are expressed, to some degree at least. This is evident since log ///q behaves as P. It is to be expected that this higher order damage exists in the haploid and in the diploid (••) but cannot be observed because of the lethality due to first order damage. The survivorship has the same P behavior for the other ploidies, but a curious feature is that this higher order damage is expressed to a greater degree in the higher ploidies, contrary to expectation. Perhaps this is a model-sensitive phenomenon (as higher order phenomena often are). If that is so further experimentation may tell us more about polyploidy. It was pointed out in Section II that j{X) is related to the interaction of radiation and matter. This indicates that repeating Owen and Mortimer's experiments with other deleterious agents may be very fruitful. For example, Uretz (15) has shown that the ultraviolet survivorship of haploid yeast is sigmoidal. If this means in the case of haploid survival /(A) = J^X, these errors will probably be shielded in the zygote. We choose the next higher term J{X) = J^X^ so that log JJIq may be expected to behave as X^. Higher powers in X may be found in the expansion of /(A) depending on the effectiveness of shielding of recessive lethal mutations in the zygote. These results should apply to organisms other than yeast and in particular to the survivorship of F^ progeny in mice. F^ progeny with one irradiated parent should have a shorter life span than the unirradiated parents. Fj progeny with two irradiated parents should have a still shorter life span. That this is at least partly the case is shown by recently reported results by Russell (34). He reports a life shortening in the offspring of male mice exposed to neutron irradiation from a nuclear detonation. The dose was rather low; the highest to the parent was 186 rep, but only two such oflfspring were obtained. Rather small numbers of individuals were obtained from other parents also so that the estimate of the magnitude of the effect is rough, although its existence seems to be established. The life shortening seemed to be of the same order of magnitude in the father as in the offspring, however. Wallace (62) has reported work on Drosophila in which he has irradiated several populations for as many as 150 generations. His criterion of viability is survival from egg to emergence and this work refers only to the second chromosome. He finds that the fitness of a population does not necessarily continue to decrease under the influence of radiation. These experiments together can be understood from the point of view A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 3 1 3 developed in my previous article in this volume without aJ hoc assumptions. Furthermore, certain interesting predictions can be made. Lansing's remarkable work on the rotifer is a particularly interesting beginning to understanding the problems discussed in this article, if aging, thermal killing, and radiation damage are really aspects of the destruction of information content then there should be, as discussed above, a reciprocity between the respective agents. It would be particularly interesting to know if Lansing's results could be obtained by suitable x-, gamma- or ultraviolet- irradiation, or also by a thermal or chemical treatment. These organisms should be well adapted to this type of research. Among the diploid organisms, of course, mice and Drosophila are of para- mount importance. It would be extremely pertinent to look for the same reciprocity in this material. In addition, one should expect it to be possible, given a strain of one of these animals with a rectangular survivorship curve, see Fig. 5, to change it by irradiation to one of the type corresponding to equation (9) in several generations. Acknowledgements — It is a pleasure to acknowledge the help of Dr. A. W. Kimball in matters concerning the statistics of data analysis. The analysis of T. H. Wood's data was greatly aided by his sending his exact experimental values. Thanks are due also to Dr. Arthur C. Upton and Dr. A. W. Kimball for furnishing the numerical data on the LAF^ normal aging survivorship. I am very grateful to R. K. Mortimer for the use of his data. REFERENCES 1. O. Rahn: The size of bacteria as the cause of the logarithmic order of death. J. Gen. Physiol. 13, 179-205 (1929). 2. O. Rahn: The order of death of organisms larger than bacteria. /. Gen. Physiol. 14,315-337(1930). 3. O. Rahn: Death of bacteria by chemical agents. Biodynamica 5, 1-116 (1945). 4. D. E. Lea: Actions of Radiations on Living Cells, Cambridge University Press, Cambridge, England (1946). 5. M. R. Zelle, and A. Hollaender: Effects of radiation on bacteria. In: Radiation Biology, ed. by A. Hollaender, Chap. 10, Vol. II, McGraw-Hill, New York (1955). 6. A. Hollaender (ed.): Vol. I, Radiation biology, McGraw-Hill, New York (1954). 7. A. Keener, W. D. Bellamy, G. E. Stapleton, and M. R. Zelle: Symposium on radiation effects on cells and bacteria. Bact. Rev. 19, 22-44 (1955). 8. S. M. Dancoff, and H. Quastler: The information content and error rate of living things. In: Information Theory in Biology, 263-270, ed. by H. Quastler, University of Illinois Press, Urbana (1953). 9. C. E. Shannon, and W. Weaver: The Mathematical Theory of Communication, University of Illinois Press, Urbana (1949). 10. H. P. Yockey: An application of information theory to the physics of tissue damage. Radiat. Res. 5, 146-155 (1956). 11. T. Alper: Hydrogen peroxide and the indirect effect of ionizing radiations. Nature, Lond. 162, 615-616 (1948). 12. J. Watson: The properties of x-ray inactivated bacteriophage. II Inactivation by indirect effects. /. Bact. 63, 473^85 (1952). 314 Hubert P. Yockey 13. T. Alper: Bacteriophage as indicator in radiation chemistry. Radiat. Res. 2, 119-134 (1955). 14. F. L. Gates: A study of the bacteriocidal action of ultraviolet light. J. Gen. Physiol. 13,231-260(1929). 15. R. B. Uretz: Additivity of x-rays and ultraviolet light in the inactivation of haploid and diploid yeast. Radiat. Res. 2, 240-252 (1955). 16. E. H. Anderson: The effect of oxygen on mutation by x-rays. Proc. Nat. Acad. Sci., Wash. 37, 340-349 (1951). 1 7. A. Hollaender, W. K. Baker, and E. H. Anderson : Effect of oxygen tension and certain chemicals on the x-ray sensitivity of mutational production and survival. Cold Spr. Harb. Symp. Quant. Biol. 16, 315-326 (1951). 18. A. Hollaender, and G. E. Stapleton: Fundamental aspects of radiation protection from a microbiological point of view. Physiol. Rev. 33, 77-84 (1953). 19. G. E. Stapleton, A. J. Sbara, and A. Hollaender: Some nutritional aspects of bacterial recovery from ionizing radiations. J. Bad. 70, 7-14 (1955). 20. R. E. Zirkle, and C. A. Tobias: Effects of ploidy and linear energy transfer on radio- biological survival curves. Arch. Biochem. Biophys. 47, 282-305 (1953). 21. W. F. Friedewald, and R. S. Anderson: Factors influencing the inactivation of the rabbit papilloma virus by x-rays. Proc. Soc. Exp. Biol., N. Y. 45, 713-715 (1940). 22. S. E. Luria, and F. M. Exner: The inactivation of bacteriophage by x-rays — influence of the medium. Proc. Nat. Acad. Sci., Wash. 27, 370-375 (1941). 23. W. M. Dale: The effect of x-rays on enzymes. Biochem. J. 34, 1367-1373 (1940). W. M. Dale: The effect of x-rays on the conjugated protein d-amino acid oxidase. Biochem. J. 36, 80-85 (1942). 24. T. H. Wood: Influence of temperature and phase state on x-ray sensitivity of yeast. Arch. Biochem. Biophys. 52, 157-174 (1954). 25. T. H. Wood: Lethal effects of high and low temperature on unicellular organisms. Advances in Biological and Medical Physics 4, 1 19-165 (1956). 26. L.H.Gray: Cellular radiobiology. Annual Review of Nuclear Science 6, 353-422 (1956). 27. N. Nybom: Some experiences from mutation experiments in Chlamydomonas. Hereditas 39, 317-324 (1953). 28. B. S. Jacobson: Reversible and irreversible effects of x-rays on vegetative survival in Chlamydomonas reinhardi. Ph.D. thesis, Univ. California, Berkeley (1957). B. S. Jacobson: Evidence for recovery from x-ray damage in Chlamydomonas. Radiat. Res. 7, 394-406, (1957). 29. C. Tobias, and B. Stepka: University of California Radiation Laboratory Report No. 1922,(1952). 30. R. K. Mortimer, and C. Tobias: Evidence for x-ray induced recessive lethal mutations in yeast. Science 118, 517-518 (1953). 31. R. K. Mortimer: Evidence for two types of x-ray lethal damage in Saccharomyces cerevisiae. Radiat. Res. 2, 361-368 (1955). 32. G. A. Sacher: On the statistical nature of mortality, with especial reference to chronic radiation mortality. Radiology 67, 250-257 (1956). 33. W. L. Russell: Genetic effects in mammals. In: Radiation Biology, ed. by A. Hollaender, Chap. 12, Vol. I, McGraw-Hill, New York (1954). 34. W. L. Russell: Shortening of life in the offspring of male mice exposed to neutron radiation from an atomic bomb. Proc. Nat. Acad. Sci., Wash. 43, 324-329 (1957). 35. R. Pearl, and S. L. Parker: Experimental studies on the duration of life, IX. New life table for Drosophila. Amer. Nat. 58, 71-82 (1924). 36. P. H. Leslie, and R. M. Ranson: The mortality, fertility and rate of natural increase of the vole (Microtus agrestis) as observed in the laboratory. J. Anim. Ecol. 9, 27-52 (1940). A Study of Aging, Thermal Killing, and Radiation Damage by Information Theory 315 37. W. D. Oliff: The mortality, fecundity and intrinsic rate of natural increase of the multi- mammate mouse, Rattiis {Mastomys) natalensis (Smith) in the laboratory. J. Anini. Ecol. 22,217-226(1953). 38. L. C. Birch: Experimental background to the study of the distribution and abundance of insects. Ecology 34, 698-711 (1953). 39. T. H. Wood: Reproducibility of x-ray survival curves for yeast cells. Proc. Soc. Exp. Biol., N. Y. 84, 446-452 (1953). 40. A. W. Kimball: The fitting of multi-hit survival curves. Biometrics 9, 201-211 (1953). 41. S. E. LuRiA, and R. Dulbecco: Genetic recombinations leading to production of active bacteriophage from ultraviolet inactivated bacteriophage particles. Genetics 34, 93-125 (1949). 42. A. C. GiESE, and D. C. Heath: Sensitization to heat by x-rays. /. Gen. Physiol. 31, 249-258 (1947). 43. A. C. GiESE, and E. B. Grossman: The action spectrum for sensitization to heat with ultraviolet light. /. Gen. Physiol. 29, 79-87 (1945). 44. A. C. GiESE, and E. B. Grossman: Sensitization of cells to heat by visible light in presence of photodynamic dyes. /. Gen. Physiol. 29, 193-202 (1946). 45. W. F. Baldwin: Similarities in killing by heat and by x-radiation in the insect Dahl- bominus fiiscipennis (Zett.). Radiat. Res. 5, 46-51 (1956). 46. A. M. Glark, and E. B. Herr Jr.: The effect of certain gases on the radiosensitivity of Habrobracon during development. Radiat. Res. 2, 538-543 (1955). 47. J. M. Glarkf, and M. J. Smith: The genetics and cytology of Drosophila siibobscura. XI Hybrid vigor and longevity. /. Genet. 53, 172-180 (1955). 48. J. Furth, a. G. Upton, K. W. Ghristenberry, W. H. Benedict, and J. Moshman: Some late effects in mice of ionizing radiation from an experimental nuclear detonation. Radiology 63, 562-570 (1954). 49. A. G. Upton, and J. Furth: Induction of pituitary tumors by means of ionizing irradia- tion. Proc. Soc. Exp. Biol., N.Y. 84, 255-257 (1953). 50. A. G. Upton, and J. Furth: Spontaneous and radiation induced pituitary adenomas of mice. /. Nat. Cancer Inst. 15, 1005-1015 (1955). 51. E. P. Gronkite, V. P. Bond, W. H. Ghapman, and R. H. Lee: Biological effects of atomic bomb gamma radiation. Science 122, 148-150 (1955). 52. W. S. Murray, and J. G. Hoffman: Physiological age as a basis for the com- parison of strains of mice subject to spontaneous mammary carcinoma. Cancer Res. 1, 298-307 (1941). 53. W.S.Murray: Studies on the inheritance of mammary carcinoma in the mouse. Cancer Res. 1, 123-129 (1941). 54. L. J. GoLE, P. G. NowELL, and M. E. Ellis: Incidence of neoplasms and other late lesions in mice protected against lethal x-ray doses by spleen homogenate. J. Nat. Cancer Inst. 17, 435^t45 (1956). 55. B. Wallace: Dominant lethals and sex-linked lethals induced by nitrogen mustard. Genetics 36, 364-373 (1951). 56. S. Pomper, and K. G. Atwood: Radiation studies on fungi. In: Radiation Biology, ed. by A. Hollaender, Ghap. 11, Vol. II, McGraw-Hill, New York (1955). 57. O. G. Fahmy, and M. J. Fahmy: Ghemical and radiation mutagenesis in Drosophila melanogaster. In: Progress in Radiobiology, 205-210. Proceedings of the Fourth Inter- national Gonference on Radiobiology, Oliver and Boyd, Edinburgh and London (1956). 58. A. M. Gloudman, K. A. Hamilton, R. S. Glayton, A. M. Brues: Effects of combined local treatment with radioactive and chemical carcinogens. /. Nat. Cancer Inst. 15, 1077-1084(1955). 59. A. I. Lansing: A transmissible, cumulative, and reversible factor in aging. J. Geront, 2, 228-239 (1947). 316 Hubert P. Yockey 60. A. 1. Lansing: Evidence for aging as a consequence of growth cessation. Pioc. Nat. Acad. Sci., Wash. 34, 304-310 (1948). 61. M. E. Owen and R. K. Mortimer: Dominant lethality induced by x-rays in haploid and diploid Saccharomyces cerevisiae. Nature, Lond. Ill, 625-626 (1956). 62. B. Wallace: Studies on irradiated populations of Drosophila melanogaster. In: Progress in Radiobiology, 201-204, Proceedings of the Fourth International Conference on Radiobiology, Oliver and Boyd, Edinburgh and London (1956). ENTROPIC CONTRIBUTIONS TO MORTALITY AND AGING* George A. Sacher Division of Biolo^^ical and Medical Research, Argonne National Laboratory, Lemont, Illinois Abstract — All dynamic physiologic processes are attended by fluctuations. The magnitude of these fluctuations is determined by the inherent regulatory capacity of the specific process and by the magnitude of random disturbances arising both in the environment and within the organism. A system with these characteristics has, in a given environment, a determinate probability of failure per unit time. As a consequence of the ubiquitous random component in physiologic performance, a population of individuals that are indistinguishable by any combination of physiologic measurements will nevertheless manifest time-survival and dosage- survival curves with finite dispersions. This is illustrated by means of a one-dimensional model system subjected to a stationary Gaussian random noise disturbance. In real biological populations, there is a component of variance between individuals. This can be taken into account by a straightforward generalization of the basic equations for homogeneous popula- tions. In this approach, aging is interpreted as a secular change in the values of the parameters of the regulatory mechanisms. These secular changes are ultimately due to irreversible changes in permanent or self-reproducing macromolecules. The rate of such irreversible change is in turn dependent in part on the magnitude of local fluctuations away from ideal steady state conditions for biochemical syntheses. There are thus two aspects to the stability of organisms — the probability of mortality per unit time and the rate of increase of this probability with time (age). Both are intimately dependent on the fluctuation characteristics of physiologic performances. I. INTRODUCTION This paper discusses mortality and aging insofar as they depend on certain statistical characteristics of organisms and populations. These characteristics, which may be subsumed under the closely related concepts of fluctuation, entropy, and information, have their origin in the dynamic nature of physiologic processes. Much of the current methodology for the analysis of survival curves is founded on the theory that the observed distributions of survival are due to the existence of a distribution of sensitivities in the populations tested. The present discussion is intended to emphasize the statistical nature of the mortality process within the individual, or in populations of indistinguish- able individuals. Only those aspects of behavior are considered that have to do with the establishment and preservation of the steady state of physiologic function, and that can be described by a set of fixed relations among a finite, and in fact quite small, number of physiologic processes. Implicit in this approach is the conception of physiologic process as functional unit rather than as ultimate enzymatic reaction-step. * Work performed under the auspices of the U.S. Atomic Energy Commission. 317 318 George A. Sacher II. PHYSIOLOGIC REGULATIONS The ability to maintain the physiologic steady state in the face of an unfavor- able environment is called homeostasis (1). A number of quantitative indices of homeostatic capacity are in use. In ecological studies the tolerable range of an environmental variable, such as ambient temperature or salinity (of sea waters), is widely employed (2). Resistance to transient stresses is a more common measure in experimental physiology. If the response can be followed continuously, measures such as the amplitude of displacement of function, and rate of return to normal may be obtained. The above may be referred to as determinate measures of homeostatic capacity, for they reflect the fact Fig. L Schematic representation in two dimensions of the probabihty distri- bution of physiologic states and their relation to the boundary delimiting viable from non-viable states. The probability distribution is indicated by elliptical contours of equal probabihty. The dehmiting boundary (L-L), called the lethal bound, is indicated as a sharp line, and is treated as a precise value in the derivation of equation 15. A more realistic representation is given in Fig. 2. that the homeostatic mechanism, even if it functions perfectly and without error, has but a finite regulatory capacity, set by the physical limitations of the mechanism. In summary, there is a closed region in the physiologic configuration space within which some degree of stable physiologic function may persist, and beyond which stable function is impossible. This is indicated schematically in Fig. 1. The boundary surface of this region will be called the lethal bound, and denoted byL. The quantitative properties of the lethal bound diff'er for different physiologic processes. In the case of white blood cells, the lethal bound on the low side, either for number of circulating cells or for number of proliferative cells, is only a small fraction of the normal level. Similarly the lethal range on the high side is considerably above normal levels. The boundary values for erythro- cytes lie somewhat closer to the normal values. Blood glucose is rather more Entropic Contributions to Mortality and Aging 319 sharply limiting on the low side than on the high. Blood pH must be held to very close tolerances on both sides of normal. III. SOME PROPERTIES OF FLUCTUATIONS From a consideration of the components of variation within and between individuals under different environmental conditions, it can be inferred that the observed variation can be attributed to (a) random fluctuations in the common environment that have a uniform influence on all animals maintained therein and (b) independent random fluctuations of each animal that must originate either within the animal or in local fluctuations of the environment that are independent for each animal. The magnitude of the fluctuation arising within the animal due to internal random noise is reasonably well known in a few experimental situations of a psychophysical or neurophysiological nature. Except for certain obvious aspects, such as temperature and humidity, the nature and properties of the environmental random variables is for the most part unknown. More significant perhaps than the purely environmental random variables are these that might be classified as organism-environment inter- actions. Such relations as pathogenicity, parasitism, dominance-submission, predator-prey relationships, etc., are in this class, and make contributions to the variability of individual perfonnance that defy estimation. For the present purpose, the fact that intra-individual fluctuations exist is sufficient : the question of their nature can be deferred. IV. FLUCTUATION AND THE PROBABILITY OF MORTALITY The set of physiologic processes can be written formally as ^i = ^i k„ ^,;, ^;, /] (/,;■= 1, ..., n) (1) where the X^ denote the set of physiologic variables, the a,_,- denote internal parameters, and the A^ external parameters. The state of an individual at a given moment is specified by the values at that moment of the n physiologic variables X^. One can conceive of, and in principle construct, a population made up of indistinguishable individuals, in which the value of each internal parameter for every member lies within an arbitrarily small range. In such a population under constant environmental conditions the time-average of any function of the physiologic variables, X,, for one individual is equal to the average over the population of the same function of the A'j at any moment in time. If we locate a frequency distribution of physiologic states in the configuration space, the result is as seen in Fig. 1. The contours enclosing percentages of the distribution (such as 50, 90, 99) are drawn approximately in accord with the supposition that the bivariate distribution of states is Gaussian, and the situation is roughly in scale for a 'healthy' population, i.e. there is only a small probability of observing states near the lethal bound. Since contact with the lethal bound removes an individual from the popula- tion, the distribution of states must be modified in the neighborhood of the boundary. Furthermore, the frequency distribution of states is not by itself 320 George A. Sacher enough to permit a calculation of the probability per unit time that fluctuations will reach L. To answer these questions, we turn to a consideration of the dynamic nature of the fluctuation process within the individual. The complete description of a fluctuation process is given by specifying its correlation function, which is in one dimension P(t) = {X{t)x{t + T)>av/<.v2(r)>av (2) where x is a deviation from the mean, ^=^0 + ^^ (3) The correlation function is a measure of the degree to which a fluctuation present at time t persists at a later time r -\- t, averaged over all values of t. The nature of p(t) depends on the nature of the system. A process that obeys the differential equation ^ + ^x = 0 (4) returns to equilibrium as X = :Co e-^^ (5) Corresponding to this, if a stationary pure random Gaussian noise source f{t) is applied, ^ + ^x =/(/) (6) the resulting correlation function is p{r) = e-P^ (7) It can be shown (3) that if the correlation function in one dimension is given by equation (7), then the fluctuation process is Markoffian and is completely described by the joint probability distribution W.{x^xS = 27ra2 (1 _ p2)* X exp 2a\\ - p2) ^1^ + -^2^ — 2p.YiX2 (8) This gives the joint distribution of observations of x separated by time t, where p is defined by equation (7). The variance, cr^, of the distribution of .y satisfies a2 = Dl(i (9) where 4D is the (constant) spectral density of the random noise (white noise) source. The conditional probability distribution, which describes the distri- bution of ^2 when x^ is fixed, is n-Yo/x, /) = [27702(1 - p2)]i X exp [-(.Y - xfl2a\\ - p"-)] (10) Entropic Contributions to Mortality and Aging 321 where a~ and p are deiincd as above and X = x^e'^'^ (11) As /— > 00, this becomes the stationary distribution of fluctuations, This distribution is Gaussian because of the linearity of mechanism specified by equation (4). Fluctuation processes are not in general Gaussian if the dynamical equations are non-linear. Equation (12) gives the stationary frequency distribution of fluctuations over the entire :>c-axis. We now introduce the condition that there is at Xg^ an absorbing barrier at a distance X from the mean X— X^ — Xq An individual remains in the distribution only as long as the path described by his fluctuation process remains in the region .v < ).. If this situation prevails for a time sufficiently longer than the relaxation time of fluctuations, 1//?, a stationary distribution is again established and there will be a stationary probability q per unit time that the path will intersect x = /I. This 'absorption rate' is the mortality rate for the model fluctuation process. The stationary frequency distribution in the presence of an absorbing barrier may be obtained from equation (12) by the following argument. In the steady state there is a stationary diffusion current, j, into the barrier. The desired frequency distribution Q{x) must satisfy the steady state difl'erential equation for diffusion in the presence of a force field (4). This equation is, in the notation of equations (6) and (9), j = KQ{x)-(^a''^^^Q{x) (13) where K = —fix is the restoring force. A solution of equation (13) satisfying the boundary condition Q(x) = 0 (x^ A) is 3:2 (x-2;)2-l Q(x) 1 (27702) 2\i 2(72 . ^ 2(T2 (14) The mortality rate, q, is equal to the diffusion current, j, normalized by the area under the distribution, Q{x), so that from equation (13) we find the constant mortality rate to be A rigorous discussion of this class of stochastic processes (4) indicates that the validity of equation (15) is subject to the limitation 322 George A. Sacher This restriction is met if we have X > 3a (as is the case in the appHcations considered). The normahzing integral in equation (15) is then not appreciably less than unity, so equation (15) for the mortality rate reduces to (A>3a) (1-a) Equation (15) gives the dependence of the mortality rate on the parameters /?, A and a (or /i, A and D) in the stationary state of a system specified by equation (6) and subject to a stationary random force function with spectral density 4D. Although this model is too simple and artificial to be an adequate description of an actual mortality process, it should be noted that equations equivalent to equation (4) give an approximate description of a number of different physiologic mechanisms. Equation (15) can be extended to the case of time-dependent mortality rates, as they are observed in animal populations, if the parameters are sufficiently slowly changing functions of time, so that stationariness of the fluctuation process is preserved. This is a reasonable assumption with regard to the life tables of animal populations in their normal environments. It is also considered for the purpose of this discussion that the fixed and the random components of environmental forces are stationary throughout life. Experimental data on homeostatic capacity for a variety of mechanisms as a function of age indicate that this capacity diminishes during adult life (5). We therefore expect a steady decrease in the value of />. Since a^ = Djft, the value of a will be increased by a decrease in /?. The observed dispersion of physiologic variables does not increase markedly with age. This may imply that the recovery constant does not diminish much during the life span, but it may also be due to the eff'ect of the distribution of parameters in the population, since it can be estimated that about half the total variance in a typical outbred population is variance between members, and this variance is reduced by selec- tion, for as mortality proceeds in a heterogeneous population the subpopulations with the more disadvantageous parameter values will experience heavier mortality and thus be preferentially eliminated from the surviving population. We have also examined (6) one simple mathematical model of a homeostatic mechanism that introduces a plausible type of non-linearity of recovery. In this model there arises a relation between the location of the mean state and the value of the recovery constant. In the notation used here, 1 and /9 would decrease concomitantly. The methodological difficulty in the study of mortality processes is that mortality data are not by themselves sufficient for the unique determination of their parameters, even in the simplest cases. In earlier treatments (7) the expedient was therefore adopted of assuming that the mean state, X, is the only parameter that changes with age. There is abundant evidence that the mean values of physiologic variables change with age (8). Advantage was also taken of the fact that changes in mean physiologic state with age are usually small in degree. This justified taking the linear term of the expansion of A^ about the initial value Aq, A2 = V + 2AoAA (16) Entropic Contributions to Mortality and Aging 323 Then taking logarithms and collecting the constant terms into lumped constants the relation was obtained G(0 — log <7(/) - « -f /> AA (17) where G(t) is called the Gompertzian. The linear approximation to A^ is satisfactory for a first-order description of the relation of injury to age and to dosage of agents that cause permanent injury, such as x-rays (7). The fact that /i also tends to decrease with age does not alter the generaliza- tion made previously (7) that the Gompertzian is a linear measure of mean physiologic state. The entire exponent 2^2 ~ 2D can be expanded, yielding an expression of the form ^=^a„ + a,A?. + a.,Aft (18) Furthemiore if the mechanism depends on several variables, the same expansion procedure again yields an exponent term that is a linear function of the dis- placements of all of the parameters. Thus, within the range of parameter values that occur in the course of natural aging, the Gompertzian is an approximate linear measure of the mean physiologic state. V. DESCRIPTION OF THE «-DIMENSIONAL FLUCTUATION PROCESS The consideration of the general 77-dimensional case will take as its starting point the empirical description of the w-variate process in tenns of its moments. The observational data consist of a large number w of sets of observations on one individual or on w indistinguishable individuals, where each set is a measurement of each of n variables at a given time. The first moments are the n mean values, 1 m x,o = - 2 X,, (19) m J = 1 The second central moments are the covariances m Vi, = — (20) The covariances are related to the standard deviations and correlation coefficients as Vik = f^i (^k Pik (21) where c, are the standard deviations and p^,^ is the total correlation coefficient between the /th and ^th variables. The covariance matrix nhi • ' • I'm ■ ... p III '^ II n (22) 324 George A. Sacher is a non-negative quadratic form, as is the correlation matrix iPn • • • Pin R (23) IPnl Pnn Given the covariance matrix V, the frequency distribution of the displacements in n dimensions is determined. In the case that V is positive-definite, so that the rank is equal to the order n, the distribution is (9) p(xj_, • • • , x„) 1 T^exp 2K tk ^''"'''"''' (24) wherQ V is the determinant of V and K,j. is the cofactor of v,^. in V. The coefficients, K,JF in the exponent of equation 24 are terms in the inverse of the covariance matrix, l7{U=Y-i=A={A,,} V (25) If V is positive semi-definite, the rank r is less than the order n, and the frequency distribution is an /--dimensional distribution in /• independent linear functions of the x^ (9) yx = l ai;,\-, (;. = 1, •••,/•) (26) or y ^ Ax (27) and the moment matrix for y is M = AVA' (28) The frequency distribution for y is then — lAfl^^^y^yi (29) ^Oi, OV) - (.2^y«/2(^^/y/2 exp The case that the rank of the matrix of covariances is less than the order is frequently encountered in the initial description of biological systems in terms of the variables of direct observation. VI. FLUCTUATION, ENTROPY, AND INFORMATION The entropy of a system at equilibrium has, in classical thermodynamics, a precise value, i'oCvjg, • • • , x„q) where the .y,o are the values of the state variables at equilibrium. However, if thermal agitation or other disturbance causes small displacements of the state variables, the entropy decreases by an amount, AS. Expanding AS in a Taylor series in terms of the displacements, .y, (10), we obtain (30) Entropic Contributions to Mortality and Aging 325 plus higher order partials. At equilibrium the first partial is equal to zero, so ^S^lljf^x,x, (31) = -i2 S,/^x,x, (32) where 5*,/ is a positive-definite matrix. There is a formal equivalence between the S^j^ and the /■; defined by equation (24), SO = ^A (33) where k is Boltzmann's constant. Thus the X^j, which we may term the partial coefficients of the frequency distribution of fluctuations, are proportional to the coefficients 5",/ of the quadratic form for the mean entropy decrease due to fluctuation in the system. The physiological systems that are under consideration and are not completely described by a small number of variables, and accord- ingly the complete fluctuation distribution and fluctuation entropy would not be estimated. However, the S^j, or the A,^, are additive, so an initially incomplete description can be completed as knowledge of the system increases. From the definition of entropy by Boltzmann S = k J p{x) log p(x) dx (34) it follows that the fluctuation entropy coefficient S^^^ in equation (32) can be written where /J = p{x^, • • • , x,). In one dimension this reduces to S» = -J^%./.v (36) This is identical with the definition of information given by R.,A. Fisher (11). The equivalence continues to hold in the n-dimensional case. It should be noted, however, that the Fisher infoiTnation is a defined quantity, whereas the 5',_," are terms in an approximation formula. There is a close relationship between information theory and the analysis of fluctuation processes as can be made evident in terms of the equivalences brought out above. Where there are distinct classes the information is H=-Y.p,\ogp, (37) In the one-dimensional continuous case we write H{x) = -J p{x) log p{x) dx (38) In a large number of cases, the representation of log/?(.T) in terms of three terms of a Taylor series is a good or even an exact description. The expression for the information then becomes ^W = - J /?(a-) log /X'Yo) + \x^ p{x) ^^ log /j(.Vn) dx (39) 22 326 George A. Sacher The information function is thereby resolved into separate terms for the expected values and for the deviations from expectation. The analysis of fluctuation processes falls into the latter class. The formal equivalences between fluctuation entropy and Fisher information does not imply complete equivalence of the concepts. The theory of entropy fluctuations deals with the stationary fluctuation process in a single individual or in a group of indistinguishable individuals, where in either case the ergodicity condition is satisfied. There is no such restriction on the applicability of the Fisher information. The case of non-ergodic populations (individual differences in parameters) can be covered by obvious generalizations of the fluctuation theory, so this distinction is not a permanent one. Determination of the Lethal Bound Thus far in the presentation the existence of the lethal boundary surface has been a postulated property of physiologic mechanisms. In terms of the linear models of fluctuation processes that have been discussed the lethal bound is of necessity an arbitrarily assumed property, for a continuous linear process by its nature has no failure point. The escape from this unsatisfactory situation is by way of a more thorough mathematical analysis of homeostatic properties. The lethal boundary has a natural interpretation as a 'divide' on a potential surface (compare with Fig. 2). When it is possible to discuss the homeostatic processes as non-linear systems with multiple equilibria, the lethal bound, and also the boundaries between different viable steady states, will appear as necessary topological properties of the physiologic mechanisms. We have under way some investigations of simple non-linear stochastic mortality models, and the early results are quite interesting (6). VII. ENTROPIC CONTRIBUTION TO THE AGING PROCESS Brief consideration was given above to the direction of change of homeo- static parameters with age. This section will deal with the influence of physiologic fluctuation on the rate of aging. It is an intuitive judgment that physiologic steady states of organisms tend to maximize the efficiency of physiologic function in the environments to which the organisms are fitted. The approach to greatest efficiency is presumably by means of natural selection operating on the genetically controllable thermo- dynamic properties of enzymes. The characteristics of physiologic performance, and in particular the values of the phenomenological rate constants are ultimately dependent on the activities, specificities and stabilities of the constituent enzymes. Thus, to give an account of the age changes in the values of the phenomenological parameters one must turn to the consideration of function at the biochemical level. The rate at which irreversible change occurs in a biological system will be discussed for three situations: (a) as a function of temperature, independent of metabolic activity; (b) as a function of metabolic activity in an undisturbed steady-state; (c) in a steady-state disturbed by ffuctuations much greater than thermal, i.e. by the fluctuations of physiologic state discussed above. The analysis of irreversible molecular changes as a function of temperature Entropic Contributions to Mortality and Aging 327 is a part of the general theory of absolute rate processes, and is also the object of a great deal of experimental work, particularly on proteins. It is discussed in another paper in this volume (12). It has been suggested by a number of investigators that the rate of aging is a function of the level of metabolic activity. In evidence of this is the relation m < (r > < o o o _i o en >- X a. -2 0 2 4 6 8 10 12 PHYSIOLOGICAL VARIABLE X Fig. 2. Interpretation of the probability distribution of fluctuation and of the lethal bound in terms of a potential which is a function of the physiologic state variables (6). The solid curves are the isopotential contours. The dashed line L'-L" is the lethal hound. This is the highest point ('divide') on the potential surface in any direction from the steady state, O. Different parts of the lethal bound can be at different potentials. If the potential is markedly lower in one part of the divide, most escapes will occur through this pass. Such preferential directions of escape may be identified with the occurrence of specific disease conditions. The contour lines are isopotential contours of the potential function E = a^x" — biX^ + a^y^ — biV^ A stochastic mortality process has already been investigated for the one- dimensional case of the cubic potential (6). between metabolic rate (or body size) and life expectation among the Mammalia, and the dependence of life expectation on environmental temperature in cold- blooded forms. However, the relation is not a simple one. Birds, with body temperatures as much as 5°C higher than mammals (13), tend to have life 328 George A. Sacher expectations (14) considerably greater than mammals (15) of equal body size and metabolic rate. Some primates (16) outlive carnivores or herbivores of equivalent body size by a considerable margin. This is obvious in the case of man compared with lower animals. In a study on grasshoppers, in which growth rate and Hfe expectation were investigated as functions of temperature (17), the Arrhenius coefficient for growth rate was /^ = 18,400 cal and that for mean death rate was jj, = 6300 cal. Though both are temperature-dependent, the difference in /j, values suggests that survival is not directly dependent on metabolic rate. An association between the level of metabolic activity and the rate of accumulation of irreversible molecular change is certainly to be expected on physico-chemical grounds. The presence of poisons in the environment, and the ever-present possibility of incorrect reactions, imply the existence of a non-zero error rate per molecular event. Finally we consider the influence of macroscopic fluctuations in physiologic state on the rate of accumulation of irreversible molecular changes. The calculation of the error rate due to fluctuation for a particular biochemical reaction would require a more detailed specification of the fluctuation process than is envisaged in the previous development, which dealt with a comparatively small number of important physiologic functions. This fluctuation in state of the organism as a whole would certainly play a part, but it would be necessary to consider in addition the independent fluctuations of small regions. These would usually have little immediate influence on the state of the whole organism, but they would be significant for the probabilities of irreversible change within the regions. The consideration of the problem of local fluctuations cannot be undertaken here. It is presumed that the physiological steady state condition is one in which, through the action of natural selection, the ratio of incorrect to correct reactions is a minimum. This minimum rate is the metabolic error rate e^i, defined above. Deviations from the steady state in any direction bring about conditions in which the probability of incorrect reactions increases. This component of the error rate is called the fluctuation error rate, e^. The fluctuation error rate would then in general be a monotone increasing function of the displacement, and the simplest assumption is that this function is a quadratic. In one dimension this is Ej, = /77.\-2 (40) where x is the displacement from the steady state . Then, for the one-dimensional model process discussed above, with stationary distribution of displacements given by equation (12), the mean error incidence per unit time is r)0-< ■ x^e 2rT2^.v (41) (2770- We find Ep = ma^ (42) This is not a solution of the problem, for the evaluation of m cannot yet be carried out. However, the essential point for the present discussion is that the fluctuation error rate is an increasing function of m and of the dispersion of displacements, g. Entropic Contributions to Mortality and Aging 329 The mortality rate for the same model system. -m e 2a2,(;>3(;) (15a) is also an increasing function of a'-, for the exponential factor in equation (15) varies much more strongly with g than does the constant factor -. The effect of accumulating errors will be to reduce A and jS and to increase a (see above, Section IV). All of these changes tend to increase the mortality rate as age increases. Therefore it is concluded from this quahtative discussion that the mortality rate of different species, and the rate of increase of mortality rate with age for the same species are positively correlated. There are too many uncertain- ties to pemiit a statement of the functional relation between these quantities at present. However, we have here a possible basis for the relative constancy in the form of the life table for species as widely different as fruit fly, mouse and man. The total error rate includes all three terms discussed above £total = fr + ^.1/ + ^F (43) where the subscripts denote temperature, metabolic rate and fluctuation, respectively. The existence of contributions to the error rate arising from back- ground ionizing radiations and other environmental noxae must also be acknow- ledged. Perhaps the best viewpoint is that the physical basis for each term demonstrably exists, but we do not know the absolute contribution of any of them. This will be a major experimental problem. All of these contributions arise when the environment and the population are in a steady state of fluctuation. The course of aging is also influenced to an important extent by very large disturbances that occur infrequently in the lifetime of the individual. Illness and crippHng accident are examples, but changes of nutrition, etc., have equally important effects, as do also insults such as adventitious poisoning. The unique nature of these events requires that they be treated historically rather than on the basis of statistical uniformity of occurrence. Under experimental conditions it can be shown that exposure of a population to ionizing radiations leaves a pennanent residue of injury (7). Jones (18) has demonstrated that human sub-populations selected on the basis of a history of given diseases have a permanent increase in their mortality at later ages. Some writers have attributed aging in general to the action of such major disturbances. Against this position it can be argued that the large common factor in the aging of human or animal populations points to an agency that acts with comparative uniformity on all members of the population and within each individual over the course of life. This is compatible with the statistical uniformities that appear in the summation of a large number of small indepen- dent events as proposed herein. REFERENCES 1. W. B. Cannon: Organization and physiological homeostasis. Physiol. Rev. 9, 399-431 (1929). 2. C. L. Prosser: Physiological variation in animals. Biol. Rev. 30, 229-262 (1955). 330 George A. Sacher 3. M. C. Wang and G. E. Uhlenbeck: On the theory of Brownian motion II. Rev. Mod. Phys. 17, 323-342 (1945). 4. S. Chandrasekhar: Stochastic problems in physics and astronomy. Rev. Mod. Phys. 15, 1-89 (1943). 5. W. B. Cannon: Aging of homeostatic mechanisms. In: Problems of Aging, 2nd ed. ed. by E. V. Cowdry, Williams and Wilkins, Baltimore (1939). 6. G. A. Sacher: Generation of a model dosage — survival curve from a stochastic hypothesis (abstract). Radiat. Res. 6, 447 (1957). 7. G. A. Sacher: On the statistical nature of mortality, with especial reference to chronic radiation mortality. Radiology 67, 250-257 (1956). 8. M. Burger: Altent und Krankheit, Ind ed., G. Thieme, Leipzig (1954). ed. by A. I. Lansing: Problems of Aging, 3rd ed. Williams and Wilkins, Baltimore (1952), L. BiNET and F. Bourliere: Precis de gerontologie, Masson et Cie, Paris (1955). 9. H. Cramer: Mathematical Methods of Statistics, Princeton University Press (1946). 10. P. Epstein: Textbook of Thermodynamics, J. Wiley and Sons, New York (1937). 11. R. A. Fisher: Statistical information. In: Information Theory in Biology, 12-13, ed. by H. Quastler, University of Illinois Press, Urbana (1953). R. A. Fisher: Probability, likelihood and quantity of information in the logic of uncertain inference. Proc. Roy. Soc. {A) 146, 1-8 (1934). M. G. Kendall: The Advanced Theory of Statistics, Vol. II, Chas. Griffin and Co., London (1946). 12. H. P. Yockey: A study of aging, thermal killing and radiation damage by information theory. This volume. 13. S. Brody: Bioeuergetics and Growth, Reinhold Publishing Co., New York (1945). 14. S. S. Flower: Contribution to our knowledge of the duration of life in vertebrate animals. V. Mammals. Proc. Zool. Soc. Lond. 1931, 145-234 (1931). F. Bourliere: Natural History of Mammals, Knopf, New York (1954). 15. S. S. Flower: Contributions to our knowledge of the duration of life in vertebrate animals. Birds. Proc. Zool. Soc. Lond., 1915, U65-\A12{\925). 16. E. P. Walker: The Monkey Book, Macmillan, New York (1954). R. M. Yerkes and A. W. Yerkes: The Great Apes, Yale University Press, New Haven (1929). 17. J. R. Parker: Bull. Univ. Mont. Agric. Res. Sta., no. 223 (1930). 18. H. B. Jones: A special consideration of the aging process, disease and life expectancy. In: Advances in Biological and Medical Physics, ed. by J. H. Lawrence and C. A. Tobias, 4, 281-337, Academic Press, New York (1956). A QUANTITATIVE DESCRIPTION OF LATENT INJURY FROM IONIZING RADIATION* H. A. Blair Department of Radiation Biology, University of Rochester School of Medicine and Dentistry Abstract— A group of hypotheses previously discussed by the writer to account for the kinetics of radiation injury in mammals is reviewed. That radiation injury is proportional to dose, is partly irreversible, and that irreversible injury adds to new acute injury to produce lethality, appear to be vaHd. Recovery is not a single process for the whole animal but proceeds at different rates in different regions. The lethal threshold diminishes, presumably to zero, in old animals but not in proportion to life expectancy throughout adult life; rather it changes more slowly at first and then more rapidly. Irreversibility of injury differs with different radiations. With x- or gamma-rays it appears to be a similar fraction with doses smaller than about 100 r, but increases with larger brief single doses. The data in general are not sufficiently extensive and accurate to test hypotheses critically. Over the past several years (1, 2, 3, 4, 5) I have discussed the adequacy of certain hypotheses to provide an empirical mathematical description of radiation injury and its effect on the duration of life. These hypotheses have been fairly successful in outlining a broad picture of radiation injury, in correlating many of the data and in suggesting critical experiments. It has become obvious, however, that they are deficient in some details and require amplification or revision. I propose at this time to discuss those changes in these hypotheses which appear to be necessary and also to point out some of the areas in which the data are inadequate to form the basis of quantitative correlations. The hypotheses in question are as follows: (a) The total injury produced by ionizing radiation is proportional to the dose. (b) This injury is reparable in part and irreparable in part. (c) Recovery from reparable injury occurs at a rate proportional to its magnitude. (d) In consequence of (a) and (b), irreparable injury accumulates in pro- portion to total dose. (e) Reparable and irreparable injury add in all proportions and death occurs when their sum attains a level which is proportional to the remaining life expectancy. The injury defined here is a latent form observable at present only in terms of additional radiation dose. With acute exposures this injury has largely disappeared in most species before the clinical syndrome of radiation sickness has fully developed. There is presumably a quantitative causal relationship * This paper is based on work performed under contract with the United States Atomic Energy Commission at the University of Rochester Atomic Energy Project, Rochester, New York. 331 332 H. A. Blair between this latent injury and the clinical syndrome, but this has not yet been established. The advantages of developing non-lethal methods for detecting latent injury will be mentioned later. It should also be remembered in what follows that the minimal lethal injury to an animal may not initially be manifest clinically at all, and that death occurs only after many days during which clinical signs develop. Because most mammals, if they are going to die from irradiation, do so within three or four weeks, it is customary to describe the lethal dose as that one which will kill one-half the experimental group within thirty days and to designate it LD50 or LD5Q 30 days. The events which happen between the time of exposure, when presumably a lethal threshold for primary injury must be reached, or exceeded, if death is to occur, and actual death, are outside the scope of this discussion. The LD50 for most mammals using whole body exposure is within the range of 400 to 800 roentgens for the young adult. In accord with the above hypotheses, the rate of development of injury / under exposure at constant dose rate y is f =^7-^(/-oc}'0 (1) in which /> is the rate of recovery per unit injury and A and a are constants. Integration of equation (1) gives for the level of injury after exposure for time t (^ — a) „, I = ^^—^y{\--e-f^^) + y.yt (2) If the time of exposure is sufficiently short that no significant recovery occurs during exposure, as is usual in determining the acute median lethal dose or LD50, e"''' may be replaced by 1 — /:»/ so that equation (2) becomes / = Ayt = Aot. a being the total dose. If, now, a is the LD50, the injury / is the lethal injury and according to postu- late (e) I=A(x^ So- S (3) in which 5*0 is the normal life expectancy of the animal and S is its age at radiation death. The constant of proportionality associated with Sq — S is taken arbitrarily as unity For animals irradiated at daily constant rates for periods of some months e~^^ may be neglected. This reduces equation (2) to . (A - a) r + ay/ (4) or, on using equation (3), to 5*0 — 5" A — a + a/ (5) Because nearly all chronic radiation experiments are begun on the young adult animal and also because postulate (e) catmot possibly be valid in very young animals in which the lethal dose rises instead of diminishes with age, A Quantitative Description of Latent Injury from Ionizing Radiation 333 it is convenient to measure Sq and S from the beginning of irradiation so that t in equation (5) is replaceable by 5* to give a /^ + OI.S (6) This equation represents existing (1, 2) data on chronic irradiation of mammals well within their possible errors. Such errors may be large in long- term experiments owing to infections and other accidents. An example of the fit is given in Fig. 1 for the data in Table I. IT bJ a. K cc tli a. o I- tr o X in 10 20 30 40 SURVIVAL TIME WEEKS 50 Fig. 1. Data by Henshaw (17) on chronic irradiation of mice plotted according to equation (6). The data are given in Table I. They cover a wider range than most. The scatter in such experiments frequently increases as the duration of the experiment gets long. Table I. Data by Henshaw (17) on Mice Irradiated Chronical/}' Five Days per Week until Death No. of animals Daily dose (r) Survival time (S) (weeks) So-S observed Total dose (r) So-S y y{A - a) (weeks) aa (weeks) So-S calculated 10 0 45.8 0 0 0 0 15 5 37.0 8.8 925 .352 2 6.2 8.2 15 10 34.6 11.2 1730 .224 4.1 11.7 15.8 15 15 27 18.8 2025 .250 6.1 13.7 19.8 14 20 23.8 22.0 2380 .220 8.1 16.1 24.2 14 25 19.3 26.5 2410 .212 10.1 16.3 26.4 10 40 11 34.8 2200 .174 16.2 14.9 31.1 10 60 7 38.8 2100 .129 24.3 12.1 36.4 10 80 4 41.8 1600 .105 32.4 10.8 43.1 It is obvious now, however, that this equation should fail, providing all the other postulates are valid, because of the inaccuracy of postulate (e) even in the region of adult ages. Actually this postulate could have been written: 'LD50 diminishes in proportion to life expectancy.' Consequently it can be tested directly by measuring LDjo as a function of life expectancy. In Fig. 2 are plotted LD50 data on Rochester rats (6) as a function of age. It will be seen that LD50 increases with age in young animals, is maximal in young adults, then declines slowly with age. As was mentioned above, postulate 334 H. A. Blair (e) could possibly apply only to the adult stage. It should be noted that this curve is not convertible into a LDgo-life expectancy relation because, owing to mortality among the animals, the sample at each succeeding age is different from those going before. Those dying early have the shortest life expectancies and presumably the lowest LDgo's, although this latter point cannot be proven directly. In the legend to Fig. 2 are also given the days of hfe expectancy for the adult data only. These are fairly linear but do not extrapolate to LD50 = 0, when Sq — S — 0, but to about 300 r. Presumably later points will diverge 800 £600 3 400 200 0 100 300 500 700 AGE IN DAYS Fig. 2. Median lethal dose in roentgens as a function of age in rats of the Rochester strain (6). The animals at different ages are not directly comparable because, for example, of a group selected at 100 days, only about two-thirds survive to 500 days. The actual median survival times of control animals for the groups irradiated at 5, 11 and 16 months, respectively, are 450, 375 and 330 days after the time of irradiation. Therefore, life expectancy does not decrease as rapidly as the age of selection increases. toward zero. Because it requires maintenance of animals for about three years to obtain data for a point at the advanced ages, it may be some time before the curves of Fig. 2 are well determined even in short-lived animals. However, Grahn and associates (7) at Argonne National Laboratory have shown in mice that the lethal dose as measured by repeated daily doses decreases rapidly from middle age with an apparent tendency toward zero at old age. At the present time it is not possible to state the situation more clearly than that the lethal threshold in the adult is some diminishing function of life expectancy, not a linear function throughout as required by postulate (e). This can be expressed also as LD50 = F(So - S) (7) and this as yet undetermined function should replace 5*0 — 5" in equation (6). However, there is considerable indirect evidence which will be discussed later that LD50 = k(So - S) (8) in fairly close approximation,/: being a constant for values of ^q — Sup to 20 per cent of 5*0. It is important to establish the form of equation (7) in several species so that estimates can be made of variation of LD50 with age in man. It is not clear whether equation (6) fits chronic data because equation (7) is sufficiently linear in the region in which most of the data lie (shortening of life span by one-half or less) or because of some other compensatory factor. In A Quantitative Description of Latent Injury from Ionizing Radiation 335 any case putting k from equation (8) equal to unity, as is done in equation (6), may modify the constants A and a. This possibility should be considered when comparing the numerical values of these constants in different species and considering their absolute values. The constants ft and clJA of equation (1) and their variations with age, if any, can be determined directly. According to equation (1) the injury /, when exposure is stopped, should be repaired exponentially between its initial value and its irreversible residual. This repair was first studied in mammals by Hagen and Simmons (8) using the rat. They assumed exponential repair to Radiotion Injury /^ ~- — -^^^^lethal threshold ( reversible ^^v injury \irreverslble injury^*^ N, AGE S So Fig. 3. A schematic representation of the LD50, or lethal injury threshold, for an individual animal with a representation of injury from a single exposure. Data discussed in the text indicate that the irreversible injury is the same, and remains constant independent of the adult age at which it is laid down. This threshold curve cannot be measured directly owing to the change by death of the sample as the age of selection is made older and older. The related curve which is measurable is that of LD50 as a function of remaining life expectancy. zero. Usually in making this determination a single substantial sub-lethal dose is given to a large group of animals followed by test doses to sub-groups at increasing intervals. Actually, the residual injury should be determined separately by testing one group after all repair presumably has taken place as illustrated in Fig. 3. The test doses less the residual in roentgens should then demonstrate simple exponential repair according to equation (1). However there is still another complicating factor in that it has been demonstrated recently that all parts of the animal do not recover at the same rate. Carsten and NooNAN (9), for example, have shown that following irradiation of the rat abdomen alone, recovery occurs with a half time between one and two days whereas it occurs in the animal with abdomen shielded with a half-time of about five days, and in the whole animal in about one week according to Hagen and Simmons (8). Possibly, however, the strain used by Carsten and Noonan would demonstrate whole-body recovery at the same rate as abdomen- shielded recovery. In any case we are confronted with the fact that partial body recovery, at least for some tissues, is different from that for whole body. We do not yet know whether recovery of the parts is independent of whether or not other parts have been irradiated. It is fairly certain, however, that faster abdominal recovery can occur following whole-body irradiation, and probably accounts for the several observations (10, 11, 12) that recovery during the first day, as measured by whole-body test doses, is considerably faster than backward exponential extrapolation of later recovery. Strictly then equation (1), in some cases, if not in all, should be written with two or more constants [> and the data analysed appropriately. It will be seen that the existence of more than one constant, /i, will not alter the form of equations (4) to (6) but the apparent value of ft determined from 336 H. A. Blair chronic data will have no exact counterpart in recovery measured directly by test doses. According to the hypotheses, recovery should not exceed the irreversible component ay/. The data confirm that at least after some months following exposure the irreversible component is demonstrable as a decrease of LD50 and it is ultimately demonstrable as a decrease in life span. Nevertheless, there are some data on recovery showing that in the first few weeks test doses for lethality may attain or even exceed values for animals not previously irradiated. The natural conclusion from these data is that recovery may be complete or even more than complete in that an apparent tolerance to radiation is developed. Owing to a number of factors, the nature of this apparent transient complete or over-recovery is not clear. One of the factors is that if the experi- ments are done on young animals which have not attained maximal LD50 (Fig. 2), increase of LD5Q during recovery will obviously make recovery appear greater than it really is. This defect may not be obviated by comparison with controls at each stage of the experiment, because LD50 may increase differently with age in the irradiated and control groups. Another disturbing factor is that fast recovery of the abdominal region will make the earlier part of the recovery curve fall faster than is appropriate to the remainder of the body and the later part of the recovery curve will be lower, because, after the abdomen has recovered considerably, the dose required to kill will be greater than it would be if the whole body were recovering together. This factor will tend to obscure an irreversible remainder until all recovery has proceeded as far as it will. Another possibility is that the animal may develop a transient physiological reaction to acute radiation injury which temporarily raises the lethal threshold for a second dose. For all these reasons the irreversibility of radiation injury probably cannot be evaluated properly until at least several weeks after a substantial dose The question of whether parameters in biological systems are age dependent should always be raised. In the case of recovery, for the reasons given above, evaluation of the constants or constant ^ is difficult by direct measurement. Nevertheless if the unanalysed recovery curve itself is similar at different ages this is an indication that the constants have not varied. Hursh and Casarett (13) have shown in the rat that the recovery curve at 546 days (beyond middle age) is similar to that at 107 days (young adult). More study should be given this problem, but at present there is no indication that the rate of recovery is age-dependent. The problems of whether irreversible injury is the same per unit injury at all ages, whether it slowly diminishes or increases, whether it gives rise to shortening of life because it is identical with ordinary aging or because it promotes ordinary aging, and whether it can be altered in any way, once laid down, are of considerable interest with respect to the setting of permissible levels for human exposure. If, for example, the irreversibility of radiation injury could be reduced the consequences of exposure would be reduced similarly. Referring to Fig. 3 the indications at present, though far from complete, suggest that irreversible injury once laid down remains at constant level, A Quantitative Description of Latent Injury from Ionizing Radiation 337 as depicted, until it intersects the curve of diminishing lethal threshold to cause the animal to die prematurely. One indication of this has been obtained by Baxter (14) in fruit flies. These flies normally live for about fifty days and lose half their life span if exposed to 75,000 r in a single dose. They die on day 26 approximately whether irradiated on day 1, day 25, or any day in between. Presumably recovery is very rapid in this species, and the irreversible component has the same eff'ect when laid down at any time which is early enough in life to allow the whole potential life-shortening to be made manifest. OSINGLE DOSE • DIVIDED DOSES 100 200 300 400 DOSE IN PERCENT OF LDsq-SO DAYS Fig. 4. Life shortening in per cent of normal span as a function of LD50 for rodents exposed to single doses or divided doses of x- or gamma-radiation. In the case of divided doses the radiation was stopped sufficiently long before death, or was at a sufficiently low daily level, that life shortening was caused only by irreversible injury, all, or nearly all, acute injury presumably having been repaired. The scatter of data is quite high for low divided doses, there being almost as many (omitted for simplicity) which show prolongation as shortening of life. The single dose curve rises more rapidly than linearly as LD50 is approached. The sources of the data are given in (16). LD50 is from 500 to 700 r for most of these strains. There is no established reason why data from different species should form a consistent pattern in this mode of plotting. They are less consistent than data on single strains. Incomplete observations by Hursh and Casarett (13) indicate that a given dose shortens life by about the same fraction of the normal expectancy in groups of rats exposed in early adult life or beyond middle age. Direct measurements of residual injury as reduction in LD50 have been made no later than a few months after an initial dose. Such direct determinations when extended will be a more satisfactory test of the validity of the hypotheses depicted in Fig. 3 than the life-span data mentioned above. Another factor to be discussed is whether the irreversible injury, or the constant a, is independent of dosage. It appears definitely to be larger with fast neutrons and alpha rays than with x- or gamma-rays (3). As measured by life-span shortening, it is also greater for single substantial doses of x- or gamma-rays than for divided doses even though all are delivered at the same dose rate in roentgens per minute. Figure 4 shows the after eff'ects of single and divided doses on life span in a number of strains of rats and mice. These data are plotted on the assump- tion that strains of different life spans and different LDjo's will lose the same fraction of their life spans per unit dose measured in LD5Q. Existing data 338 H. A. Blair are not sufficiently accurate to decide whether this assumption is more correct than the one that the effects per roentgen are more similar in going from strain to strain or species to species. It will be observed that according to Fig. 4 divided doses cause only about one-third the life-shortening per unit dose as that produced by single doses. Because some of the divided doses were given in increments as great as 120 r and because existing data are not sufficiently accurate to define small effects, it is probable that the curve for the smaller single doses coincides with that for multiple doses. It is certain, however, that substantial single doses such as 200 r (one-third LD50) or more, have considerably more effect than the same dose in smaller increments. The reason for this difference that immediately suggests itself, is that the irreversibility of the injury is some increasing function of its magnitude rather than the linear function assumed here. That this is not the correct explanation is indicated by the fact that repeated daily doses calculated to produce as much injury of the type defined here as a single substantial dose do not have the same effect on life span. There may be some unidentified dose dependent concomitant of injury which affects its reversibility. At this time, however, all that can be said is that a appears to be a constant independent of dose for doses of daily increments up to about 100 r but that it increases with dose with greater daily doses. That this larger effect of substantial single doses occurs at the time of irradiation and is not due to a dose dependent subsequent develop- ment is indicated by a single set of data (13). Such observations should be extended. As predicted, the multiple dose curve of Fig. 4 is probably nearly linear. The single dose curve increases more rapidly than linearly if carried to higher doses than those depicted. This is to be expected because, according to the hypotheses, life shortening will be linear with dose only to the extent that the threshold curve of Fig. 3 is linear. As irreversible injury becomes substantial it will have more effect on life span per unit magnitude according to this curve. YocKEY (15) has postulated the identity of radiation damage with reduction of somatic genetic information, and has related the present formulation to the consequences of such damage in tenns of information theory. CONCLUSIONS The hypotheses used appear to give a fairly accurate over-all description of radiation injury. The only one which is definitely known to be inaccurate is the last, which probably should be restated: Reparable and irreparable injury add in all proportions and death occurs when their sum attains a level which is some function, not fully detennined, of the remaining life-expectancy. Certain details, such as recovery rates, probably must be regarded as tissue- or region-specific rather than whole-body specific. This may also be true of irreversibility which has not been systematically studied in this regard. This latter problem is of particular interest with respect to human exposure, much of which, especially from internal emitters, is partial-body. However, even if each tissue, for complete description, requires a different set of constants A, a and ^, this adds only complexity of detail and not of concept. A Quantitative Description of Latent Injury from Ionizing Radiation 339 Irreversible radiation injury has the special property that it is closely related to premature aging abruptly laid down and probably persisting thereafter at a level constant or nearly so. This suggests the possibility that premature aging may be studied in young animals without waiting for them to die naturally. Of special interest is the possibility that irreversible injury may be prevented, at least in part, or altered once it has been laid down. This possibility should be studied in relation to exposure problems in man and also with respect to its bearing on natural aging. If, however, irreversible injury is wholly in the form of somatic mutations, as is often suggested, the possibility of altering it or its consequences would presumably be remote. The acute injury described here in terms of radiation dose has antecedents in the form of disturbances of cellular structure and function from absorbed radiation and consequences in the form of the clinical syndrome of radiation sickness. Only the last stage has been at all well described in physiological terms, and the connections between the stages has not been elucidated at all. The ability to measure latent injury in terms of radiation dose should assist in deriving its description in biochemical or physiological terms. This is also true of irreversible injury. Nearly all aspects of the long-term effects of radiation injury are markedly deficient in data, especially of those based on sufficient numbers of animals to be reasonably exact. For this reason no formulation of the kinetics of the injury process can be adequately tested at present for its quantitative exactness. The virtue of a particular scheme is measurable rather in its ability to designate the phenomena involved, to make useful predictions and to serve as a basis for designing critical experiments. REFERENCES 1. H. A. Blair: A formulation of the injury, life span, dose relations for ionizing radiations. I. Application to the mouse. University of Rochester Report UR-206 (1952). 2. H. A. Blair: A formulation of the injury, life span, dose relations for ionizing radiations. II. Application to the guinea pig, rat and dog. University of Rochester Report UR-207 (1952). 3. H. A. Blair: The shortening of life span by injected radium, polonium and plutonium. University of Rochester Report UR-274 (1953). 4. H. A. Blair: Recovery from radiation injury in mice and its effects on LD50 for durations of exposure up to several weeks. University of Rochester Report UR-312 (1954). 5. H. A. Blair: A formulation of the relation between radiation dose and shortening of life span. Proc. Int. Conf. Peaceful Uses of Atomic Energy 11, 118-120, United Nations, New York (1956). 6. J. B. Hursh, and George W. Casarett: The lethal effect of acute x-irradiation on rats as a function of age. Brit. J. Radiol. 29, 169-171 (1956). 7. D. Grahn: Personal communication. 8. C. W. Hagen Jr., and E. L. Simmons: Effects of total body irradiation on rats. Part 1, Lethal action of single, paired and periodic doses. University of Chicago Metallurgical Laboratory Report CH-3S15 (1941). 9. A. L. Carsten, and T. R. Noonan: Determination of the recovery from lethal effects of lower body irradiation in rats. University of Rochester Report UR-455 (1956). 10. J. F. Thomson, and W. W. Tourtellotte: Effect of dose rate on LD50 of mice exposed to gamma radiation from cobalt 60 sources. Anier. J. Roentgenol. 69, 826-829 (1953). 340 H. A. Blair 11. J. Storer: Personal communication. 12. J. P. Storaasli, S. Rosenberg, J. S. Krohmer, and H. L. Friedell: Effect of short interval fractionation on the lethal properties of total body radiation in rats. Western Reserve University Atomic Energy Medical Research Project Report NYO 4919 (1955). 13. J. B. HuRSH, and G. W. Casarett: Unpublished observations. 14. R. Baxter: Unpublished observations. 15. H. P. Yockey: An application of information theory to the physics of tissue damage. Radiat. Res. 5, 146-155 (1956). 16. H.A.Blair: Data pertaining to shortening of life span by ionizing radiation. University of Rochester Report UR 442 (1956). 17. P. S. Henshaw: Experimental roentgen injury. IV. Effects of repeated small doses of x-rays on blood picture, tissue morphology and life span in mice. J. Nat. Cancer Inst. 4, 513-522(1944). SOME NOTES ON AGING Hardin B. Jones Division of Medical Physics and Donner Laboratory University of California, Berkeley, California Abstract — Evidence of physiologic change with age uniformly points to a cumulative deteriora- tion as age increases. Further degenerative change may occur proportionally to the amount of change already acquired. As age increases, incidence of degenerative disease and death increases exponentially. It is pointed out that, whatever aspect of body function is considered, e.g. functional members, metabolism, cellular activity, or blood flow characteristics, a relatively exponential increase in degeneration of body function occurs with increasing age of the individual. It is possible that each of these separately considered systems of aging is in partial equilibrium with the others, so that all general characteristics of change in functional vigor with time follow a similar course. Increments of change in body structure and function occur as a phenomenon of aging. Usually, the term 'aging' is associated with deteriorative change, and as such is distinctively set off from those changes with age that are respon- sible for growth and development. However, even the period of development may be considered to have associated with every step some hazard that this step may not be achieved fully, thus adding an increment of imperfect function to the body. Such a deletion from full function, whether arising from genetic inheritance, developmental processes, or accidental mishap, may count just as much toward the accumulated deterioration we can manage to tolerate as does the deterioration of advanced age. Experience of mishap accumulates throughout life. Some events, to be sure, have as little residual effect upon us as the whistle of the wind, but occasion- ally something of consequence occurs. As an example, it may be the crushing of a finger; although we usually recover, we can remember the event because of some persistent change^perhaps a scar, or a distortion of the nail, or even the loss of the finger. Since, on the average, we live each day in a situation where there is some definite but slight chance that an event of misfortune may disturb us, then the longer we live under this average circumstance of risk, the more likely we are to find among us individuals showing physical impairment. Inspection of such a system leads to the probable conclusion that: Accumulated impairment = Mishap risk x Time of exposure X Fraction of function lost per mishap. But, the risk of occurrence of an unfavorable event is subject to increase as age increases, and the fraction of function lost per mishap may also increase as age increases. In this system, therefore, we can expect a relatively non-lincrr accumulation of average physical impairment as age increases; physical impair- ment may increase as some higher power of time hved than unity. This example 341 23 342 Hardin B. Jones of a system contributing to aging reflects the exponential increase of morbidity and mortality that regularly is observed with increasing age. There are other examples of impairment of body function that depend upon time lived and upon morbidity experience. A very general theory of impairment can be argued in which morbidity leads to morbidity, and mortality risk is some function of the integrated morbidity experience (la). The following examples of functional disturbances may be cited to illustrate such relationship between morbidity and morbidity and between morbidity and mortality: (a) The severity of toxic reaction usually increases more than proportionately to the poison dose. (b) Radiation exposure induces ionization in tissues, and this morbidity in turn can induce morbidity proportional to the dosage. This holds both for acute effects and for life-span and carcinogenic changes. (c) Risks of degenerative vascular disease are proportional to the extent of obesity (5). (d) Risks of degenerative vascular disease are proportional to the dis- turbances of serum lipids in individuals followed over a segment of the adult life span. (e) Death risks in diabetes throughout the past forty years have been undergoing a progressive reduction apparently proportional to the goodness of diabetic control. (f) Dimming of primary senses (vision, touch, pain, and hearing) is associated with enhanced risks of trauma. (g) Susceptibility to infectious disease is believed to be directly propor- tional to exposure intensity, and inversely proportional to defense mechanisms such as antibody levels and antibody generating capacity (lb); also, susceptibility to infectious disease can be quantitatively off'set by administration of antibiotic agents. (h) Proportional differences in death-rate risks among population samples throughout life span are related to sums of environmental and genetic factors. Having noted examples of how morbidity and mortahty risk can be depen- dent upon functional impairment, we can consider in greater detail evidence pointing to a widespread interdependence of physiologic systems. In vascular disease, occlusion may directly diminish blood flow in a small but critical segment of the body, as in coronary thrombosis. However, even though there is a measure of recovery from the acute episode, there may be a generalized insufflciency of circulatory function. Changes in blood flow caused by narrowing of the arterial channels may be expected to exact an effect upon function of the extremities, and Dobson (2) has recently shown evidence for general dependence of the body's homeostatic mechanisms upon the proportional balance of regional blood flow. Thus, especially for the circulatory system, we can be certain that functional changes can influence the entire quality of body function. A similar example of interdependence of disease is in the complications Some Notes on Aging 343 of diabetes mellitus. This disease is not limited to the classic confines of its relationship to carbohydrate and intermediary metabolism: serious disturbances of lipid metabolism may also occur, linked with enhanced tendency for vascular changes; the term of pregnancy is frequently lengthened in diabetic mothers retinal changes may occur in diabetics, and the disease in general may be associated with somewhat early changes related to aging. There seems to be no reason to suspect that diabetes is a more complicated disease fundamentally than loss of islet-cell function or absence of insulin; but it does seem that the results of this functional deficiency can produce several different conditions that may even interact to compound the pathologic impact of the basic deficiency. Another example of general disease being associated with a specific disease is observed in the follow-up of cancer patients. In cancer of the rectum, death from intercurrent disease may be just as likely as death from recurrence of the mahgnancy. There is also general evidence, from comparisons of mortality from disease in nineteen western countries, that high incidence of any one kind of disease is associated with high incidence of other diseases (la). Some factors affecting adult health and life expectancy might be expected to be common to several kinds of overt disease; other factors influencing health may have a limited efiTect upon a single system. For example, in overweight individuals the increased risk of death is attributed to increased incidence of arterio- sclerosis and hypertensive disease, while the tendency toward cancer is not significantly changed from the average of the population. In radiation exposure, all major diseases may be enhanced. Leukemia, however, may be increased by a factor of 10, while other degenerative diseases are elevated less than twice. It is quite possible that some kinds of disease are less likely to occur following radiation exposure, even though the general trend is toward more severe and earlier degenerative disease following significant radiation exposure. Similarly, smoking generally enhances degenerative disease by a factor of 2 while lung cancer is increased tenfold. These observations point to the interrelationships in etiologic factors in disease, and the possibility that causative factors in development of degenerative disease may have interactions that accelerate the appearance and consequences of disease change. Vascular Disease GoFMAN and associates (3) have been able to show that the change in the wall of the artery in arteriosclerosis is essentially a linear thickening throughout aging. Thus, the shift toward occlusive change results from narrowing of a cylindrical tube by a progressive thickening of the mass, reducing the radius of the lumen. The function describing the reduction of blood flow in the artery involves the cross-sectional area of the artery, which is proportional to the square of the radius of the artery. Since blood flow in the artery is related to cross-sectional area, blood flow changes in arteriosclerosis are not proportional to time lived but rather vary as a power function of time. The fact that elasticity of the artery may fall off" sharply as sclerotic thickening occurs probably accelerates the process. Thus, from several points of view, vascular change is not likely to produce a linear accumulation of disturbance with time lived, even though the basic feature of the disease is reasonably 344 Hardin B. Jones established as a thickening of the artery wall proportional to lipoprotein elevation and duration of the condition. Since vascular disease is a large component of degenerative disease and a con- tributing cause to other diseases, it is quite possible that exponentially-decUning blood-flow capacity may in part determine the exponential pattern of increas- ing incidence of overt disease other than vascular disease. Cancer and Aging Throughout adult life, cancer incidence and cancer death rate are increasing exponentially; in most ways, this increase is remarkably similar to the above- described increase in heart disease tendency. Armitage and Doll ('4) have ascribed this property to the fact that a succession of small changes necessarily precedes cancer. It is of interest to construct population samples of individuals known to have died of a given kind of cancer. When this is done, the increase in death rate in the cancer-destined population is remarkably like the increase in incidence of cancer in the population out of which it was taken (la). Thus, we can be reasonably certain that the risk of cancer is increasing exponentially with age. In contrast to the exponentially-increasing incidence of cancer with increas- ing age, individuals identified as having overt cancer have a constant death risk approximately independent of chronologic age. Therefore, it is a reason- able argument that changes characterizing the period prior to onset of cancer may be of many different kinds, each making cancer occurrence more likely, but the change representing incidence of cancer effects a single abrupt decrease in life expectancy. It follows from this reasoning that many of the changes that accompany aging may be of consequence only as they allow a drastic and irreversible change into overt disease to take place. In vascular disease, the average degenerative change in the walls of the artery is of less consequence than the infarctions or vascular occlusive episodes that destroy peripheral tissue. Death may occur as a consequence of a random occlusive event, even though average changes in the arterial structure may be minimal. Cellular Change and Aging Cancer is usually considered to be an example of cellular change associated with aging, very possibly upon a basis of somatic mutational change. It should be noted that evidence for this is based upon an incidence of cancer expon- entially increasing with age. While I, too, subscribe to this view, a similar phenomenon is seen in diabetes melHtus, a disease of deletion of function. It is quite possible that, in addition to changes in the quahty of cells surviving Vv'ith time, certain kinds of cells may survive aging with different likehhood. Shock (6) has evidence, for example, for a decline both in functional quality and numbers of cells in the kidney with age. It is reasonable to explore further the effects of dechning numbers of cells with age. Instances, as in the case of disappearance of islet tissue in diabetes, may be observed in various tissues and are represented by epilation, appearance of channels in the fingernails, and disappearance of cells supplying sensory function of various kinds. These cells may disappear, but we do not know why. Some Notes on Aging 345 In radiation effect, radiation exposure is related directly to enhancement of degenerative change, thus simulating the effect of aging. The similarity may be due in part to the random destruction of cells and partly to the alteration of function of cells. Within certain cells such as the marrow and the lymphatic tissue, or in embryologic development, radiation over a wide range of dose and for several species of mammals destroys approximately three cells out of every 1000 cells per roentgen of whole-body exposure (7c). At less than lethal exposures, this random destruction of cells proportionally to radiation exposure does not have a lasting effect upon the blood-forming tissues, since these cells rapidly regenerate. However, the average lethal dose of whole-body radiation exposure is estimated to involve a 50 per cent reduction in these cells. Some- what the same changes occur in other body cells, the degree being dependent upon radiation sensitivity. (Some cells are known to be much more resistant to radiation than blood-forming cells.) The effects of radiation in diminishing the numbers of cells also seem to be about the same upon mammalian germinal cells as on blood-forming tissues; in both tissues, approximately two to three cells are affected per 1000 cells per roentgen. Thus it appears that each roentgen of exposure to tissues like the blood-forming system, the gonads, and the develop- ing embryo may have about equal probabiUty of either kilhng the cell directly or altering its chromosomal structure if it survives. Such changes may be sus- pected as having a role in inducing age change in the somatic tissues. Leukemia induction by radiation, as evidenced from the analysis of Court- Brown and Doll and others (7a,b,c,d), is increased proportionally to radiation exposure. These changes are such that approximately 50 r of whole-body exposure produces a frequency of leukemia equal to its natural incidence and the effect is proportional to dose over a wide range. This is in remarkable agreement with genetic change in mammals ; here, too, an exposure of 50 r produces approxi- mately the same number of mutations as occur naturally in one generation. REFERENCES la. H. B. Jones: A special consideration of the aging process, disease, and life expectancy. In: Advances in Biological and Medical Physics, 4, 281-337, ed. by J. H. Lawerence and C. A. Tobias, Academic Press, Inc., New York (1956). lb. G. A. Sacher: Reference to aging and antibody levels. Donner Laboratory Lecture, University of California, Feb., 1957. 2. E. L. Dobson: Homeostatic regulation of body fluid volumes: Role of a blood flow dependent control mechanism. Fed. Proc. 16, 31 (1957). 3. J. W. Gofman: Serum lipoproteins and the evaluation of atherosclerosis. In: Experi- mental methods for the evaluation of drugs in various diseases. Ann. N. Y. Acad. Sci. 64, Art. 4, 590-595 (1956). 4. P. Armitage and R. Doll: The age distribution of cancer and a multistage theory of carcinogenesis. Brit. J. Cancer 8, 1-12 (1954). 5. L. I. Dublin and H. A. Marks: Mortality among insured overweights in recent years. 6th Annual Meeting, Association of Life Insurance Medical Directors. MetropoUtan Life Insurance Company, October 1951. 6. N. Shock: Physiological aspects of aging. Symposium on Aging, Amer. Assoc. Advanc. Sci. Meeting, New York, Dec. 1956. 7a. W. M. Court-Brown and R. Doll: The incidence of leukaemia among the survivors of the atomic bomb explosion at Hiroshima and Nagasaki. In: The Hazards to Man of Nuclear and Allied Radiations, 84-87, Medical Research Council, London. (1956). 346 Hardin B. Jones 7b. H. B. Jones: A summary and evaluation of the problem with reference to humans of radioactive fallout from nuclear detonations. In: Hearings before the Special Sub- committee on Radiation of the Joint Committee on Atomic Energy, Congress of the U.S., June 4-7, 1957, 1116-1136 (also discussion, 1100-1108), Government Printing Office, Washington (1957). 7c. H. B. Jones: Estimation of effect of radiation upon human health and life span. Proc. Health Physics Soc. 1, 114-126 (1956). 7d. E. B. Lewis: Leukemia and ionizing radiation. Science 125, 965-974 (1957). CANCER AS A SPECIAL CASE OF A GENERAL DEGENERATIVE PROCESS* Harry Auerbach Division of Biological and Medical Research, Argonne National Laboratory, Lemont, Illinois Abstract — Death rates or life table q^^ values for populations throughout the world tend to exhibit a sixth power linear relationship with age when plotted on a log-log basis. It is shown that the total deaths can be broadly separated into chronic degenerative causes and acute causes, with death from the degenerative causes increasing as the sixth power of age and death from the acute causes increasing in simple exponential fashion. In many cancer studies at this laboratory and elsewhere, attempts have been made to discover the underlying mechanism by which tumors come into exis- tence. While a great many of these studies have been directed toward describing the process of carcinogenesis in biological terms, the statistical approach has also been productive. A recent study at Argonne National Laboratory, using analysis of vital statistics, indicates that cancer has some characteristics in common with the degenerative diseases. The study was suggested by observations of others (1, 2, 3) that when the logarithm of death rate from cancer (either the total or that involving a specific site) was plotted against the logarithm of age at death, the result was usually a straight line. The slope of the line indicated a sixth power relationship, a fact that has been used to support several theories of carcinogenesis. Another interesting possibility — that the same linear relationship might be present in other causes of death — was recognized and investigated in the present study. The question was first examined by analyzing the United States death rates for the years 1949-1951. Plots were made on the same log-log basis for several broad groups of causes of death. Five groups (circulatory system, mahgnant neoplasms, nervous system and sense organs, respiratory system, and genito- urinary system) showed a relationship of approximately the sixth power of age to a marked degree for age thirty and older, with departures from linearity being restricted to ages under thirty. The sum of these groups gave an almost perfect linear relationship from the age of thirty upwards (Fig. 1). These five groups represent the overwhelming majority of the chronic degenerative causes of death. The remaining three (infective and parasitic diseases, digestive system, and accidents) which did not show the linear relationship, represent the acute causes of death. In order to find out whether the same situation obtained in other countries, a slightly different method had to be used. Life table data, which are available for most of the countries of the world, were used in the absence of reliable specific cause death rate statistics. The value used was q^, the proportion of * Work performed under the auspices of the U.S. Atomic Energy Commission. 347 348 Harry Auerbach persons alive at the beginning of the year of the specified age, who die during that year of all causes. The countries tested were United States, Canada, Israel (Jewish population), India, Union of South Africa (Asian population), Brazil, Japan, Portugal, Belgian Congo (African population), Costa Rica, El Salvador, Argentina, Ceylon, Finland, France and Norway. Similar sixth power relationships were exhibited by all. Deviations from linearity at 10 000 r 5000 - 2000 - 1000 r ^___ o o o o o a: UJ a. < < UJ o 500 200 - 100 - 12.5 17.5 22.5 27.5 32.5 42.5 52.5 62.5 72.5 82.5 37.5 47.5 57.5 675 775 675 AGE Fig. 1 . Log-log plots showing the relationship of death rate to age in United States white males (1949-1951) for broad groups of causes of death. Solid lines, chronic causes ; dotted lines, acute causes. younger ages were always in the direction of the actual figures being higher than the extrapolated values. On the basis of the demonstrated log-log sixth power Hnear plot of the chronic degenerative diseases, and the nonlinearity of the acute causes of death, an attempt was made to determine the relationship of acute causes to chronic degenerative causes in the total death rate. Life table values of q^ for United States wliite males for three periods, 1900-1902, 1929-1931, and 1949-1951, were plotted against age on the log-log basis. The usual departures from linearity at earher ages were marked in the 1900-1902 period, less so in the 1929-1931 period, and still less in the 1949-1951 period, but all three curves tended to merge into a common straight sixth power line at the age of forty Cancer as a Special Case of a General Degenerative Process 349 and older (Fig. 2). This merging of the three plots at the age of forty and older demonstrates a well-known fact, that chronic degenerative diseases become most important as cause of death at the older ages; the values show essentially no change over the period from 1900-1951. On the other hand, at the earlier ages, the acute causes constitute almost 100 per cent of the total death rate, and chronic degenerative diseases represent only a minor fraction. Even though 3UU - I □ 1 1 I 1 1 1 I 1900-1902 O 1929-1931 n 1949-1951 n- 200 - X X e 100 ~ x_ 8 - X 8 - 50 - X — 8 - X K or — 8 " X o o 20 - D O o DO 10 — ° X - D o" D - D ° Ox : 5 - o - Ox - 0 X 2 X 1 X *■ 1 1 1 1 1 1 J 20 40 60 80 100 AGE Fig. 2. Log-log plot of 1000^^ against age for United States white males for 1900-1902, 1929-1931, and 1949-1951. Lines are omitted in order to show more clearly the good fit of the three sets of points on the straight portion of the curve. it is apparent that acute causes have been decreasing steadily over the fifty-year period from 1900-1951, their presence is evident as departures from the straight line in the plots, with highest values at 1900-1902 and lowest at 1949-1951. The equation of the straight line for the forty years and older group was calculated on the assumption that the 1949-1951 values represented the least effect of acute causes of death on the over-all figures. This was regarded as giving q^ values for the degenerative causes of death which were then subtracted from the total q^ values for other countries. The resulting numbers were assumed to represent death rate attributable to acute causes of death. The ^a,'s associated with these acute causes of death were found to plot as a simple exponential increase with age. The slopes were equal for the countries tested, but the intercepts were different and correlated in a general way with the levels of public health and medical care in the country concerned. Plots for four of the countries are shown in Fig. 3a. When the sum of the five major groups of degenerative diseases was similarly subtracted from the total causes of death in United States white males 1949-1951, leaving a residue 350 Harry Auerbach representing the acute causes of death, a similar exponential increase with age was apparent (Fig. 3b). It therefore appears that the death process in man can be separated broadly o o o 200 — r ~1 1 1 r ■ 1 ' 1 1 1 ' 1 1 — Mauritius (1942-1946) — India (1941-1950) ^ — FnrmnQn flQ'^fi_l94n^ ^ 100 ^■"■^ ruiriiuau \\^J^~\^*^^] ^ — 70 - -- Trinidod (1945-1947) ♦^ ,/* . •/ ^' ^^ 50 — 30 20 — y' -'j;>V-' 10 — - ^ ,•''' : ~ ^ 7 — - ^'' 5 — 1 , 1 , 1 , L. J — 1,1, 20 (A) 30 40 50 60 AGE 70 80 90 2000 (B) Fig. 3. a. Typical semi-log plots of acute portions of 1000^^ (total minus chronic), for four of the countries tested, b. Semi-log plot of death rate per 100,000 population from acute causes of death (total minus chronic), for United States white males, 1949-1951. into chronic degenerative causes and acute causes, with death from the degenera- tive causes increasing as the sixth power of age and death from the acute causes increasing in simple exponential fashion. Total and specific causes of death have previously been fitted by the Gom- pertzian or semi-log plot (4). However, it would appear from this work that the degenerative causes increase with age not at a constant rate of increase as predicted by the Gompertzian, and are therefore better fitted by the log-log Cancer as a Special Case of a General Degenerative Process 351 plot. On the other hand, the acute causes of death clearly are fitted by the Gompertzian function. The presence of the sixth power relationship in a large number of different situations suggests a general underlying principle. Since we have no knowledge whatever of what this principle is in biological terms, we can only speculate that it could be a very general organizational scheme which provides about five redundant elements within each essential unit. An element might be a molecule, a cell or organelle (internal structural and functional unit of a cell), a group of cells, or a whole organ. The essential units might be separate or overlapping. Carcinogenesis may be a special case of this general degenerative process. REFERENCES 1. J. C. Fisher and J. H. Holloman: A hypothesis for the origin of cancer foci. Cancer 4 916-918 (1951). 2. C. O. Nordling: A new theory of the cancer inducing mechanism. Brit. J. Cancer 7, 68-72 (1953). 3. P. Armitage and R. Doll: The age distribution of cancer and a multi-stage theory of carcinogenesis. Brit. J. Cancer 8, 1-12 (1955). 4. H. B. Jones: A special consideration of the aging process, disease and life expectancy. In: Advances in Biological and Medical Physics 4, 281-337, ed. by J. H. Lawrence and C. A. Tobias, Academic Press, New York (1956). DISCUSSION Quastler: Two ditTerent functions have been proposed as representative of the course of the Gompertz function, G{t), for the later part of the life span: Gi(0 = fli + b^t Gi{t) = ^2 + ^2 In /. No author claims that either function is a perfect fit even for a limited interval. Still, it is worth showing that the difference between the two formulae is very small over a limited range. Let the center of this range be /*; then and for small values of A//r*, AGi = ±^1 Ar .c.^*.,n(q^') b. AC, ±^A' It is said that the mortality rate (in the later part of the life span) doubles about every 8.5 years; hence b^ ■--= 0.082; and that it increases approximately with the 5.2th power of age, or b., = 5.2. These two values are compatible around /* = 63 years, which characterizes the neighborhood in which both are claimed to be valid. Yockey: If one plots survival data as Auerbach does, one obtains curves which corre- spond to the Gompertz function for man and for many out bred wild-type organisms. For some in bred strains, particularly those which have a genetic defect, the survival curve may be of the form log ///q = — oc/'-. In Fig. 3 of my paper in Part V, I have plotted log ///o against the square of the age for several strains of mice. The dilute brown strain reported by Murray and Hoffman follows the 352 Harry Auerbach above equation quite closely for almost the entire life span, excluding only the first few months. On the other hand the DBFi hybrids (dilute brown female x C57 male) exhibit the Gompertz function as may be seen in Fig. 5 of that paper. The dilute brown strain is characterized by a high rate of mammary cancer, while the hybrid has a low rate. The Marsh albino is another high-cancer-rate strain, which, although it does not have a survival curve of the form log ///o = —a.X^, does, when crossed with the C57, produce hybrids with a much longer life span. The survival curve is of the Gompvertz type. Changes in the genetic characteristics associated with hybridization do not just change the constants of an equation of the Gompertz form, but rather the survivorship curve is of a different form. FREE RADICALS AS A POSSIBLE CAUSE OF MUTATIONS AND CANCER* Walter Gordy Department of Physics, Duke University, Durham, North Carolina Abstract — The hypothesis set forth in this note is that free radicals produced outside the body may find their way into the body and produce mutations and/or cancer. The evidence for support of this hypothesis is the presence of radicals as detected by microwave paramagnetic resonance in several carcinogenic agents, and the fact that free radicals are now recognized by radiobiologists as being responsible for a large portion of mutagenic and carcinogenic effects of ionizing radiations. Free radicals may be loosely defined as molecular fragments which are charac- terized by a free valence or an unpaired electron. Because of their highly reactive nature they are not thought to exist in any significant quantity within the organic matter about us, although they are often postulated as important, transient intermediaries in organic and biochemical reactions. Within the past few years, however, microwave spectroscopists (I) have shown that free radicals can be readily detected in organic matter which has been subjected to some form of pre-treatment that can break chemical bonds. Such free radicals are produced in the combustion of organic matter — wood, paper, tobacco, coal, oil. They are produced in excessively cooked foods such as charred steak or scorched toast. They are produced in various forms of matter by ultraviolet light, by x-rays, or by atomic radiation. The radicals are detected through their resonant absorption of microwave or radio-wave energy when they are placed in a magnetic field of the proper strength. This type of absorption spectrum is known as paramagnetic resonance or as electron spin resonance (2). Electrons in normal chemical bonds are paired in such a manner that their spins and magnetic moments cancel, and hence they exhibit no paramagnetic absorption. Paramagnetic resonance occurs only for the unpaired electrons of the disrupted chemical bond. It therefore provides a specific and powerful means of detecting and studying reactive free radicals within organic matter without interfering absorption or confusing signals from the normal stable molecules of the matter. The surprising new evidence from paramagnetic resonance is not that free radicals can be easily produced but that they become trapped and stabiUzed and can be transported from place to place, even through the air within tiny particles of soUd matter such as those in smoke. The nature of neither the radicals nor their cages is yet known definitely, although some radicals produced * This research was supported by the United States Air Force through the Air Force Office of Scientific Research of the Air Research and Development Command under contract No. AF18(600)-497. Reproduction in whole or in part is permitted for any purpose of the United States Government. 353 354 Walter Gordy in amino acids and proteins by x-irradiation have been tentatively identified from the fine structure of their microwave resonance patterns (3). The infor- mation pertinent to the present discussion is that organic radicals produced by physical forces such as heat or irradiation outside the body can be taken into the body through the processes of eating, smoking, or normal breathing, or even by diffusion through the skin. Once inside the body, these radicals may themselves penetrate the cells or they may be converted to other radicals which do so. A radical containing an odd number of electrons must, in effect, meet and react with another radical before its free valence or uncancelled electronic moment is nullified. If it reacts with a normal organic molecule (which has an even number of electrons), another radical is produced. In fact, it is just this odd character which suggests that a lone radical might start a significant chain of events within a cell. Many types of radicals which have been detected by microwave resonance are stabilized mainly within solid particles of matter. Normal chewing and mixing of food with saliva would tend to destroy them. This destruction may not always be complete, however. We have made tests which show that ordinary chewing of charred toast, beef, and other foods does not entirely kill the reso- nance signal of the radicals. Extremely small solid particles carrying radicals may diffuse into the tissues of the skin, stomach, or lungs where they would gradually dissolve and perhaps bring about damaging reactions as their radicals are released. Furthermore, these radicals are possibly stable in certain organic solvents which dissolve the solid cages and 'float' the individual radicals into the tissue. Such a solvent might assist in the production of cancer without being a primary cause of it. Strong resonances, like that shown in Fig. 1 for tobacco tar, are found for wood tar, coal tar, and other tars. H. Shields and the author have dissolved tars in organic solvents including benzene, acetone, and croton oil, and have found that the resonance of the tar radical remained strong. The role of agents such as croton oil, which are not themselves carcino- genic agents but which augment the effects of certain carcinogenic agents, may be that of facilitating the entrance of carcinogenic radicals into the body. Radiobiology experiments (4) indicate that much of the effect of ionizing radiations on the cells themselves may be indirect; that is, irradiation produces a free radical in one part of the cell which later migrates to a more vital part of the cell where it may react to bring about a mutation. Alternately, the first radical formed may react to form a second radical, or a third, which finally causes the mutation. In particular, OH and OOH radicals have been postulated as important intermediaries in radiation damage. Of course a mutation might be brought about by a so-called direct hit, but indirect effects also appear to have significant consequences. We are proposing an extension of the indirect effects to include cases where the primary irradiation occurs entirely outside the injured body, in our laboratory, microwave evidence has been obtained to indicate that hydrocarbon radicals, R, produced by irradiation, are often converted to peroxide radicals, ROO, where they come in contact with oxygen. In the tissue such radicals might be further converted to the OH or OOH radicals, already under suspicion by radiobiologists. The striking evidence which prompted this communication is the abundant paramagnetic resonance data for the existence of free radicals in many agents y -.-X^ TOBACCO TAR ~ *•■ 'A .V ■ CHIMNEY SOOT J SCORCHED BREAD X-IRRADIATED BREAD Fig. 1. Microwave electron spin resonances of radicals in some common substances. The tobacco tar was taken from an old pipe stem. Coal, wood, and other tars give similar resonances. The chimney soot was taken from the flue of an oil-burning furnace. Similar resonances were obtained for soot taken from the exhaust pipe of an automobile and from a wood-burning fireplace. Ordinary bread, unscorched and not irradiated, gave no detectable resonance in the same spectrometer. to face p. 354 Free Radicals as a Possible Cause of Mutations and Cancer 355 known or suspected to cause cancer. Among these are cigarette smoke, tobacco tars, various other tars, exhaust fumes from cars, smoke from home furnaces or industrial plants, and charred foods. I shall not attempt to cite the literature references for the various evidences that these are carcinogenic agents. It is well known that x-rays and other ionizing radiations can cause genetic mutations and are likewise carcinogenic agents. It is now well known from electron spin resonance that these ionizing radiations also produce radicals which in many biochemical solids (3) (including various proteins, carbohydrates, and fats) persist for long periods after the irradiation. The carcinogenic effects of severe chemicals which produce burns of the flesh may possibly result from subsequent diffusion into the healthy cells of free radicals produced in the original, more violent chemical reaction causing the burn. It is known that a burn of the flesh from any source of heat has carcinogenic and mutagenic effects. Since we now know that the charring of any organic matter produces long-lived radicals, it seems probable that some of the carcinogenic and mutagenic effects may result from secondary activity of radicals produced by the burn. Of course chromosome linkages are broken as direct effects of the heat, but it seems probable that most of the cells exposed to the elevated temperatures in the burned area would be killed. Certainly many known carcinogenic chemicals are not radicals, and I do not suggest that all cancer may be caused by radicals. However, many chemicals recognized as carcinogenic agents, not themselves radicals, may exert their carcinogenic activity indirectly through the production of radicals within the body. This would be analogous to the indirect effects of ionizing radiations already mentioned and might account for the seemingly parallel action of certain chemicals with ionizing radiations which has led to their being caUed radiomimetic chemicals (5). Many carcinogenic chemicals are large, aromatic, polycyclic hydrocarbons from which it would seem that free radicals might be easily produced. The radicals are not convicted from 'guilt by association' with carcinogenic agents. Our proposal is not intended to be accepted per se, but is offered as a working hypothesis which can be put to rather objective test because of the powerful method of electron spin resonance now available for detection of radicals. That certain radicals are likely to be carcinogenic agents, or that some types can lead to genetic mutations, probably will not be questioned. Others, possibly some or all of those which are sufficiently stable in organic matter to be detected with paramagnetic resonance, may be perfectly harmless. I do not therefore recommend that we become suddenly alarmed about the radicals around us. I do think there is some justification for the careful study of these radicals which can be produced, transported, and taken into the body so easily. This study is made easier by the powerful new method of paramagnetic resonance for detection of such radicals. If externally produced radicals are indeed dangerous, we can fortunately detect and avoid most of the ones we now are eating, breathing, or rubbing into our skins. Ingram (1) has shown that the number of radicals produced by heating organic matter is a sensitive function of temperature. Tests in our laboratory on common foods such as meat and bread show that those cooked in a normal manner have no detectable resonances or only very weak resonances, 356 Walter Gordy whereas burned food, scorched toast, charred steak, etc., have strong radical resonances. The temperature at which a cigarette is burned should have significant effect upon the number of radicals produced, although it may be impossible to produce smoke without producing radicals. If it proves harmful, we do not have to preserve our food by atomic irradiation. Acknowledgement — Several of my students, Howard Shields, Harvey N. Rexroad, Frank Patten, and Gene McCormick, assisted with microwave magnetic resonance experiments in connection with the hypothesis proposed here. I am indebted to my wife for encouragement and for assistance with library reference work. REFERENCES 1 . See, for example, comments by : G. E. Pake, J. E. Ingram, J. Cambrisson, and J. Uebersfeld, R. Livingston, W. Gordy et al.: Disc. Faraday Soc. 19, 179, 184 (1955). 2. General treatments of the subject are given in: W. Gordy, W. V. Smith, and R. F. Trambarulo: Microwave Spectroscopy, Chap. 5, J. WUey and Sons, New York (1953). W. Gordy: Microwave and radio-frequency spectroscopy. In: Techniques of Organic Chemistry, Vol. 9, 71-185, ed. by A. Weissberger and W. West, Interscience, New York (1956). D. J. E. Ingram: Spectroscopy at Radio and Microwave Frequencies, Butterworths, London (1956). 3. W. Gordy, W. B. Ard, and H. Shields: Microwave spectroscopy of biological substances. I. Paramagnetic resonance in x-irradiated peptides; and 11. Paramagnetic resonance in x-irradiated carboxylic and hydroxy acids. Proc. Nat. Acad. Sci., Wash. 41, 983-1004 (1955). 4. E. C. Evans: Effects of hydrogen peroxide produced in the medium by radiation on spermatozoa of Arbocia punctulata. Biol. Bull. 92, 99-109 (1947). O. Wyss et al. : Role of peroxide in the biological effects of irradiated broth. /. Bact. 56, 51-57 (1948). W. M. Dale: Some aspects of the biochemical effects of ionizing radiations. In: Symposium on Radiobiology : The Basic Aspects of Radiation Effects on Living Systems, 177-178, ed. by J. J. Nickson, J. Wiley, New York (1952). R. E. Zirkle: Speculations on cellular actions of radiations. In: Symposium on Radiobiology: The Basic Aspects of Radiation Effects on Living Systems, 333-356, ed. by J. J. Nickson, J. Wiley, New York (1952). 5. A. Haddow: Comparative studies of the biological effects of ionizing radiation and of radiomimetic chemical agents. Proc. Int. Conf. Peaceful Uses of Atomic Energy, 11 213-218, United Nations, New York (1956). PART VI INFORMATION NETWORKS A PROBABILISTIC MODEL FOR MORPHOGENESIS* Murray Eden Laboratory of Technical Development, National Heart Institute, National Institutes of Health, Bethesda 14, Maryland Abstract — A program has been outlined for establishing relationships between the form of an organism and the minimal information content of the germ cell from which the organism was derived. A simple two-dimensional model has been chosen in order to explore the feasibility of such a program. A suitable information measure has been defined for this model and computations of information have been made for small aggregates of cells, as well as estimates for larger aggregates. An arbitrary growth process has been formulated which is analogous to an assignment of virtually no information to the germ cell. The properties of this growth process have been studied and suggest that even such minimal information content in the germ cell is sufficient to specify the over-all form of the organism with high probability after a certain number of divisions have taken place. Possible ways of extending the model and increasing its embryo- logical relevance have been suggested. One of the problems that has been touched in this symposium is the particularly elusive subject that has been with biology since its beginnings as a science; that is, the general relationship between function and form and between growth and form. In the particular terms of discourse of this symposium the question might be phrased in this way: 'What is the minimal amount of information that is required in a fertilized egg, so that after a certain number of divisions and a certain length of time the egg will have developed into an organism that is recognizable as being a member of a certain species?' A number of workers have estimated the information content of biological objects. Dancoff and Quastler (1) computed values of information content relative to four different models; on the basis of atomic orientation, molecular structure, chromosome volume and a genotype catalogue. These authors were careful to specify the limitations of their computations. They write, 'We have arrived, by very tentative methods, to the result that the essential complexity of a single cell and of a whole man are both not more than 10^^ nor less than 10^ bits; this is an extremely coarse estimate, but is better than no estimate at all.' LiNSCHiTZ estimated the 'physical entropy' of a bacterial cell as being 10^^ bits (2) and Yockey (3) has computed the information content of DNA based on its molecular size and on a postulated cryptographic relation between proteins and nucleic acids. Most of the values obtained have been very large, and one would presume that they are large enough to describe adequately the observable properties of a living organism, it may be that the growing organism requires a lot less information in the germ cell than is indicated by * The work presented here was begun during the tenure of an United States Public Health Service Special Fellowship in the Department of Mathematics, Princeton University, Princeton, New Jersey. 359 360 Murray Eden estimates made on a molecular level. In the terminology of communication theory, the redundancy of the source m.ay be extremely high. An examination of the literature indicates that there is some support for this view. Studies of properties of monozygotic twins have special relevance to this point. It may be noted in passing that the existence of twins or high multiplets derived from a single germ cell is in itself strong evidence of the presence of at least a small amount of redundancy in the germ cell (4). Monozygotic twins presumably arise from an identical genetic background and they develop into mature organisms that can be compared with regard to certain of their properties. As long ago as 1876 Galton (5) studied what he called 'The History of Twins as a Criterion of the Relative Power of Nature and Nurture'. Newman, in a long series of pubhcations begun in 1912 (6) has studied both human twins and armadillo quadruplets. The nine-banded armadillo is exceptional in that the female gives birth to monozygotic quadruplets. The scales or scutes on the back of an armadillo are regular and easily counted, even in the fetus. Newman prepared a fairly large statistical study on these quadruplets and he found a correlation coefficient for fifty-six sets of male quadruplets of 0.93 and for fifty-nine sets of females of 0.91. Still there was no identity in the scute counts. Work of a similar character has been done by Hancock (7) on mono- zygotic calf twins, and by Went (8) on genetically identical seedlings. The conclusion that may be reached on the basis of studies such as these is that even when embryonic growth starts from genetically isomorphic cells, by the time the organism has developed to maturity there is, it is true, a great similarity in the large, but at the cellular level there is very httle similarity. This would suggest that, aside from the genetic signals or instructions, there are certain statistical variables or environmental factors operating that permit the development of an organism to an ultimately recognizable form but require a good deal less information than would be required if every element in the structure of the organism had to be specified with microscopic exactitude. An attempt to construct a theoretical model was made by Turing (9), who in 1952 posed the following problem. Given a group of identical cells arranged in some symmetrical configuration, e.g. a ring or a sphere; assume that each cell contains the same concentrations of certain chemical substrates and that the laws of diffusion and other classical physical laws hold. How can one devise a procedure whereby this homogeneous collection of cells could develop and differentiate so as to produce asymmetric or periodic forms? Turing proposed accomplishing this in a way that does not do too much violence to biological understanding. He postulated certain hypothetical chemical reactions involving substrate, inteiTnediates and enzymes, and built into this set of hypothetical reactions appropriate reaction rate constants so that the resulting reaction system would exhibit a special property: namely, that statistical fluctuations in the concentrations of chemical components in various cells would increase in amplitude so as to produce an instability and result in an asymmetric form or a form exhibiting periodicity. By use of a specific example he showed how a ring of cells might grow into something more or less petal-shaped with three or four lobes or petals. In another example he developed a mottled pattern on a two-dimensional surface. The model described below is entirely mathematical; physical or chemical A Probabilistic Model for Morphogenesis 361 phenomena are not considered. The principal concern of this model is the domain of forms an idealized organism can assume and the likelihood of an egg developing into such a form. Tlie mathematics employed is elementary, and as in so much of combinatorial analysis, it is ad hoc*. The attempt to establish a relationship between the form of an organism and the information content of the germ cell ancestor is treated here from a point of view that has some resemblance to that of statistical mechanics. It is assumed here that there are only a finite number of different kinds of cells in any organism and a finite number of cells of each kind. We neglect the dynamic processes occurring continuously in an organism: changes within cells, the migration and movements of cells, the death of certain cells and the cleavage or maturation of others. If there is some well-defined way of des- cribing the orientation of each cell in any organism relative to the other cells in that organism or relative to some arbitrary system of coordinates, then it is possible, in theory at least, to enumerate all the possible ways of arranging cells into different configurations. Some of these arrangements would be recognizable organisms, the overwhelming majority would not. In any case, these objects, both the recognizable and otherwise, are elements of the set of all possible configurations. This procedure might represent a means of defining a given species by certain restrictions on the possible orientations of cells and thus to identify the given species with a well defined subset of all possible configurations. Most multicellular organisms can be said to arise from a single cell resulting from the fusion of two germ cells. It is true that there are certain biological objects, of which the slime-mold is a notable example, which take their form from the migration, coalescence and specialization of a number of free-living cells. However, such organisms are uncommon and will not be considered further. This single germ cell divides into two cells and these cells will divide further and so on until maturity. Throughout the course of this branching process the growing organism will pass through a sequence of configurations, each of which is an element in the set of all possible configurations. If there is a f: relationship between successive configurations which is recursive, then a generating function can be constructed to describe the branching process. Generating functions are useful because they may afford a means of assigning a probability to each possible configuration. The actual model chosen for investigation has been simplified to the extent that its relation to biological reality is largely impressionistic. Its justification is heuristic, for the study of relatively simple systems may suggest methods of approaching the real systems which are so very much more complex. The element of the model is called a cell. All cells are considered to be identical. We restrict ourselves to the consideration of arrangements of cells in two dimensions. The shape of the individual cell is unspecified (they may * I should like to take this opportunity to express my debt to a number of mathematicians both at Princeton and at the Institute for Advanced Study with whom I have discussed this problem; and in particular to Professor Valentine Bargmann, Professor William Feller, ij Dr Hale Trotter and Dr Norman Shapiro for their stimulation and suggestions. Needless to ' say, the results and errors are my own. 362 Murray Eden be thought of as squares), but their positions are restricted to the points of a (two-dimensional) square lattice. Any arrangement of cells on the lattice will be called a configuration, i.e. an arrangement of k cells will be called a /c-configuration. If each cell in a configuration is adjacent to at least one other cell then such a configuration will be called connected. We will be interested only in connected configurations. The set of all possible (connected) A'-configurations will be called the k-array. The number of distinct ^-configurations, i.e. configurations that are not iso- morphic under translations, reflections and rotations, will be called the order of the ^'-array, symbolized N[k]. Each cell will have four edges, corresponding to the four nearest-neighbor lattice points. An edge will be called open if its corresponding lattice point is unoccupied by a cell, otherwise it is covered. Each cell also has four corners corresponding to points equidistant to four lattice points. A corner will be called an inner corner if it is at the center of a cluster of four cells. The problem of enumerating all possible A'-configurations is one that has, as yet, no easy solution. Similar combinatorial problems, arising in physics in what is called the order-disorder problem, have been considered by a large number of workers. Of particular relevance to the above problem is the work of Van der Waerden (10), Kac and Ward (11), and Humans and de Boer (12). Certain bounds can be set for the order of the A-array. We can determine a lower bound for N[k] by enumerating all members of a certain subset of [k], i.e. the subset in which all save two cells have two edges covered. Two cells, i.e. the ends, have only one edge covered. It is even easier to enumerate a smaller subset of this 'two-ended' set. Consider an arbitrary lattice point as the origin of a random walk. Limit the choices for the first step and each succeeding step in this random walk to lattice points, either above or to the right. The k^^^ cell will be added after k — 1 steps are taken. At each point there are exactly two possible choices, so that in all we have produced 2*^^~^ configurations. Since each configuration (except those that exhibit internal symmetry, in any case, a small fraction) occurs four times in 2''"^ configurations, the number of distinct configurations is 2^~^. The restriction to two choice points is dictated by the necessity of avoiding cross-overs in the random walk. Obviously, each cross-over would have the effect of decreasing the number of occupied lattice points by one. However, if the random walk is permitted three choice points, i.e. above, to the right and to the left, one can estimate the number of such walks of length, k, which contain no point adjacent to more than two occupied sites*. Such walks are isomorphic to the set of 'two-ended' /c-configurations. This estimate was found to be very close to (1 + \/^)'^~^- I^ consequence the lower bound for the number of A'-configurations may be raised to this value. Upper bounds can also be computed using a somewhat diff'erent combina- torial technique. Consider any A-configuration. Arbitrarily choose one cell as the origin, and also arbitrarily choose one of the four possible orientations of the lattice. Identify this cell by 1 if it has a cell beneath, otherwise 0. Further, this cell may have a cell adjacent to it on the left; if so, assign a 1 to the next * The mathematical details of the results presented in the text will be the subject of a separate publication. A Probabilistic Model for Morphogenesis 363 digit in the identification; a cell above it, and a cell to the right. Thus, tiie first cell Q in a configuration is identified by four binary digits. Next, identify the adjacent cell which contributed the first T in the designation of the first cell, as the second cell Cg, the second '1', as the third cell, C3, etc. Reorient the lattice so that the first cell is beneath the second. We construct the desig- nation number of the second cell as we did for the first. However, this time there are only three binary digits required since the adjacency of C^ to Q is already determined. Any of the cells adjacent to C^ that have not yet been assigned a position in the order can be given one now in a perfectly well-defined way. It is obvious that this procedure can be continued until designation numbers have been obtained for each cell in the configuration. We thus have a well-defined word in 3^ + 1 binary digits and a possible 2^''+^ such words. Since the initial cell and the orientation of the lattice were chosen arbitrarily, each district configuration (as usual, excepting those exhibiting some internal symmetry) will be given by Ak such words. Thus an upper bound for TV [A] is 23^-VA:. It is easily ascertained that a very large proportion of the 2^^^+^ words do not represent A'-configurations. These forbidden words arise for essentially the same reason that the unrestricted random walk on the square lattice fails to serve as an estimate of two-ended configurations. No simple relations have been found that will indicate which of the 2^'^"+^ words are permissible. However, one can generate a random sample of these words by a Monte Carlo procedure and arrive at a statistic that suggests that a satisfactory estimate of N[k'\ is in the neighborhood of 2-^. Values of the bounds and the estimate mentioned above have been com- puted for certain values of k (Table I). This serves to give some idea of the Table I. Estimate of Configurations for Large Arrays [k\ Lower bound Upper bound Estimate (1 + V2Y (2='^) (22.) 10 2.2 X W 2.4 X 10* 5.8 X 10^ 16 4.2 X 10^ 6.3 X 10» 2.5 X 101 25 1.1 X 10* 8 X 10^' 6 X 101- 100 1.7 X 101* 3 X 1085 6 X 10" 1000 1.7 X lO^'i 3 X 10««* 6 X lO^"^ magnitudes one might expect for configurations of large numbers of cells. So long as the number of cells is small, the distinct configurations can be exhibited with relative ease. This has been done up to A: = 8 and the results are given in Table II. In order to establish the assignment of a probability to each of these con- figurations, a simple and nearly featureless generating function was adopted. Starting with a single cell, equal probability is assigned to each of the four possible two-celled configurations. These are all isomorphic. This two-celled configuration has six open edges. Equal probabilities are assigned to each 364 Murray Eden edge. This lime, of the six three-celled configurations obtained by adjoining a cell to an open edge, four are isomorphic to one of two three-celled configura- tions, and two to the other. Thus, the probability of the first three-celled configuration is 0.67 and the other 0.33. This procedure can be carried out indefinitely, in each case assigning equal weight to each open edge and adjoining a single cell at a time. Table II. Number of Configurations in Each Array k N[k] N[k] N[k - 1] 1 1 2 1 1 3 2 2 4 5 2.5 5 12 2.4 6 35 2.9 7 108 3.1 8 367 3.4 While there is no biological organism that exhibits this pattern of growth, it has certain features in common with some tissue cultures, bacterial colonies or tumors, in that the cells are more or less undifferentiated. Growth in such "^-12 7^-10 7rM2 7r^I2 7r-I2 7r=l2 7M4 (1) ^ (I) ^ (2) Ug (1) 4^ (I) ^^ (2) ^ (3) d^ 0I1S9 0.0883 0.0817 0.0761 0.0594 0.0511 0.0250 (1) n^ Ocfim '" 7r=I4 Tr=l4 '^' a&n <«' crrxfl <» ^ '^' Bifi 0.0229 0.0203 0.0194 0.0125 0.0111 0.0104 0.00555 TT^U (I) 6- ARRAY Fig. 1 biological objects has no preferential direction except that it is peripheral, a condition due most likely to the fact that diffusion of nutrient is too slow to permit any large number of cell divisions in the interior of the growth. Exact computations have been carried out for the probability associated with each A'-configuration up to k = 8. As before, computations for k > S, while easily performed in principle, are prohibitively time-consuming. The configurations for k = 6, 7, 8 and their associated probabilities are assembled in Fig. 1, 2, 3. An unanticipated property of the particular generating function employed was revealed as a consequence of these exact computations. It will be observed in Fig. 1 to 3 that configurations have been grouped so that each different value of probability is recorded next to a single prototype configuration. A Probabilistic Model for Morphogenesis 365 These configurations bearing llic same probability, while they are not isomorphic in the sense mentioned earlier, have an important property in common. If each configuration is represented by a graph (13), identifying the cells as nodes ir, 12 7r_ 12 TT-U ■"•-14 TT-M T'-M '"'U s s (2) te (') ft ") ^ (1) t# (1) E§° (2) cfe <') S 9.2642 6 SS30 4.4603 3.S67S 3.3181 3.1906 2.6296 TT- 14 V 14 TT^U Tr=14 TT-H Tr^\t TT^IS (» #3 (')[ffl3^ Wdffi^ <5)cnrffl (3)%° (') ^ (') P^ 25119 2.3261 18499 1.2956 0.7837 0.7730 0.5952 7r= 16 77- 16 7r.i6 7r= 16 '"■=16 "■ 16 '" ^ '" ^' '^''rrftTT. '" 4™' ""rftrrn "' rfWi ^ ''' ^n^ 0.5357 0.5060 0.4067 0.3869 0.3075 0.2976 0.2679 TT 16 V 16 TT 16 7r=16 TT; 16 IT 16 S „ S rn rn S S S (') °#^ (i9)ftTTTn (ugi™ (I) H (luElnfl H I I I I II 0.2530 0.1587 0.1554 0.1488 0 0794 0.0397 7- ARRAY Fig. 2 "■=12 rr=i4 ir=u ir=i4 v,u ir=u w-i4 in^u itm ir^n v^u (">& <'>^ ("^ <«^ 0)^ ">^ ("te^ 0)^ (»fe (l)ffiffl (4)rT4ffl 5-12 3.50 3JS 3J4 3.01 2.61 2J5 2J« 2.19 1.12 ■ir=i4 77=14 Tr-14 ir=i4 vm ir=u 7r=i4 v-}i itM ■"■=16 1.66 "■=16 (I) (1) 1.40 1.30 (1) "'cfflD '" (2) rffFP (2) (2) (I) (2) (3) 1.127 1.094 0.9462 0.6456 0.7292 0,7257 0.6841 Wz. 16 7r=)6 ir=i6 Tii6 TTiii Tr=i6 ir=]i S <«a& '"^'"'d* '« cP "'-f "'c#^ (3) (5) 16 _7r=i6 ir=i6 $, (I) (IS) 7r=i6 w>cflffl 0.62719 0.6093 0.5960 0J77O 0.5280 OJ20i 0.4716 0.4340 0.3672 0.3480 0.3364 IT: 16 Tr=u "=1« "^16 ir-l« "=i< "^16 ir=.M ir^M itm "=16 s s „„ p n PS (2)rTffP (l)'#^ (lO)rrTfFP (1) ^ (2) rrFRn (1) #5^ 112) rrrrfB (2) F#P (4) 0^040 0l2920 0J644 0JZ79 02170 Ol20K1 0.1707 0.1649 0.162S ai520 0.I4S8 "=14 "^16 "=14 "=16 V^U ■"=16 "=16 ■"-16 '"=16 "=.16 "-16 (1) ^^ (8) ^ (I) ^' (1) °f°'(13)nflg3 (3) ^ 0) !§: '(15)B^ (i) CdgJ (4) ^ (3) ^ ' ai42l 0.1190 0.1166 0.1116 0.1073 aiOtO 0.1642 0.0043 0.0613 0.0769 0.0744 T»I6 ir=I6 77^16 7r=16 "=16 osi TT-U ir^M -77=16 ir=16 '"=16 ■"=16 rrrWl (SWrrftrfl (D "^ (2) S=™ (37)^TTTf^ (3) rftrfll (5) Ftfffl 0) M (52) illB O §™^ 0) g^ aOMS 0.0546 0.0531 0.04216 0.0391 0.034« 0.0394 0.019* O.Om 0.0196 0.0192 Tml6 "--16 aoo** 0M496 8 — ARRAY Fig. 3 and the covered edges as branches between nodes*, it will be seen that all configurations represented by the same graph have the same probability. Certain other properties are also suggested by consideration of these small 'organisms'. The configurations with the largest number of inner cornersf are * This representation is analogous to the graph obtained by identifying countries on a map with nodes and common frontiers between countries with branches. t We can use the perimeter, n, i.e. the number of open edges, instead of the inner corner, C, in describing the property in question since tt ^ 2{k i- 1 — C). 366 Murray Eden most probable. It also appears that configurations with many short branches are more probable than those with a few long branches. Finally, it is also observed that as k increases, a decreasingly small fraction of the A'-array carries the weight of probability. This is shown in Fig. 4 and in Table III. The k- configurations have been ranked in order of decreasing probability, that is, .2 .3 .4 .5 .6 fractional rank Fig, 4 .8 1.0 Table III. Probability of k Probability of configuration of rank 1 least probable asymmetrical configuration Rank at T.p, = 0.5 Fractional rank at •Lp, = 0.5 1 1.00 1.00 2 1.00 1.00 — — 3 .67 .33 — — 4 .33 .167 — — 5 .40 .067 2 .167 6 .12 .011 6 .169 7 .093 .0016 12 .11 8 .051 .0002 24 .067 10 .021 2 ^ 10-« 103 .020 16 .0015 3 X 10-1^ 6400 .00002 the most probable configuration was designated 1, the next most probable 2, and so on, and the cumulative probability (as ordinate) was plotted against the rank divided by N[k] (number of distinct A'-configurations) as abscissa. Fig. 5 o face p. 367 A Probabilistic Model for Morphogenesis 367 Since any direct extension of the model to larger values of ^ does not appear feasible, another procedure was adopted. Starting as before from a single cell, the edges were numbered, a random number table (14) was consulted to find a number equal to or less than 4, and then a cell was adjoined to the appro- priately numbered edge. The open edges were renumbered, another number equal to or less than the number of open edges obtained from the table, and a new cell adjoined. In this way samples of 1000 10-configurations and 16-configurations were constructed. A few larger configurations were prepared by this procedure. One such containing 200 cells is shown in Fig. 5. In the case of the sample of the 10-array, one configuration was obtained twenty-two times. Its probability was computed by the exact procedure des- cribed above and found to be 2.06 per cent. All the configurations containing four inner corners (77-= 16) (maximal for /: = 10) appeared more than ten times each. With very few exceptions, in order of occurrence, there followed the configurations with tt = 18, 20, 22. There were eighty-three occurrences with 77 = 24 (no inner corners), but none of these was two-ended. Although it was impossible to enumerate all the configurations, by judicious use of the equality of probabihties found in configurations with the same graph, estimates were made of the numbers of configurations of each kind up to rank 1150. The data were plotted in Fig. 4. It can be seen that the portion of curve obtainable is very close to the ordinate axis. A similar procedure was followed in the case of the sample of the 16-array, Here, estimates v^ere considerably poorer, but the same general features were revealed (Table IV). The thirty-two possible configurations with 77 = 18 Table IV. Summary of \6-airay Monte Carlo Sample Perimeter Configura- Cumulative Fractional E(pd tions occurrence rank xlO* 16 1 0 18 32 .028 1 X 10-« 8.75 20 569 .202 2 X 10-5 3.00 22 6250 .455 4 X 10-* .40 24 27,300 .728 1.2 X 10-3 .16 26 148,500 .907 6.6 X 10-=» .012 28 — .967 30 — .995 32 — 1.000 34 — 1.000 appeared twenty-eight times, or an expectation of occurrence of a particular configuration of 8.75 x 10^*. (It is assumed that all configurations with identical values of tt have approximately equal probabilities of occurrence.) it was estimated that the expectation of occurrence of a configuration of 77 = 20 was 3 X 10-4; 77 = 22, 4 x 10"^; and 77 = 24, 1.6 X 10~^ In this sample of 1000 there were only five occurrences of configurations with 77 = 32, and no occurrences of 77 = 34, although a low estimate of the number of distinct 368 Murray Eden 16-configurations with tt ~ 34 would be 250,000. These estimations are plotted on the same figure as the computations for configurations of up to eight cells (Fig. 4). It can be seen that none of the estimates obtainable up to a cumulative probability of 0.907 can be distinguished from the ordinate axis. The probability of the most probable A-configuration and of the least probable A-configuration are presented in Table III, for several values of Ar. It will be noted that the probability of the most probable configuration decreases slowly with increasing k. While there is no practical way to make exact com- putations of probability for large values of k, it may be conjectured that the probabiUty of the first ranked configuration is proportional to 1/2^'. On the other hand, the probability of the configurations of the lowest ranks falls very rapidly*. As with the estimates of the number of configurations, exact solutions for probability are readily obtained only for the two-ended configurations. As was noted earlier the number of such forms approximates (1 + V^)'' but the probability associated with each such form is 2^//c ! Information theory (15) suggests methods of defining appropriate measures for the distribution of probabilities as a function of A'. If 7V[A'] is the number of distinct configurations containing k cells each, the maximal uncertainty for the A'-array can be defined as i/^." = — Ig A^[A']t. In a similar manner, an uncer- tainty can be defined for an arbitrary generating function, Gj, considered as an N[k] information source. H{G^^ = ~^Pi Ig/'o ^^ which /?j is the probability that i = l the generating function G^ will terminate after k — 1 adjunctions in configura- tion oj,. Further, a measure of relatedness (16) may be defined as I{Gj^ — [//,o - H{Gj,,)]. What does this mean in teiTns of information theory? Supposing we had a generating function or some procedure that produced every one of these unusual configurations with equal probability. Then the two numbers H^9 and H(Gjj,) would be identical. The uncertainty of such a generating function would be maximal. On the other hand, if the generating process were such as to specify, with probability l,only one out of the total number of configurations, then the uncertainty of the generating process H{Gj,.) would be 0. As can be seen, I(Gj^) for a given generating process carried out through k steps has been defined above simply as the difi'erence of these two quantities. Very briefly then, this measure would suggest that if a knowledge of the generating process does not enable us to predict which of the possible configurations to expect after the process has gone along for k steps, then knowledge of the generating process provides no information. On the other hand, if one can devise a mathematical mechanism, that is, a generating process, that can specify the ultimate form of an organism exactly, then the generating process contains all the information it possibly can. Applying this measure to the presently available data and the particular generating function introduced earlier, it is observed that liGj k) increases with * It will be noted that the probabihty of the most probable configuration exhibits a maxi- mum at A; = 5. This is an accident attributable to the fact that this particular 5-configuration is the only one containing a cluster and it is asymmetric. Such an accident is extremely unlikely to be found when k is large. t The symbol 'Ig' is used here to denote 'logarithm to the base 2'. A Probabilistic Model for Morphogenesis 369 increasing k (Table V). Estimates have been made for A; = 10 and A' ^ 16 from the Monte Carlo samples. These estimates are certainly lower than the precise values since an estimate of/?, was not available for every configuration and the means for rather large groups of configurations were used instead. A functional Table V. Entropy oj \i-arrays k H{G,) H KG,) 1 0 0 0 2 0 0 0 3 .92 1.00 .08 4 2.19 2.32 .13 5 2.90 3.59 .69 6 4.54 5.13 .59 7 5.59 6.76 1.17 8 7.00 8.53 1.53 10 10.00 12.29 2.29 16 16.68 21.77 5.09 relationship between liG^^j,) and k has so far not been found. One may con- jecture that the relatedness increment /(C7y,fc) — I(Gj^j._i) approaches 0.5 as k increases without limit. This may be interpreted to suggest that the rate of information accumulation in an organism constructed according to such a plan is half a bit per cell division. Other measures have been suggested as being useful to our purposes. Following the terminology of McGill and Quastler (16), the relative uncertainty M(C \ of the generating process is Dj^. = — ^ ^^ and the redundancy is Q^^ — \ — D H,} },k- The redundancy evaluated from the results presented in Table V increases from a value of 0 for A: = 2 to a value of 0.234. As with the measure HGj^j.) — /(G;j.,_i), it seems plausible to expect that as k increases Q^j. will converge to some value other than 0 or 1, but no procedure has as yet been found to test this conjecture and to determine the limit. In a qualitative way, this increase in /(C7,,fc) may be understood to mean that the featureless generating function considered above determines the configura- tions of large numbers of cells with a high degree of specificity. It is virtually a certainty that large configurations will be essentially circular in outline; that they will have a high density, i.e. they will contain very few 'holes' and short 'tentacles'. Thus if one considers the most probable outcomes of the generating procedure in the large, then these configurations appear to resemble one another very closely even though they exhibit no correspondence in detail. It is true that the results obtained with such a simple model are far removed from the intricacy of development of living things. A few regularities in the most probable forms may be introduced by small modifications of the initial generating procedure. Objects that are ellipsoidal or cruciform or objects characterized by large numbers of branches have been developed by such modifications. However, it is unlikely that further complexity can be introduced 370 Murray Eden into the growth process without drastic modification of the generating function, in particular without consideration of the history of a particular growing configuration. The results of embryology suggest that the generating process must contain a set of instructions that will alter the pattern of growth on the condition that a given stage or over-all configuration shall have been reached, and that such a change in pattern of development may occur a large number of times during the process of maturation. It is certain that such a modified generating process will have a higher information content than the process considered in detail in this paper. It remains to be seen whether modifications of this character can be fonnulated and whether a mathematical treatment of the consequences is possible of achievement. REFERENCES 1. S. M. Dancoff, and H. Quastler: The information content and error rate of living things. In: Information Theory in Biology, ed. by H. Quastler, 263-273, University of Illinois Press, Urbana (1953). 2. H. LiNSCHiTz: The information content of a bacterial cell. In: Information Theory in Biology, ed. by H. Quastler, 251-262, University of Illinois Press, Urbana (1953). 3. H. P. Yockey: An application of information theory to the physics of tissue damage. Radiat. Res. 5, 146-155 (1956). 4. K. S. Tweedell: Identical twinning and the information content of zygotes. In: Infor- mation Theory in Biology, ed. by H. Quastler, 215-250, University of Illinois Press, Urbana (1953). 5. F. G ALTON: The history of twins, as a criterion of the relative powers of nature and nurture. /. Anthrop. Inst. 5, 391-404 (1876). 6. H. H. Newman: The Biology of Twins {Mammals), University of Chicago Press, Chicago (1917). 7. J.Hancock: Monozygotic twins in cattle. Advances in Genetics 6, \A\-\1 5 {\95A). 8. F. W. Went: The role of environment in plant growth. Amer. Scient.AA,'il^-'i9i{\956). 9. A.M.Turing: The chemical basis of morphogenesis. Phil. Trans. (B) 237, 37-72 (1952). 10. B. L. Van der Waerden: Die lang reichweite der regelmassigen atomanordnung in mischkristallen. Z. Physik 118, 473-488 (1941). 11. M. Kac and J. C. Ward: A combinatorial solution of the two-dimensional Ising model. Inst, for Advanced Study, Princeton, N.J. 1-15 (1954). 12. J. Humans, and J. de Boer: An approximation method for order-disorder problems. P/?j5/ca 21, 471-516(1955). 13. D. Konig: Theorie der endlichen and unendlichen graphen, Chelsea Publishing Co., New York (1950). 14. The Rand Corporation: A Million Random Digits with 100,000 Normal Deviates, The Free Press, Glencoe, 111. (1955). 15. C. E. Shannon, and W. Weaver: The Mathematical Theory of Communication, University of Illinois Press, Urbana (1949). 16. W. McGiLL, and H. Quastler: Standardized nomenclature: An attempt. In: Information Theory in Psychology, ed. by H. Quastler, 83-92, The Free Press, Glencoe, 111. (1955). FUNCTIONAL GEOMETRY AND THE DETERMINATION OF PATTERN IN MOSAIC RECEPTORS John R. Platt University of Chicago, Chicago, Illinois Abstract — Every visual pattern element — straight lines, curved lines, parallel lines, angles, periodicities — shows some self-congruence under translations or rotations. A random mosaic of detector cells, like the 10* cells of the human eye, can be used as a null detector to indicate this self-congruence during scanning operations. This operational definition of pattern is called functional geometry. It underlies the generation of precision optical and machine surfaces by the Whitworth, Rowland and Strong methods and theoretically can approach infinite precision, starting from rough materials. It converts a space pattern into time pattern repetitions whose accuracy is not limited by the mosaic structure. The spherical eyeball shape is generated by functional geometry, and its almost perfect rotation operations can establish among the retinal cells an external Euclidean metric of perception-space which is independent of the distortions of mapping on the retina or the cortex. A variety of second-stage and higher-stage neuroanatomical structures would have to be grown for tracking and detecting pattern repetitions. These would almost certainly include delay lines and null-transmitter cells to transmit only the identical parts of multiple input patterns. Such pattern-perception in the mature network is equivalent to determination of the initially unknown space relationships or addresses of the random detector cells. A non- addressed mosaic requires much less initial assembly information than a pre-addressed mosaic, but requires a long learning and growth time for address-determination after operation begins. It has other quasi-human characteristics, since to determine addresses it consumes information in abstracting properties, draws analogies, shows closure, may^ts symbols, learns from experience, incorporates functional memories in the network structure, and apparently might even need to sleep. But the self-congruences of functional geometry would impose certain paradoxical and Kantian restrictions on the learning process, such that only certain congruent types of experience can be learned at all, and only certain congruent types of address-connections can be formed, regardless of what the experiences are. This paper revolves around the problem of visual pattern perception by the human eye and brain. It is an attempt to generalize the problem; to restate it in a language suitable for electrical networks; and to see what basic physical principles might be involved, what detailed neural relationships might be required, and how these principles and relationships restrict and determine the general properties of such networks. The eye has millions of simultaneously active photodetectors. The theory of connections in such a system is still in a primitive state. It is therefore necessary to begin by introducing and explaining a number of new terms which will be needed in the analysis. 371 372 John R. Platt I. MOSAIC RECEPTORS Single-element and Multiple-element Receptor Systems A feedback mechanism or a neural network or a social organization is a decision network connecting sensory-receptor inputs with motor-effector outputs. The system may have single-element receptors or multiple-element receptors. An example of a single-element receptor is a phototube. Another is a proprioceptive muscle spindle cell. In the simplest case each of these might actuate a single- channel feedback loop or reflex arc leading to a one-coordinate output function of time. There may be non-linear circuit elements in the loop that pulse or chop or clip or average or stabilize the input or otherwise transform it. Never- theless, each feedback signal from a single-element receptor remains a one- dimensional time signal except as it may be trivially or artificially split into several components. Multiple-element receptors consist of many functionally similar single- element receptors acting simultaneously. If each of these has its own private reflex arc, independent of the others, to its private motor output, the system is merely an additive system of single-element receptors. But to avoid conflict in the motor responses, it is desirable to reduce their independence. In this case, the simultaneous inputs can be combined in a decision network which selects a single complex response from the output field, with suppression of conflicting alternatives. Some of the physical and mathematical relationships in such a network were discussed earlier (1). The receptor organ of such a system becomes a mosaic receptor with a pattern and hierarchy of connections to the decision network. It is an advantage if the network is concentrated into a compact central switchboard where extensive interconnections can be made quickly and cheaply. Examples of mosaic receptors are the 10^-element retina of the human eye, the basilar membrane of the ear, and the olfactory membrane. The retina will be treated as the prototype of such systems. Mechanical mosaic receptors have also been constructed such as the lO^-element assembly of sensory pins in the reading head of a punch-card sorter or reader. A social mosaic receptor would be the 10- traveling salesmen sent out by a business organization. The relatively low complexity of these man-made systems means that they are inferior to their biological counterparts by more orders of magnitude than almost any other man-made devices. It is true that some artificial receptor systems are more elaborate than the two mentioned. A television camera iconoscope tube with its 1 0^ separate resolvable spots is an example. But at present, the iconoscope signals are scanned and sent in sequence into a single output channel, undergoing only the most rudi- mentary inter-comparisons or decisions, such as stabilization, contrast or color balance. Likewise the 10^ grains of a photographic emulsion, although they form a very fine-grained system, do not feed into any decision network until they are transduced onto the human retina. Pre-addressed and Non-addressed Mosaics An address, in computer nomenclature, designates a point in the network at which a signal may be located. But in a mosaic receptor, the address of an Functional Geometry and the Determination of Pattern in Mosaic Receptors 373 input element is only partly specified by its network-address. It is incomplete unless the location in space, or space-address, is also given, at least relative to the other elements, since both address components efTect the kinds and combina- tions of messages sent through the network. Mosaic receptors may be pre-addressed or non-addressed. In a pre-addressed system, each receptor element has a specified space address and network address. It is completely connected in a unique and permanent way to the decision net before the net begins to operate. In a non-addressed system, the space address of an element, or its network address, or both, may need to be determined after operation begins. This may be the main difference between the insect eye and the human eye. The insect eye consists of a close-packed array of uniform receptor elements. Because of their uniformity, they lie in long parallel lines. Absolute genetic determination of the connections from each element to its neighbors and to the decision net might be easy: a pre-addressed system. Straight lines in the field of view that fire all elements on one of the principal lines of such an array should be easy to distinguish from curved lines, if such a distinction were biologically useful. But straight lines in any other general direction would be hard to distinguish from curved, without very elaborate inter-connections in the network; and therein might lie the limitations of a pre-addressed system. The human retina escapes this impasse. It appears to make no such dis- tinction between straight lines in different directions. And indeed under a microscope the cones in our foveas appear to be close-packed but sufficiently non-uniform that no straight line arrangements are more than a few cones long (2). Assembly Information This useful randomness seems inevitable from assembly considerations. A non-random biomechanical assembly of 10^ elements distributed over several square centimeters of the retina with individual tolerances of better than 1 micron would be almost inconceivable. Even if this could be achieved, the complexity of a non-random wiring diagram for any system of 10^ input elements, geometrically regular or irregular, would be almost impossible for the chromo- somes to specify, as Pitts has emphasized (3). And so the randomness, if it has solved one dilemma, has evidently created another. The addresses of the retinal elements are now uncertain. All straight lines have been made equal by a device which appears to make it impossible for the eye to identify straight lines at all! On the other hand, if this new problem could be solved — and the present paper aims to show that it can — non-addressed systems would evidently have one tremendous advantage over comparable pre-addressed systems: their economy of assembly information. In pre-addressed systems, if the inter- connections among /;; elements are to be specified in advance, the assembly information must increase with a power of/;? at least as large as two and perhaps much larger. This elaboration of initial design specification and mechanical assembly detail is what makes our artificial electronic networks slow and expensive to 25 374 John R. Platt manufacture. Sooner or later the increase with increasing m will limit the size of the pre-addressed systems we can construct, no matter how much the assembly process is speeded up. But for a non-addressed system, even with 10^ or 10^ elements, a very few specifications of the general assembly or growth patterns may suffice (3). The construction is cheaper, whether measured in assembly information, in time or money. Obviously there is a price. It is that the addresses of all retinal elements must now be learned— after operations begin. The construction is speeded up; the attainment of full operating efficiency is delayed until address-determination is completed. But the non-addressed system constructed with a given amount of assembly information can eventually become far more complex and 'intelligent' than its pre-addressed counterpart. This initial incompetence may be why, in evolution, the non-addressed organisms only become prominent when parental care appears in family systems like those of birds and mammals. The long learning time for large m might be connected with the long childhood of the more intelligent species. Actually there may be no sharp boundary in biology between the pre- addressed and the non-addressed. On evolutionary grounds alone, a vitally necessary fraction of the human brain must certainly be pre-addressed. The autonomic nervous system may be largely so constructed. Reflex actions and probably color vision seem to have this character. The non-addressed sections of our networks, although perhaps responsible for our most characteristically human activities, may be a late and still secondary addition to a large pre- addressed core — as Dr. Sacher stressed in his comments on this paper. It is often asserted that nerves and synaptic connections do not grow. This might be true for the pre-addressed sections; but it should be false for the non-addressed sections. Address-learning in a network necessarily means creation or change of connections. Change of neural connections means growth or atrophy or both. If new synaptic connections do not grow, they must at least be selectively and permanently activated or deactivated during the address- detennining process. The Pattern Question Whatever the economy of assembly, the question remains: Can randomly arranged elements be used to make discriminations of straight lines or of any other types of pattern elements ? There is evidently an intimate connection between the perception of pattern and the determination of the addresses of the retinal elements. To make the question more precise, let us number the elements 123 • • -y • • • in as nearly the same way as possible in all retinas, and set up coordinate axes as nearly alike as possible. The randomness means that element y will have different address coordinates, X;, j^ in every retina. Or better, we might specify addresses by relationships rather than coordinates, giving them forms such as 'Element y is collinear between elements g and p\ This address might be right in one retina, wrong in another. Such an uncertainty of internal pattern has to be resolved within the network. The question is then: Can the coordinates x^y^ be detennined, or can the straight-line or other geometrical spatial relations of element j to many other Functional Geometry and the Determination of Pattern in Mosaic Receptors 375 elements ■ ■ • a • ■ ■ p ■ ■ • be detemiined, within the receptor network and solely by its normal functional operations? And if so, how? The present paper aims to show that at least one simple method exists for this functional detennination of addresses. It can be called the method of functional geometry. It seems feasible for use at least in an artificial mosaic system. It may or may not be the method used by the eye or by any other biological system, although many of the results here strongly suggest that it is. In any case, its existence removes a principal conceptual difficulty of non- addressed mosaic receptors. And the examination of one particular method can help sharpen up our experimental inquiries as to what methods of address- determination and pattern-perception actually are used in biological systems. II. FUNCTIONAL GEOMETRY There is a class of geometrical operations that is of great importance in the highest precision machine work and in anatomy, especially in the joints of vertebrates. The operations are related to group theory but, as we shall see, they might form the axiomatic basis of a separate systematic mathematical discipline. If this discipline were ever created, an appropriate name for it would be functional geometry. Generation of Perfect Surfaces An illustrative operation of this class is that by which an optician or an amateur telescope maker grinds and polishes a spherical lens or mirror surface (4). A rough blank of glass is placed against another rough blank of glass or metal or pitch, with grinding or polishing powder between them. The blanks are pressed and rubbed together by hand or by a rather crude and loose grinding machine, as shown schematically in Fig. 1a. The operation continues with Fig. 1. Self-congruence of a sphere or a circular arc under random translation. successively finer grades of powder. Finally each of the surfaces approaches a perfectly spherical shape to a precision which may be one-tenth of a wavelength of light, or better if desired. Theoretically, if edge effects are neglected, the method can approach infinite precision. Its practical precision is limited only by the patience of the optician and the accuracy of available testing methods. The accuracy of approximation to a perfect sphere can be many orders of magnitude higher than the accuracy 376 John R. Platt of the initial blanks or the accuracy of construction or operation of the grinding machine. Usually the optician also wants a particular curvature, convex or concave, but this is a separate question which need not concern us here. The curvature determination is not automatic and it is the automatic approach to perfection by these methods which is the point of interest. In order to produce a spherical surface, the motions of the grinding machine must be (a) relative translation of the blanks in both coordinates along their surfaces, and (b) relative rotation of the surfaces. Each motion must be randomly independent of the others, relatively unconstrained by the machine. This is why the grinding machine must be loosely coupled. A grinding machine that couples the motions in any regular way or whose translations have some arbitrary fixed relation to the axis of rotation would 'over-determine' the system and damage the rate of approach to a spherical surface or the attainable pre- cision. The surfaces are self-centering, determining their own centers more and more precisely as the polishing proceeds. The reason these particular motions generate a spherical surface is that this is the only surface that satisfies the following functional definition: A spherical surface is one of two surfaces that is everywhere in contact regardless of relative translation or rotation against each other. For one surface alone, this could be made a statement of displacement congruence: A spherical su face is self-congruent for all translations or rotations in the surface. A complete sphere is self-congruent for all rotations in the surface ; that is, about any axis normal to the surface. (Three degrees of freedom. Any two rotational degrees of freedom imply the third.) The functional geometry of such definitions is conceptually more funda- mental than either Euclidean or analytic geometry. To say with Euclid that 'a spherical surface is a surface in which every point is at the same distance from a fixed point', is to require points, fixity and measures of distance. To say that 'the equation of a sphere is x^ + y^ -f z^ = 7?^ ' is to require also a coordinate system. But functional geometry generates perfect surfaces by only using two of the most primitive notions: identity (congruence) and displacement. The motions involved in these definitions are those of the continuous transla- tion and rotation groups of group theory. The definitions can therefore be generalized to surfaces representing other group operations, including discrete groups : Real surfaces approaching indefinitely close to a mathematically perfect fonn can be generated by mechanical processes that enforce displacement self-congruence under a particular set of group operations. The set determines the shape of the surface. The surface is self-centering and defines its own special centers and axes in space more and more precisely as the operation pro- ceeds. In practice, what development of these other operations has been done has come from the makers of precision screws and ruling-engines, especially Whitvvorth, Rowland (5) and Strong (6). Strong emphasized the opposi- tion between these 'inherently precise' methods (self-congruent surfaces) and the traditional 19th-century semi-precision methods of 'kinematic design' which he had described earlier (4), and the superiority of the self-congruent method. Functional Geometry and the Determination of Pattern in Mosaic Receptors 377 'The construction methods of greatest precision are ail primitive methods' (6). The following are some examples. A surface of revolution is self-congruent for rotation about its axis. (One degree of freedom : Strong method for thrust bearings.) A screw is self-congruent for simultaneous rotation about its axis and trans- lation along it. (One degree of freedom: Rowland method of lapping.) A cylinder is self-congruent for all rotations about its axis and translations along it. (Two degrees of freedom: Strong prescription for lapping a cylinder). A cylindrical surface section is self-congruent for pure translations in the surface with no component of rotation about a line normal to the surface. A gear of n identical teeth, 360^ /n apart in angle, is self-congruent for any of n different angular displacements about its axis. (One continuous degree of freedom plus one discrete. In the Strong method, the gear is polished within a kind of open-ended squirrel cage of n lapping bars or pawls that slide between the teeth. The cage is rotated by one bar after every stroke, and any initial irregularity in either the gear or the cage is polished away.) The group operations are those of the discrete group, C„. Functional geometry can therefore generate perfect right angles or other angles. 1 Fig. 2. Self-congruence in translational periodicity. By analogy with the screw and the gear, a cylindrical surface with straight parallel equally-spaced identical grooves (possibly helical) is self-congruent for continuous translation in one direction in the surface and discrete translations in the other, as indicated in Fig. 2. (In principle, the precision ruling of surfaces might be accomplished in this way.) Perfect translational periodicities in two or three dimensions might be generated in succession. There are more sophisticated possibilities on moving beyond ordinary group theory: A plane is one of three surfaces of which any pair can be placed in contact everywhere regardless of relative translations or rotations against each other. (In making optical flats by the Whitworth method of lapping, three flats are generated simultaneously by being polished against each other, with frequent interchange of pairs to prevent development of concave or convex surfaces). Restated in terms of displacement congruences: A plane is self-congruent for all translations and rotations in the surface and for two-fold rotations about an axis in the surface. Note that three-fold rotations about such an axis would be impossible. This exemplifies a fundamental physical restriction on possible generating processes, of a kind wc will encounter shortly in the biological cases. 378 John R. Platt The sophistication of this definition of a plane is that it is antecedent to the definition of a straight fine in this geometry and requires no definitions of fines or axes or coordinate systems or rectifinear translations. Other surfaces can be generated by grinding and lapping operations that maintain only a line of contact between two self-congruent surfaces, such as two surfaces of revolution rotating about skew-perpendicular axes. Biological Examples Any two physiological surfaces that are pressed and rubbed together con- tinuously must exhibit displacement congruences approximating mathematical perfection. The familiar chicken drumstick has at its lower end a perfect surface of revolution sweeping through an angle of about 270°. (One degree of freedom. It may be slightly helical, since the revolution is not complete.) The helical grooves on the narwhal tusk may be generated by displacement congruences as it grows from its socket. Ball-and-socket joints are likewise famifiar in anatomy, with accurately spherical surfaces. (Three degrees of freedom.) The eyeball-and-socket is perhaps the most perfect instance of this type. The spheres must be very precise if there are not to be considerable changes of pressure during normal rotations. The oculomotor musculature provides all three rotations, about the Z-axis (vertical axis), the 7-axis (transverse horizon- tal), and the Z-axis (longitudinal horizontal). Functional geometry provides a precise self-centering specification of the center of the spheres and therefore of the reference point about which all the operations of the three-dimensional continuous rotation group can be carried out. What is more important, these motions provide the necessary displacements by which the displacement congruences of any pattern in the external field may be detected by the retina. To anticipate the results of the next section, if an arc in the external field produces an excitation pattern on the retina (Fig. 1b), the pattern can remain unchanged during a displacement of the fixation point along the arc if, and only if, the arc as seen from the eyeball is either a straight fine or the arc of a perfect circle, with constant curvature. This is the two-dimensional analogue of the functional definition of a perfect sphere given above. Likewise a set of lines in the field is parallel and equidistant if and only if the excitation pattern can remain unchanged as the fixation point moves from one line to the next or moves along the lines (Fig. 2). This is the analogue of the functional definition of a surface with parallel equally-spaced grooves. These are indeed the kinds of pattern judgment that the human eye makes most precisely. Our peculiar sensitivity to changes of curvature and to non- parallelism and non-periodicity is well known in model-making and in pattern tracing and analysis. An extreme case is the curvature-continuity judgment involved in 'vernier acuity'. If two ends of a line join imperfectly in the middle, the eye can still perceive the break when the lateral displacement is as small as 2 seconds of arc — l/30th the diameter of a retinal cone (7). Regardless of what neural connections might be needed to make such a discrimination, it is obvious that the judgment must depend on a physical operation of inherently high precision, inherently Functional Geometry and the Determination of Pattern in Mosaic Receptors 379 unlimited by the coarseness of the mosaic structure and the randomness and uncertainty of cone locations. Functional geometry offers such a method, since it can generate indefinitely high precision out of arbitrarily coarse materials crudely manipulated. With it, the practical limitation in precision could be a very refined signal-to-noise limitation, that is, an intensity-judgment-time limitation, as we shall see, and not a coarse mosaic structure limitation. For the internal as well as the external eye, it would be characteristic of biological systems to make use of such a method, conceptually simple, operationally precise, making only minimum demands on the accuracy of assembly, and capable of being driven to higher and higher precision as needed under the pressure of natural selection. III. DETERMINATION OF ADDRESSES A. Scanning in Vision DiTCHBURN and co-workers (8-10), and Riggs and co-workers (11-12), have shown that vision disappears unless the field is continuously scanned by the eye. The scanning is normally provided by 'physiological nystagmus', or 'fixa- tion tremor.' When a subject is fixating as steadily as possible, the following eye movements are present : '(i) a tremor of amphtude of the order of 15 sec arc and frequency ranging from 30 to 80 c.p.s. '(ii) a series of 'flicks' of up to 20 min arc occurring at irregular intervals ranging from 0.03 sec to 5.0 sec. '(iii) slow drifts in the intervals between flicks.' (Ditchburn (9)). The movements are involuntary. They continue undiminished even when an image has been stabilized on the retina, so that they do not seem to have quantitative feedback character, at least for fixation of a point source ; but the flicks do tend to produce recentering after the image begins to drift off" the fovea. The frequency, amplitude and sequence of the movements as presently known would be consistent with assigning the tremor jerks to the successive single neural spike inputs in the normal trains of spike pulses to the ocular muscles ; with assigning drift to the unbalance between these jerks in opposed muscles at slightly different spike frequencies; and with assigning flick to a final sudden burst of spikes to the less active muscle which redresses the un- balance and recenters the system. There is no vision during the flick movement. This demonstrates an intimate oculomotor interaction with the retinal output, complementary to the interaction which will be postulated later. When these movements are stopped by optically stabilizing the retinal image, vision is lost within a second or two. It can be restored by flicker modulation of the light intensity or by reintroduction of some image movement. The necessity for scanning in maintaining vision might have been anticipated. It is probably no surprise to a biologist to find such a mechanism used to counteract the effects of adaptation and fatigue in receptors that need to be continuously sensitive; nor to a biochemist to find that the electrochemical shock waves corresponding to nerve impulses are reduced in frequency as the photochemical steady state is approached; nor to a physicist to find that a.c. operation of a phototube is the best way to avoid d.c. drifting. Many retinal 380 John R. Platt potential and neural spike results obtained with steady-state illumination may have to be reexamined for their relevance to the process of vision. Insects and amphibia and other lower orders seem to hold their heads and eyes rigid for long periods. The need for scanning suggests this might permit selective detection of moving objects in the field. (The insect eye perception problem treated earlier was an artificial way of pointing up the general pattern problem, and not an attempt to describe the real workings of the insect eye.) At some point up the scale, a scanning tremor in the eye might have been an evolutionary predecessor of wide-angle motion. (It is not only vision that requires 'scanning'. A variation of input stimulus is needed to maintain sensitivity in touch and in smell. This strongly suggests that a search be made for a similar mechanism in hearing, by which the 'sound image' might be scanned up and down the basilar membrane to provide continual change of stimulation, to prevent local fatigue, and to sharpen tonal discrimina- tion.) B. Determination of Addresses In any field of study, it is always a hopeful sign to find two or more un- explainable effects and not just one; for this opens up the possibility that the two will explain each other. In the retina, we are confronted first with the pattern-perception address-determination problem and then with the strange importance of scanning. Putting these together, it appears that scanning might be a particularly straightforward method for functional determination of addresses in a non-addressed mosaic receptor. And this is functional geometry. Several theorems suggest themselves. The Fundamental Operations 1. Sequence of Elements — During random scanning over visual fields con- taining some structure such as sharp discontinuities or boundaries, if retinal elements /, y, k are triggered in similar patterns in succession far more often in the time-sequences ijk or kji than in the sequences y/7:,yA:/, ikj, or kij, then: (la) there are some boundaries in the external field that are relatively stable during the scanning motion; (lb) y lies on the image of a point in the field between the corresponding points for / and k ; and (Ic) the eye movement for one of the sequences ijk is opposite to that for the other kji. 2. Collinearity — During random scanning over visual fields containing sharp boundaries, if all the elements in a certain large set fgh • • • k are excited simul- taneously in the same way (d.c. ; or a.c, as by tremor across a boundary) and if this excitation continues unchanged throughout a short drift movement, then: (2a) there is a linear boundary in the field; (2b) elements y^/j • • • A: lie on the image of that boundary; and (2c) the drift movement is parallel to that boundary. The photodetector inputs produced by tremor movement could provide a gradient discrimination across the boundary which, when combined with drift, as suggested in Fig. 1 for a curved line, could give an especially delicate deter- mination of addresses (3). Thus, for cells distributed roughly along the image, one traverse might produce firing in a reproducible sequence fkghjfigf • • •, the Functional Geometry and the Determination of Pattern in Mosaic Receptors 381 sequence depending on some function of the relative transverse cell displace- m;nts, sensitivities and repetition rates, and on the image boundary gradient. 'Unchanged excitation' would mean successive recurrences of this same sequence, and changes in the sequence could correspond to changes in the image amounting to only a small fraction of a cell diameter. Jhc gradient-discriminating power would then be limited essentially by signal-to-noise considerations rather than by mosaic structure and it might be far higher than the static mosaic resolving power, as numerous authors have suggested. The transmitted self- congruence signal, whatever it is, need contain no trace of the static mosaic structural irregularities. It is also independent of differences in the sensitivity of different receptor cells and could remain unchanged even if a few of them should fail completely (closure). These would be important biological advantages for the self-congruence method of address-determination. 2'. Parallelism — If the elements/^/? • • • A' of Operation 2 also are grouped into /■ subsets each of whose excitations can be duplicated for r different trans- verse displacements, with a different set of displacements for each subset of elements, then: (2a') there are r parallel Hnear boundaries in the field ; (2b') each subset lies on the image of one of these boundaries; and (2c') the first drift movement is parallel to the boundaries, while the discrete transverse displacements are not. It is typical of functional geometry that it simultaneously limits (2a) the type of external pattern that can be interpreted (2b), the type of internal relation- ship that can be organized, and (2c) the operational motions that can produce a coincidence of the two. This situation for pattern structure is no different from that for the eyeball, where functional geometry simultaneously limits the shape of the external socket, the shape of the ball, and the possible movements and musculatures. We shall see over and over that the functional geometry, if it is the address-determining method, is neither experience nor structure but stands outside them both, imposing an inescapably limited selection of forms on the only experiences we can perceive and the only structures we can create. Point (2c), the establishment of retinal relationships and proprioceptive oculomotor signals relative to each other, as suggested by Helmholtz, is not the least important aspect of address determination, now that proprioceptive muscle spindles are known to be present (13, 14). Note that it is the boundaries in the external f eld that are linear, and not their retinal images, when self-congruence is the method of discrimination. Likewise the projections on the cerebral cortex can have any kind of twist, distortion or discontinuity— which they have — without destroying a functional definition of collinearity and parallelism. The ambiguous word 'linear' is used in theorems (2) and (2') so as to postpone for a moment the question of how well these procedures will distinguish a perfectly straight line from a perfect circular arc of very slight curvature. But aside from that question, a mosaic detector is seen to be in principle far more accurate than a single-element detector. With the latter, straight lines could be discriminated by tracing them out, perhaps using small hunting movements superimposed on a long sweep, but the accuracy is limited by the accuracy of 382 John R. Platt the analog position-sensing circuits. With mosaic detectors, the sensitivity to imperfections in a hne pattern is not affected by analog errors. In these theorems, a discriminated boundary will be perfectly straight if the longitudinal Z-axis of the eyeball passes through it and if there is no Z rotation during the scanning motion. Eye movements about the Z axis during fixation seem not to have been measured as yet, but it seems unlikely that physiological tremor about X and Y would be unaccompanied by tremor about Z. It will be convenient here to consider the analogue of these theorems for pure rotation about Z, with no X and Y components, and to come back later to the question of how the present theorems will be changed, and what new theorems will be valid if all three rotations are present. 3. Circularity — During pure rotational scanning about the Z-axis over visual fields containing sharp boundaries, if all the elements in a certain large set fgh • • • k are stimulated in the same way and if the stimulation continues unchanged throughout this kind of scan, then: (3a) there is a boundary in the field which is a circle or circular arc as seen from the eye ; (3b) elements y^/? • • • k lie on the image of that boundary; and (3c) the-Z axis of the rotation passes through the center of the circle. 3'. Concentricity — If the elements/g/; • • • A: of Operation 3 are grouped into /• subsets whose elements can be re-excited in the same local patterns by discrete sets of X, Y rotations (in analogy with the Operation of 2') then: (3a') there are r concentric circular boundaries in the field; and so on. Concentricity is also one of our very delicate discriminations, as is shown by many gunsight designs. 4. Translational Comparison — During a random scanning drift movement (in X and Y alone) over visual fields containing sharp boundaries, if a certain time pattern of excitation of elements bed • • • is repeated after a certain fixed time-delay (that is, a certain displacement) by elements y^/j • • • in a one-to-one correspondence with the bed • • • excitation pattern, then : (4a) there is a stable pattern, fixed or undergoing translation, in the external field; and (4b) there is a constant translational separation in the field between points whose images fall on elements b and/, e and g, d and //; and so on. 4'. Translational Periodicity — If the elements bed • • -fgh • • • of Operation 4 can be divided into r subsets, where r is greater than 2, each of whose excitations can be duplicated for r different displacements, with the excitations of several subsets simultaneously duplicated for certain displacements, then (4a') there is a stable translationally periodic pattern in the field, with r repetitions; and so on. Translational comparison is of course the theoretical procedure for estab- lishing a metric in a space of unknown geometry. The precision of translational inter-comparisons between the patterns in the two eyes is the basis of depth perception. Under the most favorable conditions it approaches the same high angular precision that was found in vernier acuity. This fact alone seems to require a physical operation that can create a high-precision translational metric Functional Geometry and the Determination of Pattern in Mosaic Receptors 383 for the retinal elements: an operation for each eye that can translate elements with fixed relations from one part of the field to another — that is, scan — during the intercomparison process. A figure with bilateral symmetry, whenever its median line is defined, has translational periodicity for lateral scanning. This might be the basis for whatever accuracy the human eye has in judging such symmetry. 5. Angular Comparison— During pure rotational scanning about the Z-axis in fields containing sharp boundaries, if a certain time pattern of excitation of elements bed • • • is repeated after a certain fixed time delay by elementsy^/? • • •, then : (5a) there is a stable pattern in the field, fixed or rotating about the Z-axis; (5b) there is a constant angular separation, with respect to the Z-axis, between elements b and/, c and g, d and h; and so on. 5. Angular Periodicity — If a relation like 4' is satisfied for pure rotational Z displacements, then: (5a') there is a stable angularly periodic pattern in the field, with /• repetitions ; and so on. Because of the limited range of Z-rotation in the human eye (about 20°), our perception of angular periodicities may lose precision rapidly for larger angles. Some of this acuity may be recovered by treating the angular periodicity as a bilateral symmetry, converting the judgment into one of lateral translational periodicity. The metric of the 'space'' of the addresses established by these operations is that of the rotation-space of the eyeball and not that of the retina or cortex surface. All these operations have been internal operations, specified so as to depend only on the internal properties of the decision net and its scanning system, and to be as independent as possible of the object and properties of the external field except for the minimum requirement that there is some variety of structure and that at least sometimes its changes and motions are slow compared to those of the eyeball. Displacements and motions in the external field could lead to another similar list of theorems which would establish similarly an external metric and an expected external behavior, whose familiar translational and other con- stancies— comparable to the congruences produced by the internal operations — we might finally interpret as invariant 'objects' in the field, uniform motions, and so on. The external metric may or may not be consistent with the eyeball- rotational metric. Probably the external metric is the primitive one, with the scanning metric providing a sophisticated refinement. Inconsistency between the two may be the source of many optical illusions. But since theorems (1) to (5) suffice to establish in several different ways that consistent address determination within the network is at least physically possible it seems more important to go on now to see how it would be anatomi- cally possible. C. Possible Types of Neural Connections Required Proprioceptive Coordinate Specif cation What anatomical connections are needed for proprioceptive sensing and control? The requirements and some possible ways of solving them, at least 384 John R. Platt for an artificial system witli some quasi-neuronal properties, will be seen if we consider the problem of scanning along a curved line, with Z-rotation of the retina to follow the curve, as in Fig. 1. If the differential muscle stress or strain or rate of change of either (whichever is the principal sensed variable) about the Y-axis has a fixed ratio r to that about the Z-axis, the eye will sweep up along a line at an angle arctan r to the horizon. If the differential neural spike frequencies/^ and/^ from the two muscle pairs give a quasi-logarithmic representation of the muscle action, a fixed ratio r corresponds to a fixed frequency difference,/^ — /»> which we can call F. A subtractive-frequency mixer tube, and perhaps a similar subtractive mixer neuron, could be devised which would combine two synaptic inputs so that an output pulse is produced only when the input pulses are simultaneous, as suggested in Fig. 3. With suitable cell sensitivity and time constant, this output is the beat f,-AF-R Fig. 3. Possible proprioceptive analog connections for scanning along a uniform curve. frequency, F, the difference of the two input frequencies (1). A neuron used in this way might be called a difference cell. (Determination of the sign of the difference might require another cell.) If the output frequency F is fed back with proper sign to the Jf and 7 muscles, a constant direction of motion can be stabilized. The rate of change of direction could be sensed by introducing a time delay and comparing r{t) with r(/ — A) or F{t) with F{t — A) by a second subtractive neuron which generates the frequency AF — a time-differential cell. With a lower sensitivity and a shorter time constant, the same type of cell could be made to fire only for a certain constant pulse interval in the inputs, equal to the time delay. This would be a constant-frequency detector, and could be called a nidi cell. A constant Z-rotation of the retina to follow the change of X, F direction in scanning a curve could be detected by a third subtractive neuron which generates the frequency difference/ — F. Call this frequency R. It can again be held at a constant value by suitable feed back into the Z motion, as shown at the right side of Fig. 3. Such a tracking motion permits eeneralization of theorem (3) and (3'): ^ Functional Geometry and the Determination of Pattern in Mosaic Receptors 385 3". Decentered Circularity — During combined translational and rotational scanning, if the conditions of theorem (3) or theorem (3') are satisfied, then (3a) and (3b) or (3a') and (3b') are vaHd; but: (3c") the Z-axis of the rotation does not pass through the center of curvature of the boundary or boundaries. The oculomotor motions that could be directed by this kind of analog control correspond to the crude motions of the grinding machine in generating spherical surfaces. They do not need to be exact. They need only to be capable of making retinal displacements across the field sufficiently good that during the tremor movements the retinal excitation will have an adequate chance to signal if it is congruent with the original pattern. This signal could also interact with the oculomotor system to make the scanning more stable and accurate. In practice, visual acuity for moving patterns is considerably reduced (15), presumably because of the increased tracking errors and decreased chance of a congruence signal. We can now see the effects of combined motions on theorems (2) and (3). Evidently Operation (3), the examination of circles, gives a functional self- centering specification of the axis of rotation, even if it is off the Z-axis, whenever congruence is maintained. Any tremor about other axes simply provides useful scanning motions. But Operations (2) and (2'), the examination of collinearity and parallelism, cannot discriminate perfect straight lines from perfect circular arcs of large radius except by invoking the accuracy of the sensing and analogue control, a discrimination of much lower accuracy than the mosaic self-congruence dis- criminations. It seems that our perception of such differences is in fact small unless there are known straight lines nearby permitting a self-congruence test for parallelism. There is a familiar illusion in which a comparison straight line appears curved in the opposite direction from a curved line. This shows an uncertainty in the analogue system, which tends to scan along the bisector so as to give the figure bilateral symmetry. The Z rotations of the eyeball during scanning of curved patterns evidently deserve examination. In the classical ZoUner illusion (parallel lines appear to be non-parallel when crossed by oblique converging lines) there might also be a Z-rotation of the retinal coordinate system, perhaps in the sense of stabilizing the local foveal pattern and the local bilateral symmetry axis as the fixation point oscillates from one of the parallel lines to the other. With further combinations of difference cells and time-differentiation cells and feedbacks, probably the tracing out of any pattern by scanning could be converted at a high enough stage into a constant output from some subtractive neuron. If adjustments of the scanning rates at various points in the pattern are also introduced, probably changes of size and distortions of shape could even be accommodated in a constant output at a still higher stage neuron. We can dimly visualize how this might proceed by stages to a neuron capable of producing a fixed output, or better, a total motion of advance or retreat, whenever so specific an object as a particular person is scanned, regardless of aspect and light. Even if in later life the proprioceptive tracing of patterns by scanning becomes subordinate to direct mosaic pattern-perception, the long stages of 386 John R. Platt finger tracing of block letters and large patterns by children and newly-sighted adults suggests its early importance. Studies of the developmental pathology of pattern-perception with partial oculomotor paralysis might be instructive.^ Null Detectors and Delay Lines What kinds of neural connections might be needed in the mosaic to deter- mine addresses as these operations are performed ? The discriminating self-congruence information is always of the form 'constant repetition of the same pattern' or 'repetition after a time delay'. What is needed from the photodetectors is the information 'absence of change'. They are being used as null detectors. Unstable detector elements in the labora- tory are often used in the same way whenever the utmost accuracy of measure- ment and simplicity of interpretation is wanted. The rather complex relation- ships that can be established when using mosaic receptors as null detectors seem not to have been explored before. To signal 'no change' we could use a null cell of the type already described. But another good way to examine stability of pattern would be to have a cell with two input channels of different lengths, like two of the channels in Fig. 4, Image Fig. 4. Cone addresses from delayed coincident pulses to a null-transmitting velocity-detector cell. or of different diameters and travel times, where the cell sensitivity is such as to require simultaneous spikes from both channels in order to produce an outgoing spike in its axon. (The word 'channels' is used to avoid the experimentally unsettled question as to whether these could be all-or-none dendrites of the cell, if such exist, or two excitatory synapses with different delays in the axons, or collateral processes from the cells of the preceding stage.) Such input channels would be delay lines like those used in nuclear physics to distinguish certain events and particles, to eliminate random spurious counts, and to measure velocities. They might be used for all these purposes here. The axon output of such a cell, as shown in Fig. 4, combines three im- portant properties. It is (a) a null indicator, firing only for those patterns that are identical in the input channels. It is (b) an indicator of a particular delay time or difference of times in the channels. And it is (c) a pattern transmitter, since the pattern is not lost. Let us call this a null transmitter cell, and if the Functional Geometry and the Determination of Pattern in Mosaic Receptors 387 delay is different from zero, a delay cell. Such cells would have several possible special applications. In the auditory system, a delay cell with binaural channels would be a useful direction-indicator . In the retina, a second-stage delay cell like that in Fig. 4 would indicate a particular image velocity-component in the plane of the paper, and could be called a velocity detector cell. One with several inappropriate delays might be sensitive to almost any motion or flicker. This may be one of the functions of the widely branching horizontal cells that are so numerous at the periphery of the retina, since this is a region particularly sensitive to motion. Velocity detector cells would give useful correction signals to the oculomotor system. Under Operation 1, it is easy to see how delay cells with input channels from retinal elements /, j, k might be preserved in the organism if their time lags are in either the spatial sequence ijk or the sequence kji, but might atrophy from disuse or at least rearrange their channels if their time lags are in any other sequence. And the ijk delay cells would be a different group from the kji cells. Such a principle of natural selection and differentiation might be applicable to all types of second-stage and higher-stage cells. It may be profitable to examine these or other kinds of time-delay connections in trying to make a model of color-vision, since it now appears that this may involve a comparison of signals from cones at different times as the photo- chemical substance in each one goes through some time sequence of spectral transformations under illumination and perhaps under scanning. To summarize these exploratory notions, it appears that the types of neural connections that would be useful for address determination, at least in an artificial system, would include: difference cells; time-differential cells; null cells; null-transmitter cells ; delay lines and delay cells ; and velocity-component cells. The outputs of such second and third-stage cells apparently can signal all the self-congruences required in the basic operations of functional geometry. All the geometrical patterns defined by local group theory congruences can be signaled without using the retinal elements in any way except as null detectors. If natural selection favors those cells, together with their oculomotor connections, which signal repeated self-congruences under scanning; then in the mature organism each retinal cell will feed into many second-stage cells, each of which expresses a useful functional relationship between that retinal cell and some others. The address of the cell has indeed been determined. The mature network-address becomes an expression cf the space-address. Evidently the functional geometry of scanning a pattern is a way of con- verting its space congruences into identical time patterns. It therefore could be said, as some have said (16), that in the mature organism the appearance of a certain time pattern at a certain point in the network creates an 'expectation' of its repetition at an adjacent point, and stimulates oculomotor and other movements normally appropriate for the accurate fulfillment of this expectation. The accuracy of address-determination in these operations depends on the temporal accuracy of the delay lines and not on the spatial structure, which can be largely eliminated from the congruence signals, and therefore from the perceived patterns. 388 John R. Platt In the adult, the need for continuous redetermination of addresses becomes smaller. As Dr. W. A. Arnold of the Oak Ridge National Laboratory pointed out when this paper was first given, the words on a printed page that is illuminated for only 10""* sec, too short for any scanning, can be read (by anyone who can read) quite normally over an area comparable with the foveal diameter. The addresses have already been established and need little if any reconfirmation from the operations that generated them. How complete and rigid they are might be discovered if we knew the limits of developmental distortion of the retina after, say, the age of 3. Or if adults could work with optical systems giving subtle distortions of a few minutes of arc in the shape and topology of patterns in the foveal region, to see the effect of the loss of addresses on line, circle and pattern perception, on reading and the identification of persons, and how soon — and how far back in the network — a new set of addresses would be learned. IV. NECESSARY PROPERTIES OF NON-ADDRESSED SYSTEMS Certain properties would seem to be necessary characteristics of at least the early stages of all non-addressed and address-determining networks. What is interesting is that many of these seem to be familiar aspects of higher human behavior, restated in receptor-network terms. We may properly inquire how far our more complex activities can be subsumed under a generalized functional geometry, and how far our more complex experiences are organized by means of generalized displacement congruences. A. Operational Characteristics The null-transmitter delay cell which it seems necessary to invoke for any network making time-delay comparisons of patterns would be a suitable prototype for much of our higher neural organization. In the mature organism, after the structure and connections of such a cell have been stabilized — that is, after the addresses of its input channels have been determined — the cell will always collate two or more input patterns in a standard way to produce a simplified vital output. Let us focus attention, first, on the nature of the outputs, then on the inputs, and finally on the process as a whole. Abstraction of Invariant Pattern Properties The output of each such cell signals to the higher stages the presence of some particular kind of simple or complex pattern in the lower stages. This implies in turn a pattern in the first-stage images, representing a pattern in the external field. The process is abstraction. Pattern is another name for congruences or invar iances in this field. At a given instant, some fifth or tenth stage cell may be signaling That is the letter R'. Simultaneously some other set of elements and delay lines is abstracting from the same retinal elements the information 'It is in my wife's handwriting'. A third neuron connected to these elements says, 'It is black'; a fourth, 'It is large'; and so on, for details and context and background and all the other components. Perhaps some still higher neuron also signals the unification of these separate pattern properties and others into a word. Functional Geometry and the Determination of Pattern in Mosaic Receptors 389 glanced past in a tenth of a second. These signals have a one-to-one corre- spondence with Platonic properties of 'R-ness', 'Blackness', 'Largeness', and so on. Of course, a pre-addressed network, might also give the same kind of infor- mation that any of these neurons gives. Or it might give information trivial or inscrutable for us, such as 'Slope of sharpest edge, 103° 8"; or 'Five corners, two arcs concave to the left'. In fact, any neuronal output in any connected network could be thought of as indicating some kind of pattern invariance. Jn this sense there are no addresses to learn! But most of these invariances in an arbitrary synthetic network would be worthless for biological survival. A major evolutionary problem for non-addressed systems must have been the facilitation of principles of connection leading to the appearance of cells, like the delay cell, capable of perceiving useful invariances, color, velocity, topology, and so on. Every internal invariance or imposed relationship of signals in time and space gives rise to essential redundancies that can be eliminated from the higher-order signals with no loss of external invariance information. There is a reduction by a factor of about 10^ between the 10^ elements of the retina and the 10^ elements of the optic nerve. Possibly this represents the elimination of some of the redundant scanning constancies of types such as those described in Section III that are implicit in the oculomotor operations and in the con- straints of the kinematic rotational metric. These regularities would then acquire an inescapable a priori character so far as the higher operations of the network are concerned. The external field may also contain many redundancies that are not impor- tant in a given situation. For many purposes it suffices to know that the animal is a wet, friendly dog, and we do not need the concurrent retinal information that he is opaque and continuous and in contact with the sidewalk. We have not considered here how such an 'attention' to certain patterns and suppression of others might take place. However, the facilitation of signals through one neuron by means of a change of its sensitivity produced by feedback from the 'expectations' of a higher-order neuron (representing an earlier wet-dog experience pattern) may not be different in principle from the facilitation of oculomotor tracking movements by feedback from the 'expectations' of second-stage or third-stage retinal velocity-detector cells. It may be helpful in many problems to think of attention and expectation as generalized tracking devices. The elimination, first, of field information that does not fit into the familiar useful second-stage patterns or categories, then of the redundant internal patterns, and finally of the temporarily unimportant and unattended-to external patterns, shows qualitatively how and why information is consumed in the course of abstracting invariances from a mosaic receptor (1). It is not lost or damaged by transmission in the sense usually considered in single-channel communication theory. Instead, it is used up, somewhat in the way that energy is used up in doing mechanical work. More mosaic input information is consumed in abstracting out a higher-level decision or invariance than in a lower-level one. It is consumed in the sense that the detailed input information is irrecoverable, non-reconstructable, from the output. Only by the almost 26 390 John R. Platt impossible process of applying m independent abstracted relations simul- taneously could the m independent inputs be inferred. But this consumption or 'loss' of information is not biological loss but a gain, since it represents the selection of the biologically relevant item from the confusing irrelevant flux. After a lifetime of suppression of the less valuable pattern and field details, adults finally attend only to the later neuron outputs or abstractions and seem to lose the eidetic ability to bring forth the exact early-stage patterns of instan- taneous retinal excitation, except as they can reconstruct them approximately from their appropriate or inappropriate collection of output neurons. This may show the cessation of early-stage rearrangements, which finally become completely pre-addressed as far as new experiences are concerned. Analogy Perception We may look not only at what has been abstracted — the outputs — but at what has been compared — the inputs. The elementary process in address- determination was the comparison of excitation patterns at two different times. A network whose neurons can signal identities or similarities of pattern — dis- placement congruences — is an analogy-perceiving network. Much, if not all, of what we call intelligence may be the abihty to perceive successive analogies at higher and higher levels of abstraction, a multiple repetition of a single basic neural process of organization. Artificial pre-addressed systems are not generally able to perceive any analogies except between those sets of inputs that they are wired up to treat as equivalent. The value of mechanical mosaic detectors such as the punch- card reading-head lies in the fact that they are wired up to perceive obscure informational analogies and not any of the space or time pattern analogies of the kind that we perform easily in retinal abstraction. Thoughts and Symbols Pattern and analogy perception resemble some important aspects of the higher-order process we call thought. We might make a limited definition of a thought as the realization of previously unperceived pattern-relationships. A thought could then be represented by an operator equation, where P^" is a pattern of the «th stage of abstraction ; Pg"* is one of the mih. stage; q is the time delay or other transformation operator which relates pattern P^ to P^. And Qph the {n + l)st or {m + l)st (whichever is higher) stage of relationship; it is the pattern of the P's, a pattern of patterns. It is primarily a realization signal — a displacement-congruence signal — but it may also contain some or all of the common elements of the P patterns. We might distinguish if necessary between the possibility of the thought and the continued existence of the thought-relationship; and between the insight, or first assertion of the thought, and the repeated use of the established thought. If P2'"* has no perceivable relationship to P-^", there is no thought, Po^qP^' = 0 for all q. Functional Geometry and the Determination of Pattern in Mosaic Receptors 391 The next stage of thought might be to perceive a relationship between two different g's, ^./'i V^i"+i = RQ m 12 and so on to stages of any order. This operator-form of the equation seems to be the simplest familiar form that expresses all the required relationships. It suggests that Q can be regarded as a characteristic value or output value of the relationship operator when the latter connects the state-functions P. Or g is a symbol of the relationship between P^ and P^. The equations suggest the possible usefulness of a formal calculus of abstraction. In the first equation, if either P^ or P^ or q is changed, Q is different and generally vanishes. A little reflection will show that it is typical of thought, as it is of such operator equations, that there is only a restricted class of pairs of P's that have any relation to each other. The ^'s are sharply limited at the same time as the P's; catalogues of the possible ^'s have been made by various philosophers. And there is only a restricted class of realization signals, Q, in any case. If the equation is to have a non-vanishing value, it imposes simul- taneous restrictions on all four variables. Some /*'s may never show any congruence. Some ^'s may operate forever in vain. And the Q's may take one or two values so repeatedly that they become independent of what particular P's are present. The significant thing here is that these same equations would also describe many of the processes and structures in an address-determining network. So, in the functional determination process, q could be the functional geometry displacement operation, P^ and P2 the geometrical surfaces or the geometrical patterns of excited cells before and after the operation, and Q the signal of self-congruence. In the velocity-component cell of Fig. 4, q could be the delay operation, Pi and P2 the pulse patterns in two of the input channels, and Q the coincidence signal output. In the neuroanatomical structure of the same hypothetical cell, q represents the delay line or lines, P^ and P, the input cells of the previous stage, and Q the cell itself or its output axon or other output processes. At a higher stage of the network, we might imagine that each of these structural elements in a particular neuron may also be connected to a verbal motor stage, with the relationships among these structural connections again represented by the same equations; and probably this structural parallel would be repeated in the language structure among the words themselves. An address-detennining system therefore seems to be necessarily a symbol- creating system. Whether regarded from the process aspect or the signal aspect or the structure aspect, a relationship or pattern of patterns in each case becomes represented by a symbol. The close parallel between neuronal transmission and logical operations has been discussed for many years and is of great importance in the com- putational and logical performance of decision nets. But the present equations and hypothetical models suggest that analogy-perception, 'closure', 'insight' and other apparent 'extrapolations' from the known — in short, thinking — 392 John R. Platt may also be a necessary normal and even rather simple aspect of the neural connection process in any system that can determine its addresses. B. Growth Characteristics Learning We saw earher that a non-addressed system must be initially incompetent and needs a long learning time. This learning or address-determination requires inputs containing pattern regularities — that is, experience. For the retina, the experience may be generated by the external environment alone, or from this environment as scanned by the eyeball; in either case it is external to the retina. In either case it generates spaces and metrics independent of the retina. Scanning is probably the visual counterpart of exploratory oral and manual manipulation which defines the 'spaces' of taste and touch. Probably the 'externahty' of the visual metric, plus the simplicity and universal identity of the scanning operations of all eyeball-spheres about their centers, help to account both for the Kantian a priori character, and for the pubhc and universal character, of visual space. This is contrasted with the situation, for example, in vocal or tone-quality space, which depends on the complex interaction of hidden muscular movements, and is perhaps the most incommunicable of our public spaces. The network can learn only those types of regularities it has experienced. Two networks should develop somewhat different pattern perceptions if their environmental regularities or scanning schemes are systematically altered. A non-addressed network which is forced to operate for a long time in a structureless environment, like a bhndfolded and insulated animal or human in the Riesen and Hebb experiments, should and does have seriously defective pattern-perception and response. One can see how the formation of simple and accurate early-stage addresses in a network would be very important in facilitating fast accurate pattern-perception at later stages. This picture of non-addressed learning exemphfies Hebb's conclusion (17) that many adult pattern-perceptions having introspectively the most instinctive and self-evident or necessary character are in fact perceptions that had to be learned at some very early age. It is early experience that selects the address- connections that are to be permanent; it is the permanent address-connections that create expectations and pattern-organizations in later experience. Nevertheless, there is a double paradox in the present picture, (a) There are possible external input experiences that cannot determine address connections. And (b) There are internal pattern-perceptions, just as there are eyeball-shapes, which have grown or have been learned and yet have not been determined by the particular experiences or motions that contributed to the learning process. Both these conclusions would follow from the operator equations of the last section that were supposed to represent network structure. The first point is obvious and almost trivial. External field patterns and their images are usable only if they fall within the limitations set by the network growth mechanisms and assembly principles. Motions of too high a velocity, patterns of far too coarse or far too fine a structure, fields of diffuse clouds with no Functional Geometry and the Determination of Pattern in Mosaic Receptors 393 sharp boundaries, dazzle patterns with the proper kaleidoscopic confusions of rapidly disappearing and reappearing spots, the diabolical fields of Ditchburn that refuse to be scanned — in fact, any patterns that deny the analogies that the network is prepared to detect — are probably all equivalent to a structureless environment in their failure to produce organization behind the retina. The second point becomes obvious when we consider functional geometry and the profound limitations it imposes. Just as the lens-grinding machine with sufficient freedom of motion necessarily produces spherical surfaces no matter what it starts with, so the scanning eyeball necessarily generates 'external' space relations or addresses corresponding to the continuous three- dimensional rotation group, no matter what is the structure of the external field. Likewise at the retinal level, any scanning retina necessarily acquires a unique perception for continuous lines of constant curvature and for parallel lines and lines or points periodically spaced, which it can never accord to patterns violating these displacement-congruences. These natural congruence relations may play the same organizing and aesthetic role in vision that octaves and simple frequency relations play in hearing. It is peculiar to functional geometry and it is extremely important for philosophy that these necessary relations are neither given to the visual system by any particular external field or experience, nor are they logically implicit in the structure of the network, even when we include the analogy-detecting structure. They are Q's that, like spheres, turn up invariably, no matter what the P's or ^'s. They have rather the character of geometrical preconditions simultaneously imposed on both the external field and the network organization if any learning is to be possible. And they are not imposed by the scanning operation, even though it does mediate between the field and the network — any more than the spherical shape is imposed on the lens by the loose random grinding machine, or on the eyeball by the muscles. They are more like a priori requirements, mathematical absolutes, that determine the only kinds of experience that can be organized and the only forms that learning can take, if learning is to be done at all. Functional geometry thus may be the origin of the Kantian epistemological limitations on thinking, as represented by 'the synthetic a priori categories of the apperceptive dialectic' Pitts has described this as 'perhaps the most fundamental problem of neurophysiology and psychology' (3). It appears that much, if not all, of Kant's theory of knowledge can be translated word for word into the language of inputs, structures and geometries in an address- determining network. Functional Storage A neuron that has been selected by experience, so that its output signals an experienced pattern, constitutes a storage of the experience — a memory. The storage is not 'dead storage' but a functional link which permanently changes the operation of the larger net and which remains part of it. The address-determining connections therefore constitute a functional storage of experience. At any instant, the net is the memory; the memory is the net. This kind of storage differs in an essential way from that of a prc-addressed 394 John R. Platt network, where the inputs do not permanently modify the connections. The latter must have a separate storage unit capable of modification and isolated from the main network except for controlled temporary interactions. So our electronic computers have their fast-access and slow-access storage units. Our social decision networks have their files and libraries more or less insulated from the functioning decision-personnel. This necessary but misleading subdivision of our artificial systems into operating units and storage units may be the reason why so many investigators in the past have searched — unsuccessfully — for a special memory organ in the human brain. Functional storage may be a typical property of biological systems, a further manifestation of their usual simplicity and efficiency. As an example, consider the genetic material of the cell, which at the present time is supposed to consist of a few species-specific macromolecules, such as DNA or RNA. In a newly-formed cell, such a molecule has two functions (although they might not be separate functions): to initiate the steps up the ladder of chemical syntheses of specific cell materials; and to duphcate itself. But this is functional storage: the chemical structure and reactions are the expression of the heredity; the heredity is the chemical structure. Likewise in the production of antibodies by antigens, the chemical record of the first antigenic experience is preserved in the antibodies (or in the chemical information in the antibody-producing cells), ready to find instant chemical expression when a second essentially identical experience occurs. The record is the specific chemical protection; the protection is the record. On a grosser scale, evolution is functional storage. The coming of the cold is shown in our fur and feathers and families. The record of the ancient temperatures and salinities may be in our blood and tears. The speed and efficiency of social decision networks might be increased if they could incorporate this lesson, and replace some of their file cabinets by continuously repeated appropriate functional modification in the decision channels. Time Constants Address-determination must go on at a certain regulated rate. This is probably faster for early-stage neurons and slower for later ones, but the order of magnitude should be well-defined for a given network. In the adult human brain, the indications are that roughly 50 milhseconds elapse between distinguishable perceptions or decisions — one 'moment', in the Stroud terminology. This is of the order of fifty of the milHsecond repetition intervals or synapse intervals of an individual cell, which seems to be a reasonable relationship (1). Knowing this time constant, we can make some numerical estimates of brain rates and capacities. These estimates are naive and probably false in detail, but they are explicit and rather instructive. Thus suppose that there is one new perception every moment and that it may be preserved in a memory, represented by a single changed neural connection. The now-classical experiment of micro-electrode stimulation during brain surgery shows at least that if certain points are stimulated, a complete, detailed and specific memory is indeed evoked. Combining this with the working hypothesis suggested by Quastler and others (18), that the Functional Geometry and the Determination of Pattern in Mosaic Receptors 395 waking brain appears to be processing input information at a constant rate, it would appear that a human brain may be making changed neural connections at rates up to 10^' per day. The necessary sequential spatial order in these con- nections might be the origin of our sense of temporal order in our memories. It may be no accident that this rate adds up to the order of 10^'' to 10" neurons per lifetime, comparable to the total number of neurons estimated to be contained in the adult brain; although of course a major fraction of these may be pre-addressed, unchanging after birth. (This number has also been computed as the minimum number of neurons required in a fully- developed decision-net serving 10^ to 10^ input elements (1). But there is no necessary conflict between these two points of view, since it is a familiar property of biological systems that they represent simultaneous optimization of different considerations— as in the two-point resolution of the eye, which is simultaneously limited by diff"raction, by aberrations, and by the mosaic cell size.) By this reckoning, less than one neural junction in a thousand would be changed per week, which might account for the difficulty of detection of histological changes. With such a specific moment-by-moment locahzation of new connections, the increasing loss of memory in older persons might be the result of cumulative damage to the neurons, such as radiation damage or microhemorrhages ; or it might be due to a kind of saturation of the address-determining connections, so that either no new relationships are perceived in the continuing flux of inputs, or else those that are perceived are no longer able to modify the network. These numerical estimates are not unreasonable; and even if the one-moment one-neuron assumption were dropped, it would not be surprising from the general dimensional considerations in the physics of the problem to find that that assumption would give correct order-of-magnitude relations between the time-constant, the lifetime and the number of neurons and its rate of change. Such a situation is common in order-of-magnitude calculations. But this estimate of the rates is defended only so that it can be attacked on other grounds: for it leads to another important biological dilemma, and one that might have an interesting resolution. For it must be remembered that the brain is not merely an electrical network; it is also a biological network — living, breathing, and growing. And a neural connection time of 50 milli- seconds is orders of magnitude too short for the usual cell growth time or atrophy time. While electrochemical channels or barriers might be formed or sudden changes of shape might take place in milliseconds, these can only occur for cells that are already present. A few years ago it was supposed that a way out of this dilemma would be to let the new perception or thought be initially established as a closed self- maintaining loop of neural electrical excitation, which could persist long enough afterward for the cell growth and structural change to take place. But a succession of apparently negative experiments seems to have caused this notion to be largely abandoned. There is an alternative. It is to let the neural growth take place, not after, but before the chemical and electrical connection to the network, as the elec- trician carries his coils of wire to the site before he hooks them up. The order- of-magnitude gap between the time constants can be got over by supposing 396 John R. Platt that the slow growth of new cells or random potential connections occurs in parallel, thousands or milHons of cells at a time, while the fast new decisions or perceptions or insights occur sequentially, hooking up one cell at a time or a small group. Such a sequence might resemble the activity-stimulation- proliferation-organization sequence in other tissue. And while this specific suggestion may again be wrong, its accuracy is less important than its general bearing on the time-constant problem, which suggests that epochs of growth may need to be separated from epochs of decision in a biological address- determining network. This possibility seems to deserve experimental inquiry. Perhaps our limited time span of intellectual attention, and the 'subconscious' solution of problems, and the role of sleep, especially in the infant, in preparing new cells to be ready for new (waking) connections or learning or decisions, should be re- examined from this point of view. C. Artificial Non-addressed Systems The truck driver is trained in cliildhood to perceive and respond appro- priately to cars, stop-lights and pedestrians of whatever kind. In this pattern and analogy-perception he excels any arrangement of photocells yet created. A pre-addressed decision-net might be able to operate with his small high- way tolerances and high speeds if it had his 10^-element resolving power and wide tield of view. But it would not be safe in the unpredictabihties of the open road. For this job, a non-addressed mosaic is needed, capable of learning new patterns. Otherwise the appearance of a new type of car or a new type of hazard on the road will cause the machine to be sent back to the factory for a complete rewiring of the circuits to establish the new invariances and their analogies with the old cars and the old hazards. It is important that the new hazard be recognized by analogy and not by trial and error. Direct highway experience would eliminate quickly a number of types of 'learning' computers that have recently been devised, in which the internal strategies are altered according to experienced successes or failures, but in which there is no pattern-extrapolation or 'insight'. The possible construction of artificial non-addressed 10"*- to 10^-element systems with complete decision nets and with 10^- to 10^-element outputs may deserve consideration. Primitive pattern-perceiving networks might be useful for narrowing the band-width of communication channels, if not for crude vehicle guidance. They might be useful internal elements in high-speed analogue and digital computers, where their stupidity could be partly com- pensated by the speed of operation. There they might simphfy the presently elaborate programming operations; and could speed up computations requiring many simultaneous substages of qualitative judgment or identification under distortions or transformations, where the total judgment is more elaborate than can be quickly represented by the coincidence of two digital words. The complete theory of artificial non-addressed systems with their many quasi-human characteristics will be fascinating. Evidently in many respects it may be simpler and more physical than present theories of digital computers and single-channel systems. It would include questions of optimization of different aspects of mosaic detection, such as rates and cell sizes, the proper Functional Geometry and the Determination of Pattern in Mosaic Receptors 397 balance among second-stage detectors of different kinds, and of foveal versus peripheral vision, the role and mechanism of fixation and attention, and the whole output-selection problem which has been ignored here. Theoretical consideration might lead to principles of neural connection, unused in biological systems, which would produce entirely different kinds of 'intelligence' in the organization of the input fields. Actual construction of at least the first stages of a non-addressed system might even be relatively easy. Since the receptor elements do not need to be wired individually, they can be laid down en masse, like the 10^ crystals of a photographic emulsion. The first-stage neuron layer, second-stage layer, and so on, could be laid down similarly. The crystals could not be compact in shape, like those of the emulsion, but would have to be interbranching needles. But the first successful device might be many orders of magnitude more complex than anything now made. To create such a device would require a number of really penetrating chemical or electrical inventions, but perhaps not a prohibitive number. Oculomotor outputs for scanning and tracking might have solutions close to the present standard single-element solutions. The main problem would be to guarantee that the neuron connections will tend to grow in such directions as to support any congruences in the chemical or electrical time-patterns, and will tend to be dissolved otherwise. With elements having 10~^ second time-constants (comparable to transistors) the potential learning speed of such a device would be 10^ times faster than that of a human being (2 hours = 20 years). Such speeds could not be fully realized because the initial address-determination will be limited by scanning speeds and motor-output speeds and by the chemical speeds of deposition of successive layers. But these potential speeds and these limitations are comparable to those of a digital computer; the latter being similarly held back by the slowness of programming and input and by the slow storage access speeds. A pattern-perceiving device so much faster than a human being and with a full range of inputs and outputs would pose grave problems of education, manipulation and control, problems different from those of a digital computer and more difficult; but the rewards would be correspondingly greater if these problems could be solved. Acknowledgement — This study was stimulated by a number of conversations with Professor Richard L. Meier, now of the University of Michigan. REFERENCES 1. J. R. Platt: Amplification aspects of biological response and mental activity. Amer. Scient. 44, 180-197(1956). 2. S. L. Polyak: The Retina, Fig. 40, University of Chicago Press, Chicago, III. (1941). 3. W. C. Pitts: Proc. Western Joint Computer Conference at Los Angeles, 1-3 March 1955, 108-110, Inst. Radio Engng., New York (1955). 4. J. Strong: Procedures in Experimental Physics, Prentice-Hall, New York (1938). 5. H. A. Rowland: The Screw. Encyclopedia Britannica, 1 1th ed. (1913). 6. J. Strong: New Johns Hopkins ruling engine. J. Opt. Soc. Amer. 4i, 3-15 (195\). 398 John R. Platt 7. W. H. Marshall and S. A. Talbot: Visual mechanisms. In: Biological Symposia, ed. by H. Kliiver, 7, 117-164, Jaques Cattell, Lancaster, Pa. (1942). 8. R. W. Ditchburn: Eye-movement in relation to retinal action. Optica Acta 1, 171-176 (1955). 9. R. W. Ditchburn and D. H. Fender: The stabilized retinal image. Optica Acta 2, 128-133 (1955). 10. R. W. Ditchburn: Stabilized interference fringes on the retina. Nature, Land. 177, 434 (1956). 11. L. A. RiGGS, F. Ratliff, J. C. Cornsweet, and T. N. Cornsweet: The disappearance of steadily fixated visual test object. /. Opt. Soc. Amer. 43, 495-501 (1953). 12. L. A. Riggs, J. C. Armington, and F. Ratliff: Motions of the retinal image during fixation. /. Opt. Soc. Amer. 44, 315-321 (1954). 13. S. Cooper and P. M. Daniel: Muscle spindles in human extrinsic eye muscles. Brain 72, 1-24 (1949). 14. a. S. Sunderland: A preliminary note on the presence of neuromuscular spindles in the extrinsic ocular muscles in man. Anat. Rec. 103, 561-2 (1949). b. N. C. R. Merrilees, S. Sunderland and W. Mayhow : Neuromuscular spindles in the extraocular muscles in man. Anat. Rec. 108, 23-30 (1950). 15. E. J. Ludvigh: Symposium on Physiol. Psychology, Pensacola, Mar. 10-11 Office of Naval Research Symposium Report ACR-\, p. 125 (1955). 16. D. M. Mackay: Towards an information-plan model of human behaviour. Brit. J. Psychol. 47, 30-43 (1956). 17. D. O. Hebb: The Organization of Behaviour, J. Wiley and Sons, New York (1949). 18. H. Quastler, e/ fl/. : Human Performance in Information Transmission, I-V, Tech. Report 62, 68, 69, 70, Control Systems Laboratory, Univ. of 111., Urbana, 111. PART VII THE STATUS OF INFORMATION THEORY IN BIOLOGY* A Round Table Discussion H. AUERBACH, C. EhRET, S. FrEED, L. S. FrISHKOPF, E. E. jACOBSf, B. N. Jaroslow, p. D. Klein, A. L. Koch, H. R. Mahler, H. Quastler, W. S. YAMAMOTOij: Edited by Henry Quastler Information theory is very strong on the negative side, i.e. in demonstrating what cannot be done; on the positive side its apphcation to the study of Hving things has not produced many results so far; it has not yet led to the discovery of new facts, nor has its application to known facts been tested in critical experiments. To date, a definitive and valid judgment of the value of infor- mation theory in biology is not possible. The first attempts to apply information theory to biological studies have been met with varying degrees of enthusiasm, ranging from outright rejection to statements like : 'Information theory furnishes a person with a sort of thread which would allow him to sense out a continuum in the order of the universe' ; 'A means of relating the existence of life to the non-existence of fife' ; 'A quest for regularities in irregular phenomena'. This is an extremely vast span of reactions to a proposition of admittedly limited scope. Many of the reactions refer not to information theory as such but more generally to interdisciplinary endeavours, and to system sciences, both of which are characteristically represented by information theory. Interdisciplinary meetings are always, or almost always, an exhilarating experience to all. They allow some sub-groups of scientists of a number of breeds to communicate with each other in a way that is in general impossible with the rest of the breed to which the particular scientist belongs. To put it another way, interdisciplinary meetings factor out scientists in a different way than occurs normally, and allow them fruitfully so to aggregate. Information theory, with its 'interdisciplinary' generalization of the entropy concept, provides a common meeting group for many disciplines; what is more, it has in many actual instances provided strong rapport between representatives of widely separated disciplines. The value of the communications aspects * On the evening following the Conference, eleven participants gathered for an informal session to discuss how they felt about the proceedings they had witnessed. The informal debate which ensued was transcribed and re-arranged into a coherent account. In doing this the editor tried to be objective. t Dept. of Physiological Chemistry, University of California. :|: School of Medicine, University of Pennsylvania. 399 400 Edited by Henry Quastler of information theory was demonstrated by the apparent harmony that prevailed at this three-day meeting despite the varied professional disciplines represented by the papers and the participants. But the devil's advocate will argue that of potential meeting grounds there are many, and the fact that information theory actually has provided one does not necessarily imply that it was an essential ingredient of success. We must not follow the lure of each and every interdisciplinary beacon; most actual results are still obtained within the safe limits of the estabhshed discipHnes. More important is the problem of system sciences in general — that is, of all sciences that deal with the whole rather than with parts, with general principles rather than detailed specifications, with patterns rather than specific mechanisms. This is a many-faceted problem. There are, first of all, the situations of the 'forest-and-the-trees' type; for instance, in physics: a complete description of every particle in a gas would contain implicitly all thermodynamic para- meters— but in this form the information would be useless. Or, to use an example in biology: if we knew the chemical constitution of all substances in all cells, together with all details of distribution, chemical kinetics, in brief, if we had reached the biochemical millenium — then we still would not necessarily know which of all these details are significant on the next higher level of organi- zation, although presumably this information must be impHcitly contained in the known details. Are we simply up against a psychological limitation? We seem able to think only of so much detail within any single train of thought. Faced with amounts of detail considerably beyond our mental capacity we begin to select: in the course of such selection important features are eliminated almost as readily as unimportant ones. Knowledge is not usable for human minds unless it is organized in blocks with not too much detail in each. There is more behind the desire to look at the whole rather than the parts than just an awareness of psychological limitations. There are relations within a whole which cannot be expressed in terms of parts alone. This is the fact expressed in the proposition 'the whole is equal to more than the sum of its parts'. The very mixed reactions which this proposition elicits are probably due to a failure to state exactly in what way the whole is more than the sum of its parts. There should be little disagreement if the 'more' refers to propositions concerning the whole which are qualitatively different from any proposition which can be made about any of its parts. Information theory provides a very convenient formahsm to state this situation : the total information content of the whole is exactly equal to the sum of the information contents of the parts — where the description of each part includes all possibiHties of connection with other parts; the mutual dependency of parts being organized into a whole causes a mutual reduction of uncertainty; therefore, the amount of non- redundant information associated with the whole is less than the sum of the information contents of the parts; the difference is exactly the infonnation content of the constraints, or all those propositions which apply only to the whole and not to its parts in isolation. This formulation may be of some help in making clear a puzzling aspect of the whole-parts relation. The preference for dealing directly with wholes rather than with parts is greatly supported by contemplation of the way living things are organized. The most striking feature is the existence of the organizational pattern with The Status of Information Theory in Biology: A Round Table Discussion 401 several distinct levels of organization. Some levels are more sharply delined than others, and the hierarchy of levels is not always unambiguous. Still, they are pervading enough that the intelligibility and validity of any statement in biology depends on proper agreement with organizational hierarchy. Now, one of the outstanding features in biological organization is that quite obviously only a small amount of the features obtaining at a given level has observable effects on the next higher level. Hence one of the most urgent problems, on any level, is that of detennining what details are involved in the communication to the next higher level — but this is precisely one form of the problem of the 'whole and its parts'. Thus, in studying the whole rather than its parts we seem to act as organelles, cells, organs do. So we have good reasons to beHeve in the importance of the systems approach. Still, it remains no more than a belief — and there exists an equally strong belief that only intense preservation of details will yield major biological breakthroughs and that it would be a 'young miracle' if really important contributions would come to biology without intensive examination of details. So we have extreme misgivings either way — and those misgivings seem to be destined to be with us forever. There exists no rigid calculus telhng which formahsm must be used on what data to achieve a major discovery. The present conference was arranged to explore the applications of infor- mation theory to the study of living things. This is a new field, and one cannot say, at this time, which approach is going to be most successful. Accordingly, the scope of the program was extremely wide. It was natural to question how much the various papers had contributed to furthering the purpose of the meeting. There was general agreement that some papers had contributed very much, some a moderate amount, some little or nothing; there was however, notable disagreement about which papers belong in which category. There exist a few cases where information theory was used in dealing with problems which could have been solved in other ways; and there are very many cases where problems have been solved by various methods which could have been, possibly, solved more easily by using information theory. The coin problem that Rapoport talked about falls in this category; so do the cryptographic studies of Gamov/ and YcAS. Information theory is so general that its domain of applicability is very broad; one cannot name one situation in which by the use of information theory one cannot get some understanding on what is happening on an abstract basis. But one is always beset by the niggling doubt that the application may not be proper. One can in many situations obtain results which seem to clarify understanding or increase the sharpness of a description to an extent which was not possible prior to the use of information theoretic methods. On the other hand, such results seem often suspended in mid air, away from the results of conventional disciplines. The important question then, at this nascent stage of affairs, is some- thing which is repellent to the scientific mind, the assessment of the 'worthwhile- ness' of the answers information theory seems to give. It seems plausible to assume that information theory should be useful where communication is critical, where messages are to be transmitted in the presence of noise, and where one might assume that some optimization is approximated; biologists are inclined to invoke the Darwinian mechanism of random trials with perpetuation 402 Edited by Henry Quastler of successful attempts; one is reluctant to admit any basic element of purpose — yet it might be better to bring it to the surface for a dispassionate inquiry. The question arose whether it was preferable to use information theory only in a semi-quantitative fashion, to account for general trends in observed data, or buttressed by measurements. The advantage of working with actual numerical estimates is obvious, but against it is the irreducibly relative nature of information measures. No unanimity existed concerning this question. There is general agreement that data properly usable are scarce, that there is a slight risk involved in using data from the literature which are inadequate for this purpose, and that the procurement of more pertinent and better data will yield material which would hardly have been produced otherwise. Future meetings might be designed to give stimulus and continuity to production of data which are more cogent and amenable to information theory. There was general agreement that further meetings should and will be arranged — and that information theory is here to stay in biology. AUTHOR INDEX Abderhalden, E. 81 Abelson, P. H. 83 Acher, R. 80 Ada, G. L. 90 Afzelius, B. A. 222 Akamatsu, N. 148 Alcock, R. S. 125 Allen, A. O. 270 Allmann, D. W. 61,124,126 Alper, T. 299 Anderson, E. 219, 222 Anderson, E. H. 300 Andjus, R. K. 178 Anfinsen, C. B. 80, 109, 113, 114, 116, 125, 126, 278, 289 Ard, W. B. 242, 252, 253, 256, 278, 354, 355 Armington, J. C. 379 Armitage, P. 344, 347 Aschaffenburg, R. 73 Ashby, R. W. 188 Askonas, B. A. 125 Astbury, W. T. 219 Astrachan, L. 139 Atwood, K. C. 309 Aub, J. C. 148 Auerbach, H. 347, 351, 399 Augenstine, L. 51, 61, 103, 106, 109, 114, 116, 120, 193, 194, 202, 218, 239, 287, 288, 289, 292 Bachofer, C. 290 Bahn,A. 81 Baker, W. K. 300 Baldwin, W. F. 306 Balis, M. E. 138 Balser, M. 34 Bang, F. B. 222 Bar-Hillel, Y. 194 Barltrop, J. A. 283 Bamafi, L. 80, 87 Barron, E. S. G. 258, 283 Barry, J. M. 72 Batty, I. 84 Baxter, Robert 337 Beams, H. W. 219, 222 Beerstecher, E., Jr. 207 Behrens, O. K. 79, 87 Bell, P. H. 75 Bellamy, W. D. 298 Bender, M. 86 Benedict, W. H. 294,308,310 Benfrey, B. J. 87 Benzer, S. 76 Berg, P. 131 Berman, M. 181, 182, 183 Bettelheim, F. R. 72 Bier, M. 113 Bigelow, J. N. 226 Binet, L. 322 Birch, L. C. 303 Black, F. L. 90 Blair, E. A. 154 Blair, H. A. 188,294,331, 333, 337 Blank, A . A . 41,42 Bleaney, B. 242 Block, R .J. 90,93 Bollman, J. L. 152 Boltman, E. T. 83, 138 Boltzmann 4, 325 Bond, V. P. 307, 308 Born, M. 265 Borysko, E. 222 Bourliere, F. 322, 328 Bowers, K. D. 242 Boyd, G. A. 294 Boyd, W. C. 212 Brachet, J. 88 BrambeU, F. W. R. 84 Branson, H. 101, 105, 106, 108, 109, 110 117, 193, 197, 198 Brawerman, G. 88 Brillouin, L. 21,198 Britten, R. J. 83 Brody, S. 327 Bromer, W. W. 79, 87 Brown, D. M. 86 Brues, Austin M. 294, 309 Bruice, T. C. 207 Bucher, N. L. R. 148 Buettner-Janusch, V. 86 Bulbenko, A. 61,124,126 Burdette, W. J. 53 Burger, M. 322 Burnett, G. 72 Burton, M. 289 Butler, J. A. V. 262, 269, 272 Calderon, H. 222, 223 Calkins, G. N. 222, 223 Calvin, M. 283 Cambrisson, J. 353 Campbell, P.N. 125 Cannon, W.B. 318,322 403 404 AUTHOR INDEX Carlson, J. G. 55 Carnap, R. 194 Carsten, A. L. 335 Carver, E. 106 Casarett, G. 294, 333, 334, 336, 337, 338 Cavallini, D. 86, 94 Chance, B. 180 Chandrasekhar, S. 321 Chantreene, H. 125 Chapman, W. H. 307, 308 Chargaff, E. 63,83,90,91,93,94 Chase, H. H. 39 Chatton, E. 219, 221 Cheng, H.G. 127,131 Christenberry, K. W. 294, 308, 310 Christensen, B. G. 143, 148 Clark, A. M. 306 Clarke, J. M. 307 Clayton, R. S. 294, 309 Cloudman, A. M. 294, 309 Cohen, B. L. 73 Cohen, M. R. 222 Cohn, M. 125 Cole, D. 80 Cole, L. J. 308 Collinson, E. 285 Combrisson, J. 241 Condliffe, P. 113 Condon, E. U. 248, 253, 255 Consden, R. 81 Cooper, G. 114,117,120 Cooper, S. 381 Corey, R. B. 108, 109, 110, 111, 117, 254 Comsweet, J. C. 379 Cornsweet, T. N. 379 Cote, Louis J. 95 Cotter, G. J. 84 Court-Brown, W. M. 345 Cowie, D. B. 83,106,113,131 Cramer, H. 324 Crane, H. 117 Crick, F. H. C. 51, 52, 53, 63, 65, 72, 90, 91,92,95, 117, 141,298 Cronbach, L. J. 192 Cronkite, E. P. 307, 308 Grossman, E. B. 306 Dainton, F. W. 285 Dale, W. M. 258, 290, 300 Dalhamn, T. 219, 222 Danchakoff , V. 126 Dancoff, S. M. 53, 57, 111, 190, 193. 202, 218, 298, 359 Daniel, P. M. 381 Danielli, J. F. 197, 221 Davey, W. P. 293 Davie, E. W. 81 Davies, J. 290 Davis, B. D. 73 Dawson, R. M. C. 126 DeBoer, J. 362 DeBusk, A. G. 88 DeFinetti, B. 231 DeFremery, D. 90 DeGroot, S. R. 198 Deich, T. L. 73, 75, 77 DeMarko, L. 86, 94 DeMoss, J. A. 131 DeRobertis, E. 219, 222 Deutsch, T. 73 Devine, R. L. 219, 222 Dewey, D. L. 71 Diandier, K. M. 132 Dimond, A. E. 278 Dirr, K. 79 Ditchbum, R. W. 379 Dixon, J. J. 75, 79 Dobson, E. L. 342 Dolansky, L. 24 Dolansky, M. 24 Doll, R. 344, 345, 347 Donohue, J. 142, Dorner, R. W. 90 Dounce, A. L. 71 Doyle, B. 271,274,288,289 Dreyer, W. J. 79 Dublin, L. I. 342 Dulbecco, R. 305 Dunnebacke, T. H. 222, 223 Duryee, W. R. 222, 223 Du Vigneaud, V. 74,80,81,87 Eakin, R. E. 207 Ebert, J. D. 125, 126, 129, 132 Edds, M. v., Jr. 39 Eden, M. A. 188, 359 Edsall, J. T. 267 Ehrenberg, L. 258 Ehret, C. 191, 218, 219, 222, 223, 224, 224-5, 399 Eldjarn, L. 258, 259 Elias, P. 34 Ellis, M. E. 308 Elson, D. 83,90,91,93,94 Ephrussi, B. 125 Epstein, P. 324 Erickson, J. 114,117,120 Erlanger, J. 154 Evans, E. A., Jr. 138 Evans, E. C. 354 Evans, J. V. 73 Exner, F. M. 300 Eyring, H. 117, 120, 267, 287, 290 Fahmy, M. J. 309 Fahmy, O. G. 309 AUTHOR INDEX 405 Falcone, G. 71 Fano, R. M. 11 Farmer, C. J. 130 Fan, P. 159, 160, 161 Fawcett, D. W. 219, 222 Feinstein, R. N. 84 Felix, K. 74, 79 Feller, W. 213, 361 Felton, — . 44 Fender, D. H. 379 Fenton, P. F. 39 Fine, P. C. 198 Fink, R. M. 294 Finkelstein, P. 121,283 Fisher, J. C. 347 Fisher, R. A. 4, 325 Flavin, M. 72, 80, 125, 126 Fleischer, S. 127, 131 Fling, M. 73 Flower, S. S. 328 Fox, S. W. 87, 101 Fraenkel-Conrat, H. 70, 73, 76, 77, 80, 81, 117 Francis, M. D. 125, 131 Franck, J. 239, 255, 257, 262, 263, 264, 267, 273, 287, 289, 291 Freed, S. 171,173 Frenkel, A. 294 Fricke, H. 271,272,283 Friedberg, F. 86 Friedberg, W. 126,127,131 Friedell, H. L. 335 Friedewald, W. F. 300 Frishkopf, L. S. 61, 153, 161, 164, 399 Fritz, F. 44 Fromageot, C. 80, 86 Fuerst, C. A. 280 Furth, J. 294, 308, 310 Gale, E. F. 131 Galton, F. 360 Gamow, G. 8, 51, 52, 61, 63, 71, 78, 83, 91, 92,95, 101, 107, 189, 194,401 Garen, A. 222, 223 Gates, F. L. 299 Gay, H. 219, 222, 223 Gergeley, J. 113 Geschwind, I. I. 75, 79, 80, 87 Gey, G. O. 148 Gierer, A. 70 Giese, A. C. 306 Giles, N. H., Jr. 55 Gladner, J. 113 Glinos, A. D. 61, 148 Godin, C. 125 Gofman, J. W. 343 Golay, J. E. 34 Goldblith, S. A. 280 Goldstein, L. 88 Gordon, A. H. 81 Gordon, W. G. 86 Gordy, W. 239, 241, 242, 251, 252, 252-3, 253, 254, 256, 263, 278, 289, 292, 353, 354, 355 Gowen, J. W. 54, 55, 293 Grahn, D. 293, 334 Gray, L. H. 262, 300 Green, D. E. 219 Green, D. W. 73 Green, F. C. 79 Greenstein, J. 114, 117, 120 Grier, G. W., Jr. 44 Griffith, J. S. 91,92 Grindlay, J. H. 152 Gros, F. 88 Gross, J. 131, 221 Grubb, H. M. 291 Guild, W. R. 269, 274, 278, 280 Gundalch, G. 79 Guthneck, B. T. 94 Guyer, M. G. 214 Haddow, A. 355 Hagen, W., Jr. 335 Hagihara, B. 84 Hald, A. 156 Ham, A. W. 150 Hamilton, K. A. 294, 309 Hampton, M. 84 Hancock, J. 360 Handa, D. T. 101,144 Harkness, R. D. 151 Harrington, W. F. 113,116 Harris, A. 125, 131 Harris, D. G. 174 Han-is, H. 73 Harris, J. 113 Harris, J. I. 74, 75, 79, 80, 87 Hart, E. J. 270 Hartley, R. V. L. 4 Haurowitz, F. 113, 115, 126, 127, 131 Hayes, P. M. 283 Hearon, J. Z. 143 Heath, D. C. 306 Hebb, D. O. 392 Heck, W. W. 72 Heimberg, M. 125 Hemmings, W. A. 84 Henshaw, P. S. 293, 333 Herr, E. B., Jr. 306 Herriott, R. M. 125 Hershey, A. D. 144 Hetzer, H. O. 54 Hijmans, J. 362 Hill, R. 113,116 Hins, C. H. W. 81,86,87 27 406 AUTHOR INDEX Hinshelwood, C. N. 143 Hirohata, R. 79 Hoagland, M. D. 131 Hodge, A. J. 219, 222 Hodgkin, A. L. 202, 203 Hoffman, J. G. 307, 308, 351 Hogness, D. S. 125 Hollaender, A. 54, 298, 299, 300 Holloman, J. H. 347 Holmes, B. 295 Homeyer, P. G. 87 Honnen, L. 79 Horio, T. 84 Horowitz, N. H. 73 Hough, A. 86 Howatson, A. F. 150 Howe, E. E. 71 Huettner, A. F. 219, 222 Huggins, C. 283 Hughes, W. L. 219 Huisman, T. H. J. 73 Hunt, C. C. 159,160,161 Hursh, J. B. 333, 334, 336, 337, 338 Hutchinson, F. 269, 274, 278, 280, 290 Hutchison, C. A. 241 Hvidt, Aa. 113,116 Ingram, D. J. E. 242, 353, 355 Ingram, J. E. 353 Ingram, V. M. 73, 80 Itano, H. A. 73 Jacobs, E. E. 399 Jacobsen, C. 109 Jacobsen, E. 148 Jacobson, B. S. 301,306 Jager, B. U. 86 Jaroslow, B. N. 211,399 Jarrett, H. S. 252 Jensen, E. V. 283 John, E. R. 234 Johnson, M. P. 131 Jolles, G. 86 Jones, H. B. 188, 329, 341, 342, 343, 344, 345, 350 Jonix, J. H. P. 73 Jordan, D. O. 141 Jutisz, M. 81 Kac, M. 362 Kahn, J. R. 75, 80 Kalman, S. 113 Kalnitsky, G. 113,116 Karreman, G. 191 Katz, B. 159,160,161 Katz, J. J. 178,284 Kaufman, B. P. 55 Kauzman, W. 109, 268 Kay, L. M. 79, 81 Keller, E. B. 131 Kelner, A. 298 Kendell, M. G. 325 Kendrew, J. C. 74 Kennedy, E. P. 72,116 Keynes, R. D. 202, 203 Kharasch, N. 207 Kihlman, B. 146 Kimball, A. W. 55, 305, 308 Kimmel, J. R. 86 King, J. W. B. 73 Kitai, R. 71,80,87 Klein, D. 193, 204, 399 Klemmer, E. 24 Knight, C. A. 90 Koch, A. L. 61, 101, 136, 137, 138, 139, 140, 141, 144, 145, 146, 239, 283, 284, 290, 399 Koechlin, B. A. 86 Komintz, D. R. 86 Konig, D. 365 Korik, S. B. 125 Koros, Z. 73 Koshland, D. 116 Krohmer, J. S. 335 Kroner, T. D. 79 Kunitz, M. 113 Kutsky, R. J. 133 Lagerstedt, S. 150 Laki, K. 86 Lambert, W. V. 54 Lamont, W. A. 139, 141, 144, 146, 284 Lampen, J. O. 139 Landmann, W. A. 271 Landsteiner, K. 215 Lange, K. 86 Lansing, A. I. 309, 310, 313 Lark, C. T. 138 Lark, K. 88 Laurila, U. R. 80 Lawler, H. C. 74,81,87 Lea, D. E. 55, 262, 270, 272, 289, 298, 301 Leak, J. C. 72 Lee, R. H. 307, 308 LeGette, J. 79 Leone, C. 271 Lepper, R. 219 Leslie, P. H. 304 Levinthal, C. 219 Levy, H. R. 101,144 Lewis, E. B. 345 Li, C.H. 75,79,80,87,113 Lietze, A. 127, 131 Linderstrom-Lang, K. 109, 113, 116 Lindgren, F. T. 131 Linschitz, H. 201, 359 Littlefield, J. W. 131 Livingston, R. 353 AUTHOR INDEX 407 Lloyd, D. P. C. 157, 158 Locker, R. H. 79, 81 Lockingen, L. S. 88 Loftfield, R. B. 125, 131 Loiseleur, M. J. 178 Lorand, L. 75, 86 Lorentz, E. 293 Lovelock, J. E. 178 Low, B. W. 253 Lucas, F. 81,87 Luck, C. F. 252 Ludvigh, E. J. 385 Lumry, R. 117, 120, 267, 287, 290 Luria, S. E. 299, 300, 305 Luzzati, D. 138 Lwoff, A. 221,222,223 Maalow, O. 88 Maas, W. K. 73 Mackay, D. M. 387 Magasanik, B. 88 Magdoff, B. S. 90, 93, 95 Mahler, H. R. 61, 124, 126, 180, 399 Marack, J. C. 74 Margenau, H. 182, 183 Markham, R. 90 Marks, H. A. 342 Marsh, W. H. 75, 80 Marshall, W. H. 378 Maylow, W. 381 Mazia, D. 71 Merrilees, N. C. R. 381 Metcald, R. G. 294 Metz, C. B. 219 Michel, R. 72 Middlebrook, W. R. 75 Milhalyi, E. 113 Miller, D. G. 173 Miller, E. C. 94 Miller, J. A. 94 Miller, J. G. 234 Miller, W. C, Jr. 280 Minick, O. T. 219, 222 Minto, W. L. 294 Mirsky, A. E. 267, 287, 288, Molnar, S. 234 Monier, R. 81 Monne, L. 218 Monod, J. 125 Montagna, W. 39 Montie, D. B. 90, 93 Moore, D. 222, 223 Moore, S. 81,86,87 Morgan, C. 222, 223 Morgenstern, O. 38 Morowitz, H. J. 101, 107, 108, 193, 239, 276, 277, 281 Mortimer, R. K. 301, 302, 310, 311, 312 Moscona, A. 222 Moshman, J. 294,308,310 Muller, H. J. 55 Murai, K. 79 Murphy, G.M. 182,183 Murphy, J. B. 126 Murphy, J. S. 222, 223 Murray, Wm. S. 307, 308, 351 Muus, J. 86 Myerson, S. 291 McConnell, W. B. 73 McCormick, G. 242, 252, 252-3 McCulloch, W. S. 234 McFadden, M. L. 86 McGarr, J. J. 79 McGill, Wm. J. 38, 161, 163, 164, 369 Mclntyre, A. K. 157,158 McLaren, A. D. 121,274 McLean, J. R. 131 McMorris, R. O. 201 Nagel, E. 222 Naughton, M. A. 74 Neumann, J. von 38 Neurath, H. 79, 81, 113, 114, 117, 120 Newman, H. H. 360 Newton, G. 131 Nickerson, W. J. 71 Niu, C. L 76, 80, 81 Noonan, Thomas, R. 335 Nord, F. 113 Nordling, C. O. 347 North, A. C. T. 73 Northrop, J. H. 125, 222, 223 Novelli, G. D. 131 Novick, A. 136, 137, 141, 145, 146 Nowell, P. C. 308 Nozaki, M. 84 Nybom, N. 301,306 Oakley, C. L. 84 Oberle, E. M. 280 Ohno, K. 80, 82 Okunuki, K. 84 Oliff, W. D. 303 Orgel, L. E. 91,92 Orlans, E. S. 74 Ottesen, M. 80,113,116 Ouchterlony, O. 214 Oudin, J. 211,214 Owen, J. 242 Owen, M. E. 302,310,311,312 Ozawa, H. 74, 76, 77, 79, 87 Pake, G. E. 252, 353 Palade, G. E. 219, 222 Paleus, S. 74, 79 408 AUTHOR INDEX Pappas, G. D. 219, 222 Pardee, A. B. 88, 144 Parish, H. D. 86 Parker, J. R. 328 Parker, S. L. 303, 304, 307 Parkes, A. S. 178 Parrish, R. G. 74 Passman, J. M. 80 Pauling, L. 108, 109, 110, 111, 117, 254, 267, 287, 288 Pearl, R. 54, 303, 304, 307 Peart, W. S. 75 Pecher, C. 154, 155, 156, 157, 159, 160, 161 Perkinson, J. D., Jr. 126 Perlmann, G. 113,116 Perry, B. T. 90 Pihl, A. 258, 259 Pitelka, D. R. 219 Pitts, W. 234 Pitts, W. C. 373, 374, 380, 393 Piatt, J. R. 107, 123, 291, 294, 371, 372, 389, 394, 395 Platzman, R. 239, 255, 262, 263, 264, 265, 266, 267, 268, 270, 273, 287, 289, 291, 292 Plaut, W. 88 Pollaczek, H. 79 Pollard, E. C. 262, 269, 272, 274, 278, 280, 287, 288 Poison, A. 83 Polyak, S. L. 373 Pomper, S. 309 Pon, N. 113 Pontecorvo, G. 76 Popenoe, E. A. 74,81,87 Porter, K. R. 219, 222 Porter, R. R. 79 Posternak, T. 79 Potts, B. P. 222, 224-5 Powers, E. L. 219, 222, 223, 224 Preer, J. R. 226 Prestidge, L. S. 88 Price, J. M. 94 Proctor, B. E. 280 Prosser, C. L. 318 Purvis, J. L. 87 Putnam, F. 114,117,120,138 Quastler, H. 3, 38, 39, 41, 53, 58, 111, 135, 187, 190, 191, 193, 194, 197, 202, 207, 209, 211, 216, 218, 221, 226, 281, 298, 351, 359, 369, 394, 399 Raacke, I. 113 Raacke, 1. D. 75, 79 Rabinowitch, — . 255, 257 Rahn, Otto 54, 298, 309 Rail, W. 159,160 Ramachandran, L. K. 73 Ramasarma, G. B. 90 Ramler, W. J. 270 Ramsey, N. F. 247, 248 Ranson, R. M. 304 Rapaport, A. 196, 230, 401 Rashevsky, N. 193, 222, 223, 225 Ratliff, F. 379 Ray, R. 114,287,289 Rebhun, L. I. 219, 222, 223 Redfield, R. 289 Reed, I. S. 34 Rees, E. D. 178 Ressler, C. 80, 87 Rexroad, H. N. 242, 252, 356, 253-3 Rhondin, J. 219, 222 Rich, A. 71,72,78,83,93,107 Richards, F. 113 Ridenour, L. N. 6 Riggs, L. A. 379 Rimoldi, H. J. A. 234 Robbins, K. C. 74 Roberts, E. 90 Roberts, R. 106, 113 Roberts, R. B. 83 Roche, J. 72 Rocklin, S. R. 270 Rogers, W. 113,116 Romanoff, A. J. 130 Romanoff, A. L. 130,131 Ronzoni, E. 86 Roos, P. 80, 87 Roper, J. H. 76 Rose, H. 222, 223 Rose, I. A. 139 Rosenberg, S. 335 Rosenblith, W. A. 61,153,157 Rossi-Fanelli, A. 86, 94 Rost, G. 94 Roth, L. E. 219, 222, 223, 224-5 Rowland, H. A. 376 Rupe, C. O. 130 Russ, S. 293, 294 Russell, W. L. 55, 294, 301, 312 Rylander, P. N. 291 Ryle, A. P. 71,80,87,283,284,286 Sacher, G. A. 293, 294, 301, 317, 322, 323, 326, 327, 329 Sachs, H. 131 Sanger, F. 71, 74, 80, 87, 283, 284, 286 Satake, K. 74, 76, 77, 79, 87 Savage, L. J. 231 Sayre, E. V. 173 Sbara, A. J. 300 Schellman, J. 113, 116 Schoenfeld, R. 181, 182, 183 Schott, R. G. 54 Schram, G. 70 AUTHOR INDEX 409 Schroeder, W. A. 79, 81 Schroedinger, E. 117,189 Schuler, R. H. 270 Schwartz, D. 71 Schweigert, B. S. 94,139 Scott, G. M. 293, 294 Scott, J. F. 140 Seegers, W. H. 86 Seguela, J. 219 Sekuzu, I. 84 Semmett, W. F., 86 Seno, N. 79 Setlow, R. B. 269, 271, 272, 274, 278, 280, 288, 289 Shannon, C. E. 4, 21, 44, 53, 187, 234, 298, 368 Shapiro, B. 258, 259 Shaw,J.T.B. 81,87 Shields, H. 242, 251, 252, 252-3, 253, 254, 256, 278, 354, 355 Shields, J. 74 Shimura, K. 79 Shive, W. 207 Shock, Nathan 344 Shortley, G. H. 248, 253 Shumway, N. P. 75, 80 Siekevitz, P. 131 Silberstein, H. E. 294 Silverman, R. A. 34 Simmons, Eric L. 335 Simon, H. A. 92, 95, 123 Simpson, M. V. 125, 131 Sinex, F. M. 72 Singer, S. J. 178 Sinn, L. G. 79, 87 Skanidze, I. K. 132 Sleggs, L. T. 75, 80 Sloan, G. J. 252 Smart, H.F. 178 Smellie, R. M. S. 88 Smith, A. U. 178 Smith, E. 113,116 Smith, E. L. 86 Smith, H. 71 Smith, J. D. 90 Smith, L. F. 71,80,87 Smith, M. J. 307 Smith, S.G. 81,87 Smith, W. V. 242, 353 Soreni, E. T. 73, 75, 77 Spiegelman, S. 125 Sri Ram, J. 113 Stacy, R. W. 201 Stadler, J. 55, 293 Stapleton, G. E. 298, 300 Staub, A. 79, 87 Steam, A. 114, 117 Steam, A. E. 268, 287, 288 Stein, W.H. 81,86,87 Steinberg, D. 109, 113, 125, 126 Stent, G. 280 Stepka, B. 301 Stetten, M. R. 72 Stevens, K. W. H. 242 Stockell, A. 86 StoraasU, J. P. 335 Storer, John 335 Strange, R. E. 71 Strong, J. 375, 376, 377 Sunderland, S. 381 Susman, N. 148 Suter, G. M. 294 Swanson, C, P, 55 Swick R. W. 101,144 Swift, H. 219, 222, 223 Symonds, P. 86 Szafarz, D. 88 Szent-Gyorgyi, A. 113 Szilard, L. 4, 136, 137, 141, 145, 196 Szorenyi, E. T. 73 TabrofT, W. 79 Tahmisian, T. N. 219, 222 Talbot, S. A. 378 Tapley, D. F. 283 Tartar, V. 223, 225 Taylor, J. H. 219 Teller, E. 120 Terminiello, L. 113 Thompson, A. R. 80 Thompson, E. O. P. 74, 79, 80, 81, 86 Thompson, W. D'Arcy 191,219,225 Thomson, J. F. 335 Tobias, C. A. 300,301,305 Tolman, R. C. 198 Tourtellotte, W. W. 335 Townsend, J. 252 Trambarulo, R. 242, 353 Trippett, S. 80, 87 Tristram, G. 289 Trucco, E. 191 Tumanishvili, S. 132 Tuppy, H. 74, 79 Turba, F. 79 Turing, A. M. 37, 360 Turner, J. E. 127,131 Tweedell, K. S. 360 Uebersfeld, J. 241,353 Uhlenbek, G. E. 320 Umbarger, H. F. 144 Upton, A. C. 294,308,310 Uretz, R. B. 299,300,305,312 Van Der Schaaf, P. C. VanderWaals 109 73 410 AUTHOR INDEX Van der Waerden, B. L. 362 Van Roggen, A. 252 Van Roggen, L. 252 Van Slyke, D. D. 72 Vaughan, M. 109,113,125,126 Velick, J. F. 86 Velick, S. F. 86, 125 Volkin, E. 139 Waelsch, H. 131 Walker, E. P. 328 Wallace, B. 309,312 Waiter, H. 124,126,127,131 Walton, B. P. 131 Wang, M. C. 320 Ward, J. C. 362 Warren, F. L. 73 Wassermann, F. 221 Watson, J. D. 51, 52, 53, 63, 65, 90, 117, 141,298,299 Watson, M. L. 219, 222, 223 Weaver, W. 4, 53, 187, 298, 368 Webster, G. C. 131 Weinstein, H. J. 222 Weiss, P. 125, 126 Weissmann, S. I. 252 Weisz, P. B. 39, 225 Wenneker, A. S. 148 Went, F. W. 360 Wertz, J. E. 242, 252 Westfall, J. A. 219 White, H. E. 248, 253 White, W. F. 74, 75 Wicks, L. F. 86 Wiener, N. 4,21,187 Williams, D. T. 201 Williams, G. R. 180 Williams, R. B. 126 Williams, R. J. 55, 207 WiUiamson, M. B. 80 Wilson, P. W. 144 Windsor, E. 71 Winnick, T. 125,131 Winzler, R. J. 208 Wohlfarth-Bottermann, K. E. 219, 222 Wolf, G. 72 Wolken, J. J. 222 Wollenberg, A. 80 Wood, J. L. 126 Wood, T. H. 300, 305, 306 Woods, P. S. 219 Worden, R. E. 201 Work, E. 71 Work, T. S. 125 Wyatt, G. R. 63 Wyss, O. 354 Yalow, R. 290 Yamamoto, W. S. 399 Yamashita, J. 84 Yates, R. A. 144 Yeas, M. 8, 51, 52, 55, 61, 63, 68, 69, 70, 71, 78, 83, 87, 88, 92, 101, 102, 107, 108, 189, 194,401 Yearkes, A. W. 328 Yearkes, R. M. 328 Yockey, H. P. 50, 54, 55, 188, 297, 298, 299, 308,327,338, 351, 359 Yuasa, Akira 219 Zamecnik, P. C. 131 Zavoisky, E. 241 Zcheile, F. P. 174 Zelle, M. R. 54, 298 Ziegler, J. 86 Zimmer, K. G. 258 Zinder, N. D. 222, 223 Zirkle, R. E. 270,300,301,305 Zwartouw, H. T. 71 SUBJECT INDEX Acetahularia 88 Acetyl DL-valine 252-3 (Fig. 7) Acroblast 219, 222 ACTH 69, 75, 76, 104, 107 Actin 79, 86 Additivity, in information functions 21 Address 191, 371 — , determination of 379, 380, 387, 394 Address learning 374 Adenine 63, 90, 136, 137, 138, 141, 145 — deoxyribosidc 1 38, 282, 294 Adenosine deaminase 146, 286 Adenylic acid 70, 90 Adrenocorticotropin 68, 79 Aging 188, 293, 297, 308, 317, 326, 339, 341, 344, 347 — , premature 293 Alanine 70, 90 Albumin 104, 114, 128, 252-3 (Fig. 10) — ' ^gg ^^^ Ovalbumin — peptides 128 Alcohol dehydrogenase 86 Aldolase 86, 288 Allele 73,311 Allelic states 190 Allelomorph 301,302 Alphabet 8, 15, 18,211 — , binary 8, 20 Ambiguity 31 Amino acid residues 50 sequence 105,135,194 ■ , phylogenetically common ancestral 107 Amino acids 18, 65, 71, 76, 84, 105, 128, 242, 252, 253, 263, 273 , X-irradiated 252 Amino-adipic acid, a 71 Amoeba 88 Amylase 1 14 Analogy perception 390 Anaphe moloneyi 87 Antibodies 192, 342, 394 Antibody formation 211 Antigens 192,211,214,394 Antimetabolites 193 — , artificial 206 — , natural 206 Antimutagens 145 Antiserum 214 Apo-enzyme 204 Arginine 70, 90, 109 Armadillo, nine banded 360 Arrangement, linear 18 Ascorbic acid 208 Asparagine 70 Aspartic acid 70, 90, 109 Auger cascades 270 Averaging, effect of 23 Axon 386 Bacillus sitbtilis 111 Bacterium 136. 178, 215, 277, 299, 301 Binary digits See Bits Binary symbol 9 Biological material, inactivation of 262, 276 et seq., 287 Bioplast 218 Biosynthesis 117,145 — , of protein 124 — , purine 1 38 Biotin 208 Biotopology 218 bis-2, 4-dinitrophenyl-L-cystine 284 bis-DNP cystine 284 , gamma-ray irradiation of 285 Bits 21 Blepharoplast 219 Bohr magneton 247 Bonds, interhelical 1 10 — , rupture by ionization 269 Born energy of polarization 265 et seq. Bovine albumin 252-3 (Fig. 10) CO,, level in blood, as O., regulator 7 — , , as symbolization 7 Caffeine 136, 137 Cage effect 252, 255, 257, 273, 353 Calandra oryzae 303, 306 Cancer 342, 343, 344, 347, 353, 355 Cariboxypeptidase 79, 86 Carcinogenic agents 354, 355 Carcinogens 294, 309, 347, 351 Carotene 173,208 Casein, a 79, 86 Cat, cochlear nucleus 157 Catalase 84 Category 6, 10, 12, 40 Cerebral cortex 381 Channel 30 — capacity 32, 194 theorem 3 1 , 34, 39, 298 Channel communication 31 — , noise-free 31 411 412 SUBJECT INDEX Chemical potential 199 Chick 208 Chick embryo See Embryos Chlamydomonas 301, 306 Chloracetyl-DL-alanine 252-3 (Fig. 5) Chloramphenicol 88 Chlorophyll 174 Choline 208 Chromidion 218 Chromosome 63, 65, 219, 224, 312, 359, 373 — aberrations 55, 298, 301 Chymotrypsin 113 Chymotrypsinogen 79, 87, 104 Cilium 219 Citruline 71 Clupein 79 Coacervation 219 Cocarboxylase 204 Cochlear nucleus 157 Code 30, 34, 219, 298 — , antigen-antibody 211 — , binary 9 — , DNA-protein 52, 67, 194 — , error correcting 32, 281 — , Fano 11, 15, 16, 18 —, Morse 8,10,41,193 — , over-lapping 67, 92 — groups, combination of 10 Coding — problem, protein text 87 Co-enzymes 192, 204 Co-factors 204 Collagen 72, 79 CoUinearity 380 Communication 25 — channel 3 1 — -engineering 187 — ■ systems 30, 35 Conditioned reflexes 7 Congolese moth See Anaphe maloneyi Congruence, displacement 376 — , self 376 Conjugation 309, 310 Conservation principle, one-sided 189 — , two-sided 189 Constraint, measures of 27, 188 Constraints 37, 105 Continuity, in information functions 21 Control, in the animal and in the machine 187 Correlation intersymbol 83 Cortex surface 383 Creatine-phosphate-ATP system 225 Cross reaction 212 Cryptography 70 Cybernetics 5 Cysteamine 259 Cysteine 70, 101, 129, (Figs. 9, 10), 253, 255, 258 — residues 109 Cystine 71, 90, 101, 251, 252-3 (Figs. 4, 9), 253, 255, 256, 258, 259, 271, 283, 284, 288, 290, 292 Cystinyl residue 71 Cytidine deaminase 286 Cytidylic acid 70, 90 Cytochrome 74, 79 Cytoplasm 63, 65, 89, 150, 218 Cytosine 63, 142 Dahlbomimis fuscipennis 306 Dancoff's principle 50, 53, 111, 190, 192, 298 Darwinian machines 188 Death, definition of 54 Decision theory 5 Decoding 6, 31 Degenerative diseases, chronic 348 Denaturation 283, 292 — , of proteins 117,267,268,271, 274, 286, 287 Deoxyribonucleic acid See DNA Diabetes melHtus 342, 344 Diamino pimelic acid 71 Dictyosome 219, 222 Dielectric absorption 264 — properties 263 Differentiation, capacity for 125 Dipeptides, combinations of 78, 82 Diphenyl picryl hydrazyl (DPPH) 249 Diploid survivorship 298, 301 Direct action 253, 258, 272, 276, 300, 301, 354 Disaccharide 212 Displacement congruence 376 Disulfide bond 116, 239, 255, 257, 258, 259, 271, 278, 283, 286, 287 et seq., 292 Disulfide interchanges, radiation-induced 283 et seq. DNA (deoxyribonucleic acid) 51, 52, 63, 70, 76, 88, 101, 117, 138, 245, 269, 272, 282, 298, 359, 394 — inactivation of by radiation 269, 272 — synthesis, steady-state flux of 142 — Watson and Crick model of 51, 55, 63, 65,90, 142, 189 Drosophila 53, 303, 304, 312 — melatwgaster 293, 304, 307, 337 — • subobscura 307 Earthworm See Lumbricoides terrestris E. coli See Escherichia coli Edestin 104 Egg albumin See Ovalbumin Electrochemical potential 200, 261, 292 I SUBJECT INDEX 413 Electron-spin resonance 241-261, 292, 353 — hole (electron vacancy) 257, 258, 272, 292 — hole trap 257, 258, 272, 278 — trap 256, 257, 259, 265, 266 Electrophysiology 197 Embryonic protein synthesis 130 Embryos 124 Emulsin 114 Encoding 6, 12, 16, 31 Endoplasm 221, 222 Ensemble 20,191,307 — , Gibbsian 277 — , of genetic messages 53 — , of organisms 51,52,297 — , probability measures of 193 Entropy 21, 44, 52, 1 19, 171, 198, 269, 287, 298, 317,324,359 — , irreversible production of 198 — surface, equilibrium position on 200 Enzymatic inhibition 139 Enzymes 63, 111, 121, 125, 140, 192, 197, 242, 263, 276, 278 Enzyme-substrate complex 171, 180,206 Equilibrium position, on the entropy surface 200 Equivocation 31, 52, 55, 302 — , in the germ line 298, 309 Error detection and detection 32 Escherichia coli 73, 83, 92, 101, 106, 137, 138, 215 , B/r strain of 300 , purine metabohsm in 138,144 Estrogens 208 Euchlanis triquetra 309, 310 Europium chloride 173 Events, groups of 12 — , rare 42 — , real 7, 8 — , symbolic 7 Evolution 112 Excited molecules 273 Experimental design theory 5 Eye, human 373 — , insect 373 — , spot 222 Fano code See Code, Fano Fatty acids 242 F-center 241 Feather quill, irradiated Fig. 1 1 Feedback 188,223,372,385,389 — , negative 148 Fibrinogen 75, 87 Fission, binary 223 Fixation tremor 379 Flagellum 219 Fluctuations 160, 200, 317, 320, 324, 328 Fluorescence spectrum of solutions 173 "Fluxes" 199 Folic acid 208 Force of mortality 298, 299 "Forces" 199 Formate 1 38 Franck-Condon principle 255, 257 Free energy 1 1 7 — radicals 242, 243, 246, 252 et seq., 258, 259, 283, 353 Frog 208 See Rana pipiens — , sciatic nerve 154, 155 Fruit fly See Drosophila melanogaster Functional geometry 375 .^-Factor 243 Game theory 38 y-Globulin 79, 86, 212 Gastrocnemius nerve 1 58 , in decapitated cats 157 Gaunine 63 Gene 73, 88, 224 — code 310 — mutations 298, 301 General systems theory 5 Genetic information 51, 70, 190, 306, 338 , storage and transfer of 298 Genetic locus 190 — noise 52, 55, 189, 298, 309, 313 Genome 53, 54, 55, 310 Germ cell 359 — line, equivocation 298, 309 Giant starfish 215 Gliadin 73, 75, 104 Globulin 79, 129, 212 Glucagon 79, 87 Glutamine 70 Glutamic acid 70, 90, 109, 131 Glutathione 253, 255, 258, 259 Glyceraldehyde dehydrogenase 86 Glycine 70, 90, 138, 242 Gly co-protein 212 Glycyl DL-valine 252-3 (Fig. 6) Glycyl-glycine 252-3 (Fig. 10) Glycyl glycine, deuterium substitution 354 Glycyl glycyl glycine 252-3 (Fig. 5) Golgi apparatus 222 Gompertz function 323, 350, 351 Grain beetle See Calandra oryzae Gramicidin, action spectra for 288 Grasshopper See Melanoplus differentialis Green algae 306 Growth hormone 104 Guanine 63, 136, 137, 138, 141 — deoxyriboside 138 Guanosine 138 Guanylic acid 70, 90 Guinea pig 208 414 SUBJECT INDEX Habrobracon juglandis 306, 307 Haploid 300, 306 — survivorship 298 Heart, chicken 129 Helix 65,110,117,142 — , alpha 254 — , double 141, 190 Hemocyanins 212 Hemoglobin 73, 75, 79, 85, 86, 90, 94, 104, 114, 268, 272 Hepatectomy 148 Heterophile reaction 212 Histidine 70, 90 Hit theory See Target theory Holo-enzyme 204 Homeostasis 318, 326, 342 Homocysteine 71 Hormones 191,204,242 Horse mussel 215 Humoral system, of communication 148 Hydrogen bond 109, 117, 266, 287 et secj. Hydroxyglutamic acid 71 Hydroxylysine 70, 72 Hydroxyproline 70, 72 Hyperfine structure 250 Hypertensive peptide 75, 80 Hypoxanthine 137 Immunopolysaccharides 88 Inactivation of proteins, by ionizing radiation 262, 287 Independence 21 Indirect action 258, 300, 301, 354 Individuality 54 Influenza A virus 90 Information 230, 317, 324 — , chemical 394 — , conservation of 189 — , decay of 55 — , destruction of by ionizing radiation 239 — , Fisher 325 — , genetic 51, 70, 190, 306, 338, 359 — , measure of 18, 20, 25, 40, 169, 188 — , rate of production 198,199 —, redundant 33,189,292 Information, replication 87 —, storage 61,87,393 — , transfer 61, 87 — , transmission of 112,211,298 — and entropy 21, 44, 171 — and intelligence 6 — assembly 373 — content 20, 21, 39, 41, 103, 116, 212, 218,235,276, 309, 359,400 , carried by protein sequence 103, 113 , configurational 105,114 , destruction of 297 -. of molecules 193 Information content, of tracer data 181 , per printed letter 44, 281 , structural 1 1 1 , total 105 , total carried by proteins 113 — continuum 6 — functions 26, 31, 43 — measures, Shannon-Wiener 21 — measures 103 , determination of 169 , functions of ensembles 192 , of constraint between letters 281 , relativity of 40, 191, 193, 235 , the search for invariants 190 , topological representation of molecules 193 — representation 6, 7, 8, 9 — theory 5, 7, 50, 230, 276, 399 , domain of in biology 187 — transfer 65, 125 — transmission, by biological co-factors 204 , from genetic to somatic material 189 , process of 1 12, 189 , rate of 44 Inheritance, cytoplasmic 88 Inhibition, enzymatic 139 Inhibitor 208 Injury, irreparable 331 — , irreversible 336 — , radiation 271, 331 — , reparable 331 Inosine arsenolysis, inhibition of 140 — hydrolase 140 Input information 32 Insulin 71,74,104,107,114,125, 252-3 (Fig. 12), 256, 268, 288, 290, 343 — (cattle) 80 — 'A' chain 66, 69, 80, 87 — 'B' chain 69, 80, 87 Intelligence and information 6, 7, 20 Internal emitters 338 Interphase 224 Intersymbol influence 56, 67, 71, 78, 105, 115 Invertase 115 Ionization 243, 255, 264 et seq., 278 Ionization, in proteins 266 Ionizing radiation 252 et seq., 262 et seq., 273,283,331, 333, 355 • , latent injury from 271, 272, 331 Isoleucine 70, 90 Isopropylbenzene 291 Karyoplasm 221 Keratin protein, /? Kinetodesma 224 Kinetosome 224 253 SUBJECT INDEX 415 Kinety 224 Lactalbumin, a 86 Lactogenic hormone 104 Lactoglobulin, /5 80, 86, 104 Lacto peroxidase 114 Language 8,44,104,111,193 L-cystine 284 Late eiTects, from radiation damage 301 LD50 332, 333 Learning 392 LET 269, 270, 279 et seq. Lethal bound, limit 298, 318, 326 Leucine 70, 90 Leucosin 115 Leukemia 343, 345 Life expectancy 333, 334, 344 — span 312,331,334,342 Limitations, of information theory 192 Linear programming theory 5 Lipo-protein 212 Liver, chicken 129 — regeneration 148 Logical machine 37 Longevity 331 Lumbricoides terrestris 215 Lysine 70, 72, 90, 109 Lysozyme, papaya lysozyme 80, 86, 107, 113 Macrostate 118 Malignant neoplasms See Cancer Markoffian, fluctuation process 320 Mass action, law of 151 Maximum likelihood estimate 213 Maxwell's demon 120, 196 Melanophore 80, 87 Melanopliis differentia/is 215 Membrane phenomena 197 Membranes, exclusive 198 — , indifferent 197 — , responsive 198 Memory 393, 394 Message 30, 33 — , genetical 5 1 — entropy 52, 54, 298 — sets 191 Metabolism, purine 138 Methionine 70, 90 Methylcholanthrene 294, 309 Methyl xanthines 136 Micelle 109,218 Micro-organisms 101,131,208 Microsome 66, 132 Microstate 118 Microtus agrestis 303 Microwave spectroscopy 242, 353 Mitochondrion 132,219 Molecular energy 118 — dissociation process 264 Molecules, irreversibly inactivated 120 Monosaccharide 212 Mono-DNP cystine 284 Mono-2, 4-dinitrophenyl-L-cystine 284 Morbidity 342 Morphogenesis 218, 359 Morse code See Code, Morse Mortality 317,319,342 Mosaics, non-addressed 372 — , pre-addressed 372 Moth, Congolese (Anaphe maloiwyi) 87 Motoneurons 157 Mouse 293,304,309,333,337,351 — , LAFi 294, 307 — , R.F. 294, 307 Mutagenesis 136 — , chemical 145 Mutagenic agents, chemical 136, 145, 309 , DNA 137 , RNA 137 Mutagenic effects 355 Mutagenicity 137 Mutants, biochemical 300 Mutation rate, spontaneous 58, 136, 141, 145 Mutations 65, 71, 74, 84, 137, 298, 353, 355 — , point 55, 56, 146 — , recessive lethal 301 Myeloma globulin 86 Myoglobin 74, 80, 86, 104 Myosin 104,113 Nebenkern 219 Negentropy, rate of production of 199 Neoplasms, malignant See Cancer Nerve fiber 153 — membrane 201 — optic 389 - — protein 86 Nervous system 347 Neural connections 383 Neural network 372 — systems 153 Neural thresholds 153 Neurospora 13 Neuron (See also Motoneurons) 384, 389, 393, 395 Nicotinamide 208 Nissl substance 222 Noise 31, 35, 51, 53, 165, 189, 281, 319, 379 — , chemical 172 — , genetic 52, 189, 298, 309, 313 — , measures of 188 — , random Gaussian 320 — , "white" 55, 320 416 SUBJECT INDEX Noise-and-redundancy-theorem 188, 189, 298 Non-addressed systems 373, 388, 396 Norleucine 71 Nuclear hyperfine structure 244 Nucleic acid metabolism, enzymes of 140 — acid residues 191 — acids {See also DNA, RNA) 18, 138, 144, 209, 242 Nucleolus 219 Nucleotide pairs 51, 52 Nucleotides 8, 18, 65, 70, 76, 91, 298 Nucleus 219 Odd-coin problem 233, 401 OH radicals 258, 259, 284 Operations analysis 5 Order 268, 276, 362 Orderliness, living things feed on 117, 189 — , problem of destruction of 188 Organ, control 191 Organ, target 191 Organelle decision trees 221 Organelles 218 Organization 36, 39, 218, 262 Organ-specific transfer 129 Origin of life 278 Ornithine 71 Output information 32 Ovalbumin (egg albumin) 80, 1 14, 128, 212, 268, 272, 289 — , peptides 128 Ovum, fertilized 51,63 Oxytocine 66, 69, 80, 87 p-Xmmo benzoic acid 208 Pantothenic acid 208 Papain 80,86,113 Parallelism 381 Paramagnetic resonance {See Electron-spin resonance) Paramecium 220-225, 306 Paschen-Back effect 246 Pathology of aging 293 Pathology of radiation damage 293 Pattern perception 371, 374, 388 Pauli principle 243 Pellicle unit 224 Peniculus 225 Pepsin 80,104,113,116,268 Pepsinogen 87 Peptide bonds 263 Peptides 215,242,252 — , X-irradiated 253 Peroxidase (milk) 268 Phenylalanine 70, 90 Philodina citrina 309, 310 Phosphoglucomutase 1 1 6 Phosphorylase 86, 144 Phosphorylated derivatives 138 Phosphoserine 70, 72 Physiologic variables 319,322 Plasmagenes 219 Plasmapheresis 148 Ploidy 312 Point mutation 55, 56, 146 Poisson distribution 83, 107 Polarization, electronic 266 — , orientation 267 — , secondary-bond 266 — , vibrational 266 Polypeptides 273 Polypeptide structures, minimum entropy 117 Polystyrene, X-irradiated 256 Pooling, effect of 24 Praseodymium chloride 172 Pre-addressed 373 Preservation, of cells, tissues at low tempera- tures 178 Probabilities, conditional 28 — , unequal 13 Probability density 247 — distribution 231,298 , joint 320 Problem-solving process 234 Prolactin 80 Proline 70,72,90 Protamine 74 Protein 18, 202, 204, 209, 212, 242, 263, 273, 283, 287 — and peptide chains 87 — , composition of 90, 101 — precursors, incorporation into chick embryos 128 Proteins, denaturation of, reversible 117, 267, 288 — , genetic determination of 52, 73 — , homologous 74 — , inactivation of 262 et seq., 287 — , ionized 257 — , radiation damage in 253, 262 et seq., 287 — , terminal residues of 77 Protein specificity 18, 50 — structure 103,111,267 and language 105, 1 12 — synthesis 65, 125, 135, 144, 298 , mechanism for 65, 125 Prothrombin 86 Proto-RNA 89 Pseudo-holo-enzyme 204 Psychology 191,231 Punctuation mark 8, 69, 91 SUBJECT INDEX 417 Purines 63 et seq., 139 — , biosynthesis 138 — , derivatives, structure 136 — , incorporation and mutagenicity 137 — , metabolism 138 — , phosphorylases 140 Purine-purine transglycosidase 140 Purine-pyrimidine transglycosidase 140 Pyridoxine 208 Pyrimidine phosphorylase 140 Pyrimidines 65 Quandrulus 222, 225 Quantum theory 196, 231 , 242, 247 q. 347 Radiation {See Ionizing radiation, Ultraviolet radiation) — damage 241, 252 et seq., 262 et seq., 276 et seq., 283 et seq., 287 et seq., 308, 395 , protection from 273 — exposure 342 — hazards to man 297, 301, 338 — injury 253, 331 , irreversible 273, 339 Radiation sensitivity 269 Radicals, free 259 ^, lifetime of 243, 252 et seq. Radiobiology 190, 239, 252 et seq., 262 et seq., 276 et seq., 283 et seq., 287 et seq., 292 et seq., 293, 297 Radiomimetic chemicals 355 Radiomimetics 309 Random networks, theory of 1 88 — variables 230 Rat 148, 208, 293, 335, 337 Rate processes 327 Rattus natalensis 303 Reaction kinetics 56 Receptors 372 Recessive lethal mutations 301, 312 Recovery 308, 322, 335 — , from radiation damage 332 Redundance 3, 33, 111, 116, 369 — , in the germ cell 359, 360 — , measures of 188 Redundant information 3, 33, 189 Rennin 115 Replication, capacity for 125 RepresentabiUty condition 31 Representation, binary 8, 17, 20 — theorem 17, 188 Residues, correlations between adjacent 78 Reticulocytes 94 Retina 383 ReversibiHty 121 Riboflavin 204, 208 Ribonuclease 69, 81, 86, 87, 104, HO, 116, 125, 285, 288 Ribonucleic acid See RNA Ribonucleoprotein 150 — , associated basophilic bodies 150 RNA 8, 52, 66, 70, 88, 90, 92, 101, 131, 138, 278, 394 Rotifer 178, 309, 313 Ruling engines 376 Salivary amylase 86 Salmine 81, 104 Salmonella gallinarum, typhimurium 54 Salt bridges 109, 267 Sample space 231,235 Sarcosome 222 Sciatic nerve, of frog 154, 155 Screws, precision 376 Sea pen 215 Selection 322 — rules 191 Semantic information 194 Sequence 51, 107, 113, 135, 191, 194 Serine 70, 90, 138 — , protein-bound 72 Serum, normal blood 148 — albumin 74, 81, 86, 104, 128, 150, 289, 290 • , chicken 127 Silk, irradiated 353 — , X-irradiated 252-3 (Fig. 8) — fibroin (Bombyx) 8 1 , 1 04 Site, sense, non-sense 91 Solanin 115 Somatic line 309 Somatic mutation 344 Soret band 174 Southern bean mosaic virus 90, 93 Specificity See Protein specificity — , antigenic 211 — , code of 194 Spectroscopic splitting factor, g 242 Spin-orbit coupling 249 Spleen homogenate 308 Spores 276 S-Sbond 116 Staphylococcus aureus 1 3 1 Stentor 223 Stereocilium 222 Stimulus, threshold of 154 — threshold variations, possible sources of 159 Stochastic process 321 S^5 126 Sulphur linkages, in proteins 109, 116, 239, 257, 278, 283, 287, 292 Survival curves, exponential 300 , sigmoid 300 418 SUBJECT INDEX Survivorship curves 51, 297, 317 Symbol, 8, 30, 390 Symbolic events See Events, symbolic — representation 7 Symbolization 7 Synapse 157 Synaptic delay 234 System analysis 37 — parameters 182 — theory 5 Systems 36 See also Communication sys- tems — , multipart 38 • — , two-part 28, 30 Target 276, 281 — theory 276, 298, 300, 301, 305 — volume 276, 289 Tautomeric form, of purines 141 Temperatures, chemistry and biochemistry at low 171 Template 67, 71 Theobromine 136, 137 Theophylline 137 Thermal killing 297, 305, 306 Thermodynamics 52, 192, 196 — , equation of state in 307 Thermodynamics, irreversible 194, 198 — , second law of 117, 120, 196 Thiamine 208 Threonine 70, 90 Threshold, fluctuating 153, 161 Thymine 63, 141 Thyroglobulin 72 Thyroxine 72, 207, 208 Tissue homogenates, incorporation of acti- vity from 130 r-Measure 27, 29 Tobacco mosaic virus 70, 76, 81, 90, 93, 278 Toluene 291 Tomato bushy stunt virus 90 Torulopsis utiUs 126, 131 Transducer 3 1 Transforming principle 276 Transphosphorylase 86 Tribolium confiisum 293 Trichocyst 220, 221 Triticum durum, vulgare 73 Tropomyosin 81, 86, 104 Tropylium ion 291 Trypsin 114,268,288 Trypsinogen 81, 87 Tryptophan 70, 90 Tumors 309 Turnip yellow virus 90, 93 Twins, monozygotic 360 Tyrosine 70, 90, 109 Tyrosine-O-sulfate 72 Uhraviolet radiation 252, 263, 273, 274, 287, 290,297,299, 306, 312, 353 — action spectra 274, 288 Uncertainty 19, 27, 32, 41, 230, 235, 368 — , conditional 28 — , — , average 30 — , functions 207 — , joint 27 — , mappings in generalized space 184 — , measure of 19, 21 — , relative 369 —, unit of 185 Unitization 29, 218 Unsaturated fatty acids 208 Uridylic acid 70, 90 Urosil 66 Valine 70, 90 Van der Waals forces 109 Variables 25 Variance 322 Vascular disease 342, 343 Vasopressin 66,69,74,81,87 Vibriolysin 115 Virus 223, 276, 299, 301 Vitamins 204, 242 — , B 207 — , D 208 Vole 304 Wool 81 Word 8, 107, 212, 363 Xanthine 136, 137 X-rays 243, 252, 290, 293, 300, 305, 337, 353, 355 Yeast 215,300,301,310 —.diploid 301,305 Yeast invertase 268 Yule distributions 123 Zein 104 ZoUner illusion 385